var/home/core/zuul-output/0000755000175000017500000000000015114741567014540 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015114757374015507 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005544473215114757364017730 0ustar rootrootDec 06 05:51:14 crc systemd[1]: Starting Kubernetes Kubelet... Dec 06 05:51:14 crc restorecon[4683]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:14 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 05:51:15 crc restorecon[4683]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 05:51:15 crc restorecon[4683]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 06 05:51:15 crc kubenswrapper[4809]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 06 05:51:15 crc kubenswrapper[4809]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 06 05:51:15 crc kubenswrapper[4809]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 06 05:51:15 crc kubenswrapper[4809]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 06 05:51:15 crc kubenswrapper[4809]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 06 05:51:15 crc kubenswrapper[4809]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.224731 4809 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.230655 4809 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.230704 4809 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.230720 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.230731 4809 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.230740 4809 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.230749 4809 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.230758 4809 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.230768 4809 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.230778 4809 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.230787 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.230797 4809 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.230807 4809 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.230816 4809 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.230825 4809 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.230832 4809 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.230840 4809 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.230847 4809 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.230855 4809 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.230873 4809 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.230882 4809 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.230890 4809 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.230897 4809 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.230905 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.230913 4809 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.230921 4809 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.230964 4809 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.230972 4809 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.230980 4809 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.230987 4809 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.230995 4809 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.231006 4809 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.231017 4809 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.231026 4809 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.231035 4809 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.231044 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.231052 4809 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.231060 4809 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.231068 4809 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.231076 4809 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.231084 4809 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.231092 4809 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.231100 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.231107 4809 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.231115 4809 feature_gate.go:330] unrecognized feature gate: Example Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.231126 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.231134 4809 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.231142 4809 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.231150 4809 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.231157 4809 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.231164 4809 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.231175 4809 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.231185 4809 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.231196 4809 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.231204 4809 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.231212 4809 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.231220 4809 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.231227 4809 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.231235 4809 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.231242 4809 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.231250 4809 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.231258 4809 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.231265 4809 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.231273 4809 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.231281 4809 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.231288 4809 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.231297 4809 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.231304 4809 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.231312 4809 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.231321 4809 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.231333 4809 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.231343 4809 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.231829 4809 flags.go:64] FLAG: --address="0.0.0.0" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.231850 4809 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.231861 4809 flags.go:64] FLAG: --anonymous-auth="true" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.231870 4809 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.231878 4809 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.231884 4809 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.231894 4809 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.231905 4809 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.231911 4809 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.231917 4809 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.231924 4809 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.231955 4809 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.231962 4809 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.231969 4809 flags.go:64] FLAG: --cgroup-root="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.231974 4809 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.231980 4809 flags.go:64] FLAG: --client-ca-file="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.231986 4809 flags.go:64] FLAG: --cloud-config="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.231991 4809 flags.go:64] FLAG: --cloud-provider="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.231996 4809 flags.go:64] FLAG: --cluster-dns="[]" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232006 4809 flags.go:64] FLAG: --cluster-domain="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232011 4809 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232017 4809 flags.go:64] FLAG: --config-dir="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232022 4809 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232029 4809 flags.go:64] FLAG: --container-log-max-files="5" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232039 4809 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232045 4809 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232051 4809 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232057 4809 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232063 4809 flags.go:64] FLAG: --contention-profiling="false" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232069 4809 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232074 4809 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232080 4809 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232086 4809 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232093 4809 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232098 4809 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232103 4809 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232109 4809 flags.go:64] FLAG: --enable-load-reader="false" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232115 4809 flags.go:64] FLAG: --enable-server="true" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232120 4809 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232128 4809 flags.go:64] FLAG: --event-burst="100" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232133 4809 flags.go:64] FLAG: --event-qps="50" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232139 4809 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232145 4809 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232153 4809 flags.go:64] FLAG: --eviction-hard="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232161 4809 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232166 4809 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232172 4809 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232178 4809 flags.go:64] FLAG: --eviction-soft="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232184 4809 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232189 4809 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232195 4809 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232201 4809 flags.go:64] FLAG: --experimental-mounter-path="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232206 4809 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232212 4809 flags.go:64] FLAG: --fail-swap-on="true" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232217 4809 flags.go:64] FLAG: --feature-gates="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232226 4809 flags.go:64] FLAG: --file-check-frequency="20s" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232232 4809 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232238 4809 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232244 4809 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232249 4809 flags.go:64] FLAG: --healthz-port="10248" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232255 4809 flags.go:64] FLAG: --help="false" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232260 4809 flags.go:64] FLAG: --hostname-override="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232265 4809 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232271 4809 flags.go:64] FLAG: --http-check-frequency="20s" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232276 4809 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232281 4809 flags.go:64] FLAG: --image-credential-provider-config="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232286 4809 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232291 4809 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232296 4809 flags.go:64] FLAG: --image-service-endpoint="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232301 4809 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232306 4809 flags.go:64] FLAG: --kube-api-burst="100" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232311 4809 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232317 4809 flags.go:64] FLAG: --kube-api-qps="50" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232323 4809 flags.go:64] FLAG: --kube-reserved="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232328 4809 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232334 4809 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232339 4809 flags.go:64] FLAG: --kubelet-cgroups="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232345 4809 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232351 4809 flags.go:64] FLAG: --lock-file="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232357 4809 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232363 4809 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232369 4809 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232379 4809 flags.go:64] FLAG: --log-json-split-stream="false" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232385 4809 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232390 4809 flags.go:64] FLAG: --log-text-split-stream="false" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232396 4809 flags.go:64] FLAG: --logging-format="text" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232402 4809 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232413 4809 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232419 4809 flags.go:64] FLAG: --manifest-url="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232425 4809 flags.go:64] FLAG: --manifest-url-header="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232435 4809 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232441 4809 flags.go:64] FLAG: --max-open-files="1000000" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232450 4809 flags.go:64] FLAG: --max-pods="110" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232456 4809 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232462 4809 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232467 4809 flags.go:64] FLAG: --memory-manager-policy="None" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232473 4809 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232479 4809 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232485 4809 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232492 4809 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232510 4809 flags.go:64] FLAG: --node-status-max-images="50" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232516 4809 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232521 4809 flags.go:64] FLAG: --oom-score-adj="-999" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232528 4809 flags.go:64] FLAG: --pod-cidr="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232533 4809 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232544 4809 flags.go:64] FLAG: --pod-manifest-path="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232550 4809 flags.go:64] FLAG: --pod-max-pids="-1" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232556 4809 flags.go:64] FLAG: --pods-per-core="0" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232561 4809 flags.go:64] FLAG: --port="10250" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232568 4809 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232573 4809 flags.go:64] FLAG: --provider-id="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232579 4809 flags.go:64] FLAG: --qos-reserved="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232585 4809 flags.go:64] FLAG: --read-only-port="10255" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232591 4809 flags.go:64] FLAG: --register-node="true" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232597 4809 flags.go:64] FLAG: --register-schedulable="true" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232604 4809 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232615 4809 flags.go:64] FLAG: --registry-burst="10" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232621 4809 flags.go:64] FLAG: --registry-qps="5" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232627 4809 flags.go:64] FLAG: --reserved-cpus="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232636 4809 flags.go:64] FLAG: --reserved-memory="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232650 4809 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232656 4809 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232662 4809 flags.go:64] FLAG: --rotate-certificates="false" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232668 4809 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232674 4809 flags.go:64] FLAG: --runonce="false" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232680 4809 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232685 4809 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232691 4809 flags.go:64] FLAG: --seccomp-default="false" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232697 4809 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232703 4809 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232709 4809 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232715 4809 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232721 4809 flags.go:64] FLAG: --storage-driver-password="root" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232727 4809 flags.go:64] FLAG: --storage-driver-secure="false" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232733 4809 flags.go:64] FLAG: --storage-driver-table="stats" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232739 4809 flags.go:64] FLAG: --storage-driver-user="root" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232744 4809 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232750 4809 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232756 4809 flags.go:64] FLAG: --system-cgroups="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232762 4809 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232771 4809 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232776 4809 flags.go:64] FLAG: --tls-cert-file="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232781 4809 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232791 4809 flags.go:64] FLAG: --tls-min-version="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232797 4809 flags.go:64] FLAG: --tls-private-key-file="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232804 4809 flags.go:64] FLAG: --topology-manager-policy="none" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232811 4809 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232817 4809 flags.go:64] FLAG: --topology-manager-scope="container" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232823 4809 flags.go:64] FLAG: --v="2" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232832 4809 flags.go:64] FLAG: --version="false" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232841 4809 flags.go:64] FLAG: --vmodule="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232859 4809 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.232868 4809 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233041 4809 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233050 4809 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233055 4809 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233060 4809 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233067 4809 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233073 4809 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233080 4809 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233086 4809 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233091 4809 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233097 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233102 4809 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233107 4809 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233111 4809 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233116 4809 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233121 4809 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233126 4809 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233130 4809 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233135 4809 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233140 4809 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233145 4809 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233150 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233154 4809 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233159 4809 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233163 4809 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233169 4809 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233174 4809 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233178 4809 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233183 4809 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233188 4809 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233193 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233201 4809 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233207 4809 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233212 4809 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233218 4809 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233225 4809 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233230 4809 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233235 4809 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233240 4809 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233245 4809 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233249 4809 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233254 4809 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233259 4809 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233264 4809 feature_gate.go:330] unrecognized feature gate: Example Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233268 4809 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233273 4809 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233278 4809 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233282 4809 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233287 4809 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233292 4809 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233296 4809 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233301 4809 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233305 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233310 4809 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233315 4809 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233320 4809 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233325 4809 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233329 4809 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233333 4809 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233338 4809 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233343 4809 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233347 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233352 4809 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233358 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233365 4809 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233370 4809 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233374 4809 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233379 4809 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233383 4809 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233390 4809 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233396 4809 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.233403 4809 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.233420 4809 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.240031 4809 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.240084 4809 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240164 4809 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240178 4809 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240183 4809 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240189 4809 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240195 4809 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240200 4809 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240204 4809 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240208 4809 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240212 4809 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240216 4809 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240220 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240224 4809 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240228 4809 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240232 4809 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240236 4809 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240239 4809 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240243 4809 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240247 4809 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240251 4809 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240255 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240260 4809 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240266 4809 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240271 4809 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240276 4809 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240283 4809 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240292 4809 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240297 4809 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240301 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240305 4809 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240309 4809 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240313 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240318 4809 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240322 4809 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240326 4809 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240330 4809 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240334 4809 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240337 4809 feature_gate.go:330] unrecognized feature gate: Example Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240341 4809 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240345 4809 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240349 4809 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240353 4809 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240357 4809 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240360 4809 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240364 4809 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240368 4809 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240372 4809 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240376 4809 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240380 4809 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240384 4809 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240388 4809 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240392 4809 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240396 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240400 4809 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240404 4809 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240408 4809 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240412 4809 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240416 4809 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240420 4809 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240423 4809 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240427 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240430 4809 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240434 4809 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240438 4809 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240443 4809 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240447 4809 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240451 4809 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240455 4809 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240459 4809 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240464 4809 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240469 4809 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240473 4809 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.240481 4809 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240595 4809 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240605 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240610 4809 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240615 4809 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240619 4809 feature_gate.go:330] unrecognized feature gate: Example Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240624 4809 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240628 4809 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240633 4809 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240637 4809 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240644 4809 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240648 4809 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240651 4809 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240655 4809 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240659 4809 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240664 4809 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240668 4809 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240671 4809 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240677 4809 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240681 4809 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240685 4809 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240689 4809 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240693 4809 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240697 4809 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240701 4809 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240706 4809 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240710 4809 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240715 4809 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240719 4809 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240724 4809 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240729 4809 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240734 4809 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240738 4809 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240743 4809 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240747 4809 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240751 4809 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240755 4809 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240759 4809 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240762 4809 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240767 4809 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240771 4809 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240775 4809 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240779 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240782 4809 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240786 4809 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240790 4809 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240794 4809 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240797 4809 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240801 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240805 4809 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240809 4809 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240813 4809 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240817 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240821 4809 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240825 4809 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240829 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240833 4809 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240836 4809 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240840 4809 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240843 4809 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240847 4809 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240850 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240855 4809 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240859 4809 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240863 4809 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240867 4809 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240871 4809 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240875 4809 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240879 4809 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240883 4809 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240886 4809 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.240891 4809 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.240898 4809 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.241399 4809 server.go:940] "Client rotation is on, will bootstrap in background" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.244547 4809 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.244647 4809 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.245148 4809 server.go:997] "Starting client certificate rotation" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.245180 4809 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.245670 4809 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-09 06:45:34.90657907 +0000 UTC Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.245769 4809 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 72h54m19.660812171s for next certificate rotation Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.249806 4809 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.251405 4809 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.260454 4809 log.go:25] "Validated CRI v1 runtime API" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.292631 4809 log.go:25] "Validated CRI v1 image API" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.294256 4809 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.297067 4809 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-06-05-46-33-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.297114 4809 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.313856 4809 manager.go:217] Machine: {Timestamp:2025-12-06 05:51:15.312078613 +0000 UTC m=+0.201061575 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:5931df91-c1df-4f78-9e5f-391a01a76127 BootID:88a28b6b-9c85-4932-b5bc-36346303aee5 Filesystems:[{Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:83:ce:09 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:83:ce:09 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:9d:69:3f Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:6a:7b:c9 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:9b:87:f7 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:3e:12:37 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:86:f2:2d:86:fe:72 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:52:8d:22:e9:99:9c Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.314659 4809 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.314973 4809 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.315555 4809 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.315822 4809 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.315878 4809 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.316205 4809 topology_manager.go:138] "Creating topology manager with none policy" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.316219 4809 container_manager_linux.go:303] "Creating device plugin manager" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.316416 4809 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.316469 4809 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.316751 4809 state_mem.go:36] "Initialized new in-memory state store" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.317341 4809 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.317990 4809 kubelet.go:418] "Attempting to sync node with API server" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.318020 4809 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.318051 4809 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.318072 4809 kubelet.go:324] "Adding apiserver pod source" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.318088 4809 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.320467 4809 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.320991 4809 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.321285 4809 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.321340 4809 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Dec 06 05:51:15 crc kubenswrapper[4809]: E1206 05:51:15.321407 4809 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.154:6443: connect: connection refused" logger="UnhandledError" Dec 06 05:51:15 crc kubenswrapper[4809]: E1206 05:51:15.321443 4809 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.154:6443: connect: connection refused" logger="UnhandledError" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.321745 4809 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.322288 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.322310 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.322343 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.322351 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.322362 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.322370 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.322382 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.322393 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.322401 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.322411 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.322436 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.322447 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.322602 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.323247 4809 server.go:1280] "Started kubelet" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.323835 4809 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.324227 4809 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.324386 4809 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.325081 4809 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.326003 4809 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.326040 4809 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.326067 4809 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-09 02:47:00.76359908 +0000 UTC Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.326108 4809 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 812h55m45.437492696s for next certificate rotation Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.326403 4809 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.326451 4809 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.326554 4809 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 06 05:51:15 crc kubenswrapper[4809]: E1206 05:51:15.326640 4809 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 06 05:51:15 crc systemd[1]: Started Kubernetes Kubelet. Dec 06 05:51:15 crc kubenswrapper[4809]: E1206 05:51:15.327197 4809 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.154:6443: connect: connection refused" interval="200ms" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.330511 4809 factory.go:55] Registering systemd factory Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.330546 4809 factory.go:221] Registration of the systemd container factory successfully Dec 06 05:51:15 crc kubenswrapper[4809]: E1206 05:51:15.327303 4809 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.154:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187e8a66271b10d6 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-06 05:51:15.323195606 +0000 UTC m=+0.212178548,LastTimestamp:2025-12-06 05:51:15.323195606 +0000 UTC m=+0.212178548,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.331125 4809 factory.go:153] Registering CRI-O factory Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.331164 4809 factory.go:221] Registration of the crio container factory successfully Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.331250 4809 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.331281 4809 factory.go:103] Registering Raw factory Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.331297 4809 manager.go:1196] Started watching for new ooms in manager Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.333781 4809 server.go:460] "Adding debug handlers to kubelet server" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.334073 4809 manager.go:319] Starting recovery of all containers Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.336532 4809 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Dec 06 05:51:15 crc kubenswrapper[4809]: E1206 05:51:15.336608 4809 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.154:6443: connect: connection refused" logger="UnhandledError" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342112 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342200 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342218 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342231 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342244 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342257 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342268 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342280 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342295 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342308 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342322 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342337 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342352 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342368 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342382 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342395 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342412 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342426 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342440 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342453 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342467 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342480 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342493 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342509 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342524 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342538 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342554 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342568 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342582 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342596 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342609 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342621 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342704 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342717 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342731 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342743 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342755 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342767 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342780 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342792 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342804 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342817 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342831 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342843 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342854 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342866 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342879 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342892 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342908 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342925 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342952 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342965 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342982 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.342996 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343010 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343024 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343037 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343050 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343061 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343074 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343086 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343097 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343109 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343121 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343135 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343146 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343159 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343170 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343182 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343193 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343206 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343218 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343230 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343244 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343258 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343271 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343287 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343301 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343314 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343326 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343341 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343355 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343367 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343380 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343393 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343407 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343420 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343433 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343446 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343458 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343494 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343508 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343521 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343535 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343549 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343562 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343575 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343589 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343601 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343614 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343628 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343642 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343654 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343666 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343685 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343698 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343714 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343730 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343743 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343762 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343775 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343787 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343801 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343814 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343827 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343838 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343849 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343861 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343872 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343882 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343893 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343906 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343918 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343946 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343959 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343971 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343986 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.343998 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344011 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344024 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344035 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344046 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344060 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344071 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344082 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344094 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344106 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344116 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344126 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344136 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344144 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344155 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344164 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344172 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344182 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344196 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344206 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344221 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344231 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344260 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344268 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344277 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344290 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344301 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344313 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344324 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344332 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344340 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344348 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344356 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344366 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344376 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344387 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344400 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344410 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344421 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344432 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344445 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344456 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344467 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344479 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344490 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344502 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344517 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344528 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344540 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344559 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344571 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344582 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344594 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.344605 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.345274 4809 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.345297 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.345311 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.345324 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.345335 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.345345 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.345356 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.345366 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.345378 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.345390 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.345403 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.345416 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.345428 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.345440 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.345451 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.345463 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.345474 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.345485 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.345497 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.345509 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.345520 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.345533 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.345545 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.345559 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.345572 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.345584 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.345598 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.345610 4809 reconstruct.go:97] "Volume reconstruction finished" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.345619 4809 reconciler.go:26] "Reconciler: start to sync state" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.356053 4809 manager.go:324] Recovery completed Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.364816 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.367004 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.367049 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.367058 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.367947 4809 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.367966 4809 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.367990 4809 state_mem.go:36] "Initialized new in-memory state store" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.385174 4809 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.387179 4809 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.387256 4809 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.387312 4809 kubelet.go:2335] "Starting kubelet main sync loop" Dec 06 05:51:15 crc kubenswrapper[4809]: E1206 05:51:15.387370 4809 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 06 05:51:15 crc kubenswrapper[4809]: W1206 05:51:15.388561 4809 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Dec 06 05:51:15 crc kubenswrapper[4809]: E1206 05:51:15.388618 4809 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.154:6443: connect: connection refused" logger="UnhandledError" Dec 06 05:51:15 crc kubenswrapper[4809]: E1206 05:51:15.426813 4809 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 06 05:51:15 crc kubenswrapper[4809]: E1206 05:51:15.487610 4809 kubelet.go:2359] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Dec 06 05:51:15 crc kubenswrapper[4809]: E1206 05:51:15.527047 4809 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 06 05:51:15 crc kubenswrapper[4809]: E1206 05:51:15.528757 4809 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.154:6443: connect: connection refused" interval="400ms" Dec 06 05:51:15 crc kubenswrapper[4809]: E1206 05:51:15.627871 4809 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.671324 4809 policy_none.go:49] "None policy: Start" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.672505 4809 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.672572 4809 state_mem.go:35] "Initializing new in-memory state store" Dec 06 05:51:15 crc kubenswrapper[4809]: E1206 05:51:15.687861 4809 kubelet.go:2359] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Dec 06 05:51:15 crc kubenswrapper[4809]: E1206 05:51:15.728418 4809 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.824324 4809 manager.go:334] "Starting Device Plugin manager" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.824416 4809 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.824442 4809 server.go:79] "Starting device plugin registration server" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.825188 4809 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.825220 4809 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.825997 4809 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.826133 4809 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.826148 4809 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 06 05:51:15 crc kubenswrapper[4809]: E1206 05:51:15.831905 4809 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.926226 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.928664 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.928713 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.928724 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:15 crc kubenswrapper[4809]: I1206 05:51:15.928745 4809 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 05:51:15 crc kubenswrapper[4809]: E1206 05:51:15.929157 4809 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.154:6443: connect: connection refused" node="crc" Dec 06 05:51:15 crc kubenswrapper[4809]: E1206 05:51:15.929450 4809 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.154:6443: connect: connection refused" interval="800ms" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.088823 4809 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.088986 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.090456 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.090542 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.090580 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.090862 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.091268 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.091331 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.092332 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.092400 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.092434 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.092445 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.092473 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.092485 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.092671 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.092765 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.092796 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.093832 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.093893 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.093916 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.094027 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.094070 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.094093 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.094318 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.094375 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.094406 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.095251 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.095317 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.095341 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.095411 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.095425 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.095434 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.096370 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.096728 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.096794 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.098263 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.098292 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.098304 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.098268 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.098796 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.098822 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.098949 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.098995 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.099781 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.099846 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.099873 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.129656 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.131035 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.131084 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.131102 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.131134 4809 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 05:51:16 crc kubenswrapper[4809]: E1206 05:51:16.131713 4809 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.154:6443: connect: connection refused" node="crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.156058 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.156116 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.156145 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.156166 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.156185 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.156261 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.156309 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.156336 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.156401 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.156454 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.156494 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.156532 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.156586 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.156629 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.156647 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.258018 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.258075 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.258098 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.258120 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.258139 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.258160 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.258180 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.258185 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.258204 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.258213 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.258263 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.258268 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.258227 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.258307 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.258310 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.258318 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.258341 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.258294 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.258294 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.258355 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.258361 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.258343 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.258530 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.258596 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.258628 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.258654 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.258664 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.258693 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.258731 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.258764 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.325142 4809 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Dec 06 05:51:16 crc kubenswrapper[4809]: W1206 05:51:16.416015 4809 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Dec 06 05:51:16 crc kubenswrapper[4809]: E1206 05:51:16.416167 4809 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.154:6443: connect: connection refused" logger="UnhandledError" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.435848 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.442746 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.460532 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: W1206 05:51:16.465361 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-8dcdaa6dc8a3264763ab014cac60bd395f7ac58482dfc7a9c381868f061d4723 WatchSource:0}: Error finding container 8dcdaa6dc8a3264763ab014cac60bd395f7ac58482dfc7a9c381868f061d4723: Status 404 returned error can't find the container with id 8dcdaa6dc8a3264763ab014cac60bd395f7ac58482dfc7a9c381868f061d4723 Dec 06 05:51:16 crc kubenswrapper[4809]: W1206 05:51:16.469882 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-5186b3e919501ad93deb514d4660c3bd2c27a8aaeedbfb342e3e3b475e924210 WatchSource:0}: Error finding container 5186b3e919501ad93deb514d4660c3bd2c27a8aaeedbfb342e3e3b475e924210: Status 404 returned error can't find the container with id 5186b3e919501ad93deb514d4660c3bd2c27a8aaeedbfb342e3e3b475e924210 Dec 06 05:51:16 crc kubenswrapper[4809]: W1206 05:51:16.475554 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-42a4981b2d587ac03a691608a1442ba03aed6cdf37fd0061a68e5dc58fa4501f WatchSource:0}: Error finding container 42a4981b2d587ac03a691608a1442ba03aed6cdf37fd0061a68e5dc58fa4501f: Status 404 returned error can't find the container with id 42a4981b2d587ac03a691608a1442ba03aed6cdf37fd0061a68e5dc58fa4501f Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.482639 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.487464 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 05:51:16 crc kubenswrapper[4809]: W1206 05:51:16.497607 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-be566588c78ee2cd14d6c949a7067c32cfce050d7f1e7878220dea1be5963dc5 WatchSource:0}: Error finding container be566588c78ee2cd14d6c949a7067c32cfce050d7f1e7878220dea1be5963dc5: Status 404 returned error can't find the container with id be566588c78ee2cd14d6c949a7067c32cfce050d7f1e7878220dea1be5963dc5 Dec 06 05:51:16 crc kubenswrapper[4809]: W1206 05:51:16.509553 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-ac07917c0b8a09289e8a7265ae0afcc5c322602fc0f429e473d1d9d460f8d496 WatchSource:0}: Error finding container ac07917c0b8a09289e8a7265ae0afcc5c322602fc0f429e473d1d9d460f8d496: Status 404 returned error can't find the container with id ac07917c0b8a09289e8a7265ae0afcc5c322602fc0f429e473d1d9d460f8d496 Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.532351 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.534464 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.534500 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.534512 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:16 crc kubenswrapper[4809]: I1206 05:51:16.534541 4809 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 05:51:16 crc kubenswrapper[4809]: E1206 05:51:16.534950 4809 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.154:6443: connect: connection refused" node="crc" Dec 06 05:51:16 crc kubenswrapper[4809]: W1206 05:51:16.691053 4809 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Dec 06 05:51:16 crc kubenswrapper[4809]: E1206 05:51:16.691157 4809 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.154:6443: connect: connection refused" logger="UnhandledError" Dec 06 05:51:16 crc kubenswrapper[4809]: E1206 05:51:16.730587 4809 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.154:6443: connect: connection refused" interval="1.6s" Dec 06 05:51:16 crc kubenswrapper[4809]: W1206 05:51:16.748340 4809 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Dec 06 05:51:16 crc kubenswrapper[4809]: E1206 05:51:16.748423 4809 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.154:6443: connect: connection refused" logger="UnhandledError" Dec 06 05:51:16 crc kubenswrapper[4809]: W1206 05:51:16.872695 4809 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Dec 06 05:51:16 crc kubenswrapper[4809]: E1206 05:51:16.872806 4809 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.154:6443: connect: connection refused" logger="UnhandledError" Dec 06 05:51:17 crc kubenswrapper[4809]: I1206 05:51:17.324425 4809 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Dec 06 05:51:17 crc kubenswrapper[4809]: I1206 05:51:17.336007 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:51:17 crc kubenswrapper[4809]: I1206 05:51:17.337228 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:17 crc kubenswrapper[4809]: I1206 05:51:17.337269 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:17 crc kubenswrapper[4809]: I1206 05:51:17.337278 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:17 crc kubenswrapper[4809]: I1206 05:51:17.337305 4809 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 05:51:17 crc kubenswrapper[4809]: E1206 05:51:17.337742 4809 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.154:6443: connect: connection refused" node="crc" Dec 06 05:51:17 crc kubenswrapper[4809]: I1206 05:51:17.396356 4809 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="dc6965798de9aaf2b8066b7bd8abc23e12545f026bf1f0930dc0085b992553f2" exitCode=0 Dec 06 05:51:17 crc kubenswrapper[4809]: I1206 05:51:17.396458 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"dc6965798de9aaf2b8066b7bd8abc23e12545f026bf1f0930dc0085b992553f2"} Dec 06 05:51:17 crc kubenswrapper[4809]: I1206 05:51:17.396555 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ac07917c0b8a09289e8a7265ae0afcc5c322602fc0f429e473d1d9d460f8d496"} Dec 06 05:51:17 crc kubenswrapper[4809]: I1206 05:51:17.396663 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:51:17 crc kubenswrapper[4809]: I1206 05:51:17.397944 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:17 crc kubenswrapper[4809]: I1206 05:51:17.398005 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:17 crc kubenswrapper[4809]: I1206 05:51:17.398017 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:17 crc kubenswrapper[4809]: I1206 05:51:17.403724 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"1dfd977b049a8ff81192d4041adb265366ab88475ee84c5750305d061d21c8db"} Dec 06 05:51:17 crc kubenswrapper[4809]: I1206 05:51:17.403775 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"8fa59e4ec9867cf280b308ee3adb5e303eaf94455aca588f5eb74a2b6263d57c"} Dec 06 05:51:17 crc kubenswrapper[4809]: I1206 05:51:17.403792 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5720f7f2977d9af6ed5324a30d08d34a53d55460951b3875b7d7d753d202f349"} Dec 06 05:51:17 crc kubenswrapper[4809]: I1206 05:51:17.403805 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"be566588c78ee2cd14d6c949a7067c32cfce050d7f1e7878220dea1be5963dc5"} Dec 06 05:51:17 crc kubenswrapper[4809]: I1206 05:51:17.406887 4809 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7" exitCode=0 Dec 06 05:51:17 crc kubenswrapper[4809]: I1206 05:51:17.406971 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7"} Dec 06 05:51:17 crc kubenswrapper[4809]: I1206 05:51:17.407114 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"42a4981b2d587ac03a691608a1442ba03aed6cdf37fd0061a68e5dc58fa4501f"} Dec 06 05:51:17 crc kubenswrapper[4809]: I1206 05:51:17.407258 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:51:17 crc kubenswrapper[4809]: I1206 05:51:17.408286 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:17 crc kubenswrapper[4809]: I1206 05:51:17.408336 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:17 crc kubenswrapper[4809]: I1206 05:51:17.408353 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:17 crc kubenswrapper[4809]: I1206 05:51:17.409858 4809 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="9a37233831095a0ba4b488c136d3ede54ee833ab0d0141448f20ddb34fd4cc8a" exitCode=0 Dec 06 05:51:17 crc kubenswrapper[4809]: I1206 05:51:17.409904 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"9a37233831095a0ba4b488c136d3ede54ee833ab0d0141448f20ddb34fd4cc8a"} Dec 06 05:51:17 crc kubenswrapper[4809]: I1206 05:51:17.409982 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"5186b3e919501ad93deb514d4660c3bd2c27a8aaeedbfb342e3e3b475e924210"} Dec 06 05:51:17 crc kubenswrapper[4809]: I1206 05:51:17.410084 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:51:17 crc kubenswrapper[4809]: I1206 05:51:17.411323 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:17 crc kubenswrapper[4809]: I1206 05:51:17.411369 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:17 crc kubenswrapper[4809]: I1206 05:51:17.411385 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:17 crc kubenswrapper[4809]: I1206 05:51:17.412727 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:51:17 crc kubenswrapper[4809]: I1206 05:51:17.413950 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:17 crc kubenswrapper[4809]: I1206 05:51:17.413985 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:17 crc kubenswrapper[4809]: I1206 05:51:17.413999 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:17 crc kubenswrapper[4809]: I1206 05:51:17.418206 4809 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="75390f40b8b220311fda0a6203646ae89c1e9dcff4b74f350de6ac101998eca8" exitCode=0 Dec 06 05:51:17 crc kubenswrapper[4809]: I1206 05:51:17.418271 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"75390f40b8b220311fda0a6203646ae89c1e9dcff4b74f350de6ac101998eca8"} Dec 06 05:51:17 crc kubenswrapper[4809]: I1206 05:51:17.418314 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8dcdaa6dc8a3264763ab014cac60bd395f7ac58482dfc7a9c381868f061d4723"} Dec 06 05:51:17 crc kubenswrapper[4809]: I1206 05:51:17.418461 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:51:17 crc kubenswrapper[4809]: I1206 05:51:17.422530 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:17 crc kubenswrapper[4809]: I1206 05:51:17.422583 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:17 crc kubenswrapper[4809]: I1206 05:51:17.422595 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:18 crc kubenswrapper[4809]: I1206 05:51:18.423422 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"2ae336ca136d0a51627e63b1ec86bd4f916a5efaf804e5b91014188331c19802"} Dec 06 05:51:18 crc kubenswrapper[4809]: I1206 05:51:18.423466 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"2ddfff8cca35bda4796fe58ace60506f5c87e4398193d6711e81aae0d4b7d729"} Dec 06 05:51:18 crc kubenswrapper[4809]: I1206 05:51:18.423478 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"54d4b8272a98d4acd09f165939684c0990b9e56b1269f85c20d2bd41c0f75f67"} Dec 06 05:51:18 crc kubenswrapper[4809]: I1206 05:51:18.423557 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:51:18 crc kubenswrapper[4809]: I1206 05:51:18.424484 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:18 crc kubenswrapper[4809]: I1206 05:51:18.424506 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:18 crc kubenswrapper[4809]: I1206 05:51:18.424514 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:18 crc kubenswrapper[4809]: I1206 05:51:18.427692 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:51:18 crc kubenswrapper[4809]: I1206 05:51:18.427690 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"15d5801b61a87c95c9a623c3106e305920199c640efa8a32cf57698245697e54"} Dec 06 05:51:18 crc kubenswrapper[4809]: I1206 05:51:18.428387 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:18 crc kubenswrapper[4809]: I1206 05:51:18.428407 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:18 crc kubenswrapper[4809]: I1206 05:51:18.428416 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:18 crc kubenswrapper[4809]: I1206 05:51:18.430369 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad"} Dec 06 05:51:18 crc kubenswrapper[4809]: I1206 05:51:18.430394 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48"} Dec 06 05:51:18 crc kubenswrapper[4809]: I1206 05:51:18.430404 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f"} Dec 06 05:51:18 crc kubenswrapper[4809]: I1206 05:51:18.430412 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d"} Dec 06 05:51:18 crc kubenswrapper[4809]: I1206 05:51:18.431604 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"95eafea368aebd35fb8e9a20bb01938c0ee9b1109d1a4ac5c1062e0d4ce7465c"} Dec 06 05:51:18 crc kubenswrapper[4809]: I1206 05:51:18.431657 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:51:18 crc kubenswrapper[4809]: I1206 05:51:18.432288 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:18 crc kubenswrapper[4809]: I1206 05:51:18.432308 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:18 crc kubenswrapper[4809]: I1206 05:51:18.432316 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:18 crc kubenswrapper[4809]: I1206 05:51:18.433330 4809 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="fed3f2072ad9747d95c7002d8901c17e30476c03074a43d746392e65fff18927" exitCode=0 Dec 06 05:51:18 crc kubenswrapper[4809]: I1206 05:51:18.433460 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"fed3f2072ad9747d95c7002d8901c17e30476c03074a43d746392e65fff18927"} Dec 06 05:51:18 crc kubenswrapper[4809]: I1206 05:51:18.433570 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:51:18 crc kubenswrapper[4809]: I1206 05:51:18.434151 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:18 crc kubenswrapper[4809]: I1206 05:51:18.434190 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:18 crc kubenswrapper[4809]: I1206 05:51:18.434208 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:18 crc kubenswrapper[4809]: I1206 05:51:18.938180 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:51:18 crc kubenswrapper[4809]: I1206 05:51:18.939784 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:18 crc kubenswrapper[4809]: I1206 05:51:18.939833 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:18 crc kubenswrapper[4809]: I1206 05:51:18.939852 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:18 crc kubenswrapper[4809]: I1206 05:51:18.939883 4809 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 05:51:19 crc kubenswrapper[4809]: I1206 05:51:19.440255 4809 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="794b282ca73f1d2bbaa81b605f2f23d7790f3abdb5d9e2a3cc0ef7594fe97f65" exitCode=0 Dec 06 05:51:19 crc kubenswrapper[4809]: I1206 05:51:19.440330 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"794b282ca73f1d2bbaa81b605f2f23d7790f3abdb5d9e2a3cc0ef7594fe97f65"} Dec 06 05:51:19 crc kubenswrapper[4809]: I1206 05:51:19.440484 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:51:19 crc kubenswrapper[4809]: I1206 05:51:19.441334 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:19 crc kubenswrapper[4809]: I1206 05:51:19.441357 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:19 crc kubenswrapper[4809]: I1206 05:51:19.441368 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:19 crc kubenswrapper[4809]: I1206 05:51:19.447952 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:51:19 crc kubenswrapper[4809]: I1206 05:51:19.447915 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36"} Dec 06 05:51:19 crc kubenswrapper[4809]: I1206 05:51:19.448047 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:51:19 crc kubenswrapper[4809]: I1206 05:51:19.449235 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:19 crc kubenswrapper[4809]: I1206 05:51:19.449297 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:19 crc kubenswrapper[4809]: I1206 05:51:19.449316 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:19 crc kubenswrapper[4809]: I1206 05:51:19.449556 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:19 crc kubenswrapper[4809]: I1206 05:51:19.449607 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:19 crc kubenswrapper[4809]: I1206 05:51:19.449621 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:20 crc kubenswrapper[4809]: I1206 05:51:20.454803 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"64a802d6552da2644745f9514816997aeb4b2b35960628deaac866f2ed680a06"} Dec 06 05:51:20 crc kubenswrapper[4809]: I1206 05:51:20.454843 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1b3007b233fc349c6fade569e76fa97fa8f7cb311aeb2a9c557edf5df0cd45be"} Dec 06 05:51:20 crc kubenswrapper[4809]: I1206 05:51:20.454862 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"19d9efd6613f93b3a54c457dd5121c9504b4f8b5494ba3bfe6109cbaf057130d"} Dec 06 05:51:20 crc kubenswrapper[4809]: I1206 05:51:20.454872 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"05df6bc6f96907a3d1292071f938115d789d84399383f3fcee301f6def6f6f05"} Dec 06 05:51:20 crc kubenswrapper[4809]: I1206 05:51:20.454885 4809 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 05:51:20 crc kubenswrapper[4809]: I1206 05:51:20.454958 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:51:20 crc kubenswrapper[4809]: I1206 05:51:20.457154 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:20 crc kubenswrapper[4809]: I1206 05:51:20.457183 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:20 crc kubenswrapper[4809]: I1206 05:51:20.457193 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:20 crc kubenswrapper[4809]: I1206 05:51:20.548035 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 05:51:20 crc kubenswrapper[4809]: I1206 05:51:20.548156 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:51:20 crc kubenswrapper[4809]: I1206 05:51:20.555068 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 05:51:20 crc kubenswrapper[4809]: I1206 05:51:20.561568 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:20 crc kubenswrapper[4809]: I1206 05:51:20.561614 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:20 crc kubenswrapper[4809]: I1206 05:51:20.561626 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:20 crc kubenswrapper[4809]: I1206 05:51:20.703140 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 05:51:20 crc kubenswrapper[4809]: I1206 05:51:20.816771 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:51:21 crc kubenswrapper[4809]: I1206 05:51:21.464273 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"cedbde3de6a12172d93ed44ffadfe17e1c27f7987a808f1a7553ae70a91e5184"} Dec 06 05:51:21 crc kubenswrapper[4809]: I1206 05:51:21.464333 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:51:21 crc kubenswrapper[4809]: I1206 05:51:21.464316 4809 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 05:51:21 crc kubenswrapper[4809]: I1206 05:51:21.464378 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:51:21 crc kubenswrapper[4809]: I1206 05:51:21.464453 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:51:21 crc kubenswrapper[4809]: I1206 05:51:21.465414 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:21 crc kubenswrapper[4809]: I1206 05:51:21.465439 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:21 crc kubenswrapper[4809]: I1206 05:51:21.465450 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:21 crc kubenswrapper[4809]: I1206 05:51:21.465528 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:21 crc kubenswrapper[4809]: I1206 05:51:21.465570 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:21 crc kubenswrapper[4809]: I1206 05:51:21.465592 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:21 crc kubenswrapper[4809]: I1206 05:51:21.466598 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:21 crc kubenswrapper[4809]: I1206 05:51:21.466618 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:21 crc kubenswrapper[4809]: I1206 05:51:21.466626 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:22 crc kubenswrapper[4809]: I1206 05:51:22.083981 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:51:22 crc kubenswrapper[4809]: I1206 05:51:22.467058 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:51:22 crc kubenswrapper[4809]: I1206 05:51:22.467056 4809 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 05:51:22 crc kubenswrapper[4809]: I1206 05:51:22.467159 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:51:22 crc kubenswrapper[4809]: I1206 05:51:22.467229 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:51:22 crc kubenswrapper[4809]: I1206 05:51:22.470740 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:22 crc kubenswrapper[4809]: I1206 05:51:22.470794 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:22 crc kubenswrapper[4809]: I1206 05:51:22.470842 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:22 crc kubenswrapper[4809]: I1206 05:51:22.471024 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:22 crc kubenswrapper[4809]: I1206 05:51:22.471102 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:22 crc kubenswrapper[4809]: I1206 05:51:22.471135 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:22 crc kubenswrapper[4809]: I1206 05:51:22.471377 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:22 crc kubenswrapper[4809]: I1206 05:51:22.471502 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:22 crc kubenswrapper[4809]: I1206 05:51:22.471535 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:23 crc kubenswrapper[4809]: I1206 05:51:23.703750 4809 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 05:51:23 crc kubenswrapper[4809]: I1206 05:51:23.704711 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 05:51:23 crc kubenswrapper[4809]: I1206 05:51:23.951216 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:51:23 crc kubenswrapper[4809]: I1206 05:51:23.951452 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:51:23 crc kubenswrapper[4809]: I1206 05:51:23.952748 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:23 crc kubenswrapper[4809]: I1206 05:51:23.952772 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:23 crc kubenswrapper[4809]: I1206 05:51:23.952780 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:24 crc kubenswrapper[4809]: I1206 05:51:24.268592 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 05:51:24 crc kubenswrapper[4809]: I1206 05:51:24.268724 4809 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 05:51:24 crc kubenswrapper[4809]: I1206 05:51:24.268764 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:51:24 crc kubenswrapper[4809]: I1206 05:51:24.269889 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:24 crc kubenswrapper[4809]: I1206 05:51:24.269914 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:24 crc kubenswrapper[4809]: I1206 05:51:24.269921 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:24 crc kubenswrapper[4809]: I1206 05:51:24.416853 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 05:51:24 crc kubenswrapper[4809]: I1206 05:51:24.417071 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:51:24 crc kubenswrapper[4809]: I1206 05:51:24.418501 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:24 crc kubenswrapper[4809]: I1206 05:51:24.418853 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:24 crc kubenswrapper[4809]: I1206 05:51:24.419065 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:24 crc kubenswrapper[4809]: I1206 05:51:24.560726 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 05:51:24 crc kubenswrapper[4809]: I1206 05:51:24.561289 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:51:24 crc kubenswrapper[4809]: I1206 05:51:24.562426 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:24 crc kubenswrapper[4809]: I1206 05:51:24.562459 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:24 crc kubenswrapper[4809]: I1206 05:51:24.562471 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:25 crc kubenswrapper[4809]: I1206 05:51:25.194495 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 06 05:51:25 crc kubenswrapper[4809]: I1206 05:51:25.194740 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:51:25 crc kubenswrapper[4809]: I1206 05:51:25.196108 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:25 crc kubenswrapper[4809]: I1206 05:51:25.196150 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:25 crc kubenswrapper[4809]: I1206 05:51:25.196172 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:25 crc kubenswrapper[4809]: E1206 05:51:25.832063 4809 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 06 05:51:28 crc kubenswrapper[4809]: W1206 05:51:28.106297 4809 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 06 05:51:28 crc kubenswrapper[4809]: I1206 05:51:28.106403 4809 trace.go:236] Trace[1584291393]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (06-Dec-2025 05:51:18.104) (total time: 10001ms): Dec 06 05:51:28 crc kubenswrapper[4809]: Trace[1584291393]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (05:51:28.106) Dec 06 05:51:28 crc kubenswrapper[4809]: Trace[1584291393]: [10.001856811s] [10.001856811s] END Dec 06 05:51:28 crc kubenswrapper[4809]: E1206 05:51:28.106426 4809 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 06 05:51:28 crc kubenswrapper[4809]: I1206 05:51:28.325329 4809 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 06 05:51:28 crc kubenswrapper[4809]: E1206 05:51:28.331688 4809 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" interval="3.2s" Dec 06 05:51:28 crc kubenswrapper[4809]: E1206 05:51:28.681417 4809 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": net/http: TLS handshake timeout" event="&Event{ObjectMeta:{crc.187e8a66271b10d6 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-06 05:51:15.323195606 +0000 UTC m=+0.212178548,LastTimestamp:2025-12-06 05:51:15.323195606 +0000 UTC m=+0.212178548,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 06 05:51:28 crc kubenswrapper[4809]: W1206 05:51:28.876716 4809 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 06 05:51:28 crc kubenswrapper[4809]: I1206 05:51:28.876790 4809 trace.go:236] Trace[1879683336]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (06-Dec-2025 05:51:18.874) (total time: 10001ms): Dec 06 05:51:28 crc kubenswrapper[4809]: Trace[1879683336]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (05:51:28.876) Dec 06 05:51:28 crc kubenswrapper[4809]: Trace[1879683336]: [10.001947063s] [10.001947063s] END Dec 06 05:51:28 crc kubenswrapper[4809]: E1206 05:51:28.876816 4809 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 06 05:51:28 crc kubenswrapper[4809]: W1206 05:51:28.933577 4809 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 06 05:51:28 crc kubenswrapper[4809]: I1206 05:51:28.933661 4809 trace.go:236] Trace[967669419]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (06-Dec-2025 05:51:18.932) (total time: 10001ms): Dec 06 05:51:28 crc kubenswrapper[4809]: Trace[967669419]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (05:51:28.933) Dec 06 05:51:28 crc kubenswrapper[4809]: Trace[967669419]: [10.001283448s] [10.001283448s] END Dec 06 05:51:28 crc kubenswrapper[4809]: E1206 05:51:28.933683 4809 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 06 05:51:28 crc kubenswrapper[4809]: E1206 05:51:28.941129 4809 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="crc" Dec 06 05:51:29 crc kubenswrapper[4809]: I1206 05:51:29.558518 4809 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 06 05:51:29 crc kubenswrapper[4809]: I1206 05:51:29.558587 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 06 05:51:29 crc kubenswrapper[4809]: I1206 05:51:29.563008 4809 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 06 05:51:29 crc kubenswrapper[4809]: I1206 05:51:29.563050 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 06 05:51:30 crc kubenswrapper[4809]: I1206 05:51:30.795116 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 06 05:51:30 crc kubenswrapper[4809]: I1206 05:51:30.795290 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:51:30 crc kubenswrapper[4809]: I1206 05:51:30.796665 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:30 crc kubenswrapper[4809]: I1206 05:51:30.796714 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:30 crc kubenswrapper[4809]: I1206 05:51:30.796726 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:30 crc kubenswrapper[4809]: I1206 05:51:30.819882 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 06 05:51:31 crc kubenswrapper[4809]: I1206 05:51:31.494272 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:51:31 crc kubenswrapper[4809]: I1206 05:51:31.495308 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:31 crc kubenswrapper[4809]: I1206 05:51:31.495341 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:31 crc kubenswrapper[4809]: I1206 05:51:31.495353 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:31 crc kubenswrapper[4809]: I1206 05:51:31.511276 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 06 05:51:31 crc kubenswrapper[4809]: I1206 05:51:31.542151 4809 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 06 05:51:32 crc kubenswrapper[4809]: I1206 05:51:32.141703 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:51:32 crc kubenswrapper[4809]: I1206 05:51:32.142999 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:32 crc kubenswrapper[4809]: I1206 05:51:32.143046 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:32 crc kubenswrapper[4809]: I1206 05:51:32.143062 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:32 crc kubenswrapper[4809]: I1206 05:51:32.143090 4809 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 05:51:32 crc kubenswrapper[4809]: E1206 05:51:32.146798 4809 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 06 05:51:32 crc kubenswrapper[4809]: I1206 05:51:32.496488 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:51:32 crc kubenswrapper[4809]: I1206 05:51:32.497210 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:32 crc kubenswrapper[4809]: I1206 05:51:32.497234 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:32 crc kubenswrapper[4809]: I1206 05:51:32.497243 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:33 crc kubenswrapper[4809]: I1206 05:51:33.704024 4809 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 05:51:33 crc kubenswrapper[4809]: I1206 05:51:33.704100 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 05:51:33 crc kubenswrapper[4809]: I1206 05:51:33.954716 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:51:33 crc kubenswrapper[4809]: I1206 05:51:33.955054 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:51:33 crc kubenswrapper[4809]: I1206 05:51:33.958193 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:33 crc kubenswrapper[4809]: I1206 05:51:33.958242 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:33 crc kubenswrapper[4809]: I1206 05:51:33.958257 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:33 crc kubenswrapper[4809]: I1206 05:51:33.961644 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:51:34 crc kubenswrapper[4809]: I1206 05:51:34.442662 4809 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 06 05:51:34 crc kubenswrapper[4809]: I1206 05:51:34.502387 4809 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 05:51:34 crc kubenswrapper[4809]: I1206 05:51:34.557594 4809 trace.go:236] Trace[802396992]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (06-Dec-2025 05:51:19.793) (total time: 14763ms): Dec 06 05:51:34 crc kubenswrapper[4809]: Trace[802396992]: ---"Objects listed" error: 14763ms (05:51:34.557) Dec 06 05:51:34 crc kubenswrapper[4809]: Trace[802396992]: [14.763538681s] [14.763538681s] END Dec 06 05:51:34 crc kubenswrapper[4809]: I1206 05:51:34.557625 4809 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 06 05:51:34 crc kubenswrapper[4809]: I1206 05:51:34.558846 4809 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 06 05:51:34 crc kubenswrapper[4809]: I1206 05:51:34.566898 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 05:51:34 crc kubenswrapper[4809]: I1206 05:51:34.568885 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:51:34 crc kubenswrapper[4809]: I1206 05:51:34.785811 4809 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.327730 4809 apiserver.go:52] "Watching apiserver" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.334297 4809 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.335538 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-cb4vc","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-dns/node-resolver-rnzq4","openshift-kube-apiserver/kube-apiserver-crc","openshift-machine-config-operator/machine-config-daemon-npms2","openshift-network-operator/iptables-alerter-4ln5h","openshift-ovn-kubernetes/ovnkube-node-5xvvb","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-multus/multus-additional-cni-plugins-9t7pq","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"] Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.336639 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.337767 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:51:35 crc kubenswrapper[4809]: E1206 05:51:35.337857 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.338229 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.338408 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.338727 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:51:35 crc kubenswrapper[4809]: E1206 05:51:35.338808 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.338736 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.338983 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-npms2" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.339095 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 05:51:35 crc kubenswrapper[4809]: E1206 05:51:35.339407 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.339861 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.341149 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-rnzq4" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.342331 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.350569 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.350955 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.351096 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.351701 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.354341 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.354562 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.354422 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.354832 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.355029 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.355106 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.355401 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.355636 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.355721 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.355913 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.355965 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.356087 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.356106 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.356180 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.356269 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.356314 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.356357 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.356414 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.356442 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.356537 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.356632 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.356825 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.357071 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.357280 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.357380 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.357341 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.357515 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.376694 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5636dd78-3d61-4c6a-b8fd-6d2457e19234\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.397882 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.414487 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.424593 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.428044 4809 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.434023 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.441969 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.451155 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.460744 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-npms2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.463568 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.463610 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.463632 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.463651 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.463670 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.463687 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.463703 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.463718 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.463738 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.463763 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.463788 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.463804 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.463823 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.463844 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.463882 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.463907 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.463945 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.463971 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.463992 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.464013 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.464039 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.464068 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.464094 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.464116 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.464142 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.464171 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.464201 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.464228 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.464259 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.464289 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.464317 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.464344 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.464371 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.464399 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.464473 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.464503 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.464535 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.464562 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.464591 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.464619 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.464664 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.464699 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.464725 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.464751 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.464778 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.464808 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.464836 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.464866 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.464895 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.464975 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465010 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465036 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465068 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465100 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465135 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465161 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465231 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465259 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465286 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465312 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465343 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465373 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465394 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465418 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465442 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465461 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465485 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465505 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465532 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465552 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465589 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465606 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465627 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465647 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465671 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465691 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465715 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465734 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465758 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465782 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465800 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465820 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465841 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465860 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465881 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465908 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465945 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465967 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465987 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466009 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466036 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466056 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466079 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466099 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466117 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466138 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466161 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466188 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466212 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466237 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466281 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466300 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466320 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466338 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466360 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466380 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466402 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466421 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466439 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466461 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466481 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466570 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466591 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466611 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466630 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466650 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466671 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466691 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466709 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466727 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466748 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466766 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466784 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466803 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466821 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466840 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466862 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466883 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466902 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466921 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466957 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466978 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466999 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467021 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467041 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467060 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467086 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467104 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467125 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467145 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467187 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467208 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467229 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467250 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467268 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467289 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467310 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467331 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467355 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467374 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467394 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467414 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467433 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467451 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467470 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467489 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467507 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467528 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467546 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467566 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467584 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467604 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467622 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467640 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467660 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467680 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467699 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467719 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467739 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467757 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467776 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467795 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467813 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467923 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467964 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467992 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.468010 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.468028 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.468047 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.468065 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.464469 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.464643 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465124 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465340 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465413 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465458 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465482 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465503 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465608 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465724 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.468507 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465733 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.465745 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466366 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466439 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466555 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466691 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466703 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466722 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466840 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.466984 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.468640 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467116 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467164 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467285 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467348 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467498 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467764 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.468740 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.467831 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.468039 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.468287 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.468310 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.468851 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.469027 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.469042 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.469097 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.469224 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.469384 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.469709 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.469754 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.469824 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.469860 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.469879 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.469901 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.469922 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.469956 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.469974 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.469995 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.470016 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.470034 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.470058 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.470080 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.470101 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.470121 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.470142 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.470245 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.470493 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-host-run-netns\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.470520 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-env-overrides\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.470571 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/e0075197-9392-4109-b8a7-39507b15cc17-hosts-file\") pod \"node-resolver-rnzq4\" (UID: \"e0075197-9392-4109-b8a7-39507b15cc17\") " pod="openshift-dns/node-resolver-rnzq4" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.470589 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a48de155-7389-48db-8f86-d175b7db70c8-tuning-conf-dir\") pod \"multus-additional-cni-plugins-9t7pq\" (UID: \"a48de155-7389-48db-8f86-d175b7db70c8\") " pod="openshift-multus/multus-additional-cni-plugins-9t7pq" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.470609 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-host-slash\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.470627 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-run-openvswitch\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.470644 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-host-cni-bin\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.470673 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.470691 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.470790 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.471398 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.471592 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.471822 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.471707 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.472471 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.472649 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.472700 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.473029 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.473242 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.473405 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.473558 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.473717 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.473999 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.469613 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.470705 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-var-lib-openvswitch\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.475479 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-ovnkube-config\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.475516 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.475564 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mqln\" (UniqueName: \"kubernetes.io/projected/cbf4cf62-024e-4703-a8b8-9aecda9cd26a-kube-api-access-4mqln\") pod \"machine-config-daemon-npms2\" (UID: \"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\") " pod="openshift-machine-config-operator/machine-config-daemon-npms2" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.475584 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/ed42d686-41af-470e-910d-a1fabbec66b0-host-run-k8s-cni-cncf-io\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.475603 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ed42d686-41af-470e-910d-a1fabbec66b0-etc-kubernetes\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.475622 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-etc-openvswitch\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.475639 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-ovnkube-script-lib\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.475662 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ed42d686-41af-470e-910d-a1fabbec66b0-host-run-netns\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.475664 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.475674 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.475695 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.475742 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.476353 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5636dd78-3d61-4c6a-b8fd-6d2457e19234\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.477045 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.477161 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.477187 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.477246 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.477465 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.477581 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.477607 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.477584 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.478110 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.477985 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.478363 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.478715 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.478867 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.479120 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.479365 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.479589 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.479597 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.480134 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.480251 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.480307 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.480499 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.480473 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.480842 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.480854 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.480915 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.481023 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.481463 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.481810 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.482161 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.482431 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.482574 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.483108 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.483147 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.484567 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.484298 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.485730 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.485756 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.486062 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.486086 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.486281 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.486346 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.486325 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.486696 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.486835 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.487801 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.487864 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.487903 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.488169 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.488258 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.488199 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.488409 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.488430 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: E1206 05:51:35.490238 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:51:35.990214428 +0000 UTC m=+20.879197370 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.495352 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.496085 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.496245 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.496527 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rnzq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0075197-9392-4109-b8a7-39507b15cc17\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4mgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rnzq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.497014 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.497054 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.497315 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.497419 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.497445 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.497547 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.497780 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.497866 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.498030 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.475683 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/ed42d686-41af-470e-910d-a1fabbec66b0-host-var-lib-cni-multus\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.499615 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.499827 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-host-cni-netd\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.499914 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/ed42d686-41af-470e-910d-a1fabbec66b0-hostroot\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.500011 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nwpx\" (UniqueName: \"kubernetes.io/projected/ed42d686-41af-470e-910d-a1fabbec66b0-kube-api-access-2nwpx\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.500094 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.500162 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-host-kubelet\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.500231 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-systemd-units\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.500345 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.500357 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.500457 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a48de155-7389-48db-8f86-d175b7db70c8-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-9t7pq\" (UID: \"a48de155-7389-48db-8f86-d175b7db70c8\") " pod="openshift-multus/multus-additional-cni-plugins-9t7pq" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.500506 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/cbf4cf62-024e-4703-a8b8-9aecda9cd26a-proxy-tls\") pod \"machine-config-daemon-npms2\" (UID: \"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\") " pod="openshift-machine-config-operator/machine-config-daemon-npms2" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.498195 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.500538 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ed42d686-41af-470e-910d-a1fabbec66b0-multus-conf-dir\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.498280 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.498687 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.498862 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.500578 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.498907 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.500608 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-log-socket\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.500598 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.499162 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.500644 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-227vc\" (UniqueName: \"kubernetes.io/projected/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-kube-api-access-227vc\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.499420 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.499649 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.500046 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.500475 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.500622 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.500742 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cbf4cf62-024e-4703-a8b8-9aecda9cd26a-mcd-auth-proxy-config\") pod \"machine-config-daemon-npms2\" (UID: \"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\") " pod="openshift-machine-config-operator/machine-config-daemon-npms2" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.500694 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.500802 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ed42d686-41af-470e-910d-a1fabbec66b0-multus-cni-dir\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.500859 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.500897 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ed42d686-41af-470e-910d-a1fabbec66b0-cnibin\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.500997 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a48de155-7389-48db-8f86-d175b7db70c8-system-cni-dir\") pod \"multus-additional-cni-plugins-9t7pq\" (UID: \"a48de155-7389-48db-8f86-d175b7db70c8\") " pod="openshift-multus/multus-additional-cni-plugins-9t7pq" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.501061 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.501096 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-run-systemd\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.501116 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ed42d686-41af-470e-910d-a1fabbec66b0-os-release\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.501138 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.501160 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a48de155-7389-48db-8f86-d175b7db70c8-cnibin\") pod \"multus-additional-cni-plugins-9t7pq\" (UID: \"a48de155-7389-48db-8f86-d175b7db70c8\") " pod="openshift-multus/multus-additional-cni-plugins-9t7pq" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.501182 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-node-log\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.501232 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-ovn-node-metrics-cert\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.501290 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:51:35 crc kubenswrapper[4809]: E1206 05:51:35.501346 4809 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.501333 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/cbf4cf62-024e-4703-a8b8-9aecda9cd26a-rootfs\") pod \"machine-config-daemon-npms2\" (UID: \"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\") " pod="openshift-machine-config-operator/machine-config-daemon-npms2" Dec 06 05:51:35 crc kubenswrapper[4809]: E1206 05:51:35.501426 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 05:51:36.001401267 +0000 UTC m=+20.890384209 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.501459 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.501481 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-run-ovn\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.501506 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.501533 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.501552 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.501569 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/ed42d686-41af-470e-910d-a1fabbec66b0-host-var-lib-kubelet\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.501588 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.501608 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4z265\" (UniqueName: \"kubernetes.io/projected/a48de155-7389-48db-8f86-d175b7db70c8-kube-api-access-4z265\") pod \"multus-additional-cni-plugins-9t7pq\" (UID: \"a48de155-7389-48db-8f86-d175b7db70c8\") " pod="openshift-multus/multus-additional-cni-plugins-9t7pq" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.501633 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4mgv\" (UniqueName: \"kubernetes.io/projected/e0075197-9392-4109-b8a7-39507b15cc17-kube-api-access-q4mgv\") pod \"node-resolver-rnzq4\" (UID: \"e0075197-9392-4109-b8a7-39507b15cc17\") " pod="openshift-dns/node-resolver-rnzq4" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.501635 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.501656 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a48de155-7389-48db-8f86-d175b7db70c8-cni-binary-copy\") pod \"multus-additional-cni-plugins-9t7pq\" (UID: \"a48de155-7389-48db-8f86-d175b7db70c8\") " pod="openshift-multus/multus-additional-cni-plugins-9t7pq" Dec 06 05:51:35 crc kubenswrapper[4809]: E1206 05:51:35.501660 4809 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.501706 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-host-run-ovn-kubernetes\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.501727 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ed42d686-41af-470e-910d-a1fabbec66b0-system-cni-dir\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: E1206 05:51:35.501764 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 05:51:36.001730284 +0000 UTC m=+20.890713226 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.501795 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ed42d686-41af-470e-910d-a1fabbec66b0-cni-binary-copy\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.501820 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/ed42d686-41af-470e-910d-a1fabbec66b0-multus-daemon-config\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.501840 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/ed42d686-41af-470e-910d-a1fabbec66b0-host-run-multus-certs\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.501864 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.501883 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/ed42d686-41af-470e-910d-a1fabbec66b0-multus-socket-dir-parent\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.501901 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ed42d686-41af-470e-910d-a1fabbec66b0-host-var-lib-cni-bin\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.501919 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a48de155-7389-48db-8f86-d175b7db70c8-os-release\") pod \"multus-additional-cni-plugins-9t7pq\" (UID: \"a48de155-7389-48db-8f86-d175b7db70c8\") " pod="openshift-multus/multus-additional-cni-plugins-9t7pq" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.502069 4809 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.502088 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.502100 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.502113 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.502123 4809 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.502133 4809 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.502143 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.502153 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.502165 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.502175 4809 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.502184 4809 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.502195 4809 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.502205 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.502214 4809 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.502223 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.502233 4809 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.502241 4809 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.502251 4809 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.502261 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.502274 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.502334 4809 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.502348 4809 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.502802 4809 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.502841 4809 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.502853 4809 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.502865 4809 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.502877 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.502889 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.502900 4809 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.502910 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.502921 4809 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.502962 4809 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.502974 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.502958 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.502986 4809 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.503116 4809 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.503132 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.503132 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.503147 4809 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: E1206 05:51:35.503353 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 05:51:35 crc kubenswrapper[4809]: E1206 05:51:35.503384 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 05:51:35 crc kubenswrapper[4809]: E1206 05:51:35.503402 4809 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:51:35 crc kubenswrapper[4809]: E1206 05:51:35.503465 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 05:51:36.003455275 +0000 UTC m=+20.892438407 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.503590 4809 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504145 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504226 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504240 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504258 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504269 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504279 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504290 4809 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504302 4809 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504312 4809 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504324 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504336 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504355 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504366 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504377 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504387 4809 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504397 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504408 4809 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504418 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504428 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504440 4809 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504450 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504466 4809 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504477 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504489 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504500 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504510 4809 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504522 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504531 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504540 4809 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504551 4809 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504561 4809 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504571 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504580 4809 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504590 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504600 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504611 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504622 4809 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504632 4809 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504642 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504652 4809 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504661 4809 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504670 4809 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504683 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504692 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504702 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504710 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504719 4809 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504729 4809 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504738 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504746 4809 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504755 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504764 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504774 4809 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504786 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504795 4809 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504804 4809 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504813 4809 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504821 4809 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504831 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504840 4809 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504850 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504860 4809 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504868 4809 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504877 4809 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504886 4809 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504895 4809 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504906 4809 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504915 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504924 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504956 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504969 4809 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504983 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.504995 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.505004 4809 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.505013 4809 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.505024 4809 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.505036 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.505049 4809 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.505062 4809 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.505074 4809 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.505097 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.505108 4809 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.505116 4809 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.505126 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.505136 4809 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.505145 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.505154 4809 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.505164 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.505173 4809 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.505183 4809 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.505192 4809 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.505202 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.505213 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.505227 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.505237 4809 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.505246 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.505255 4809 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.505264 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.505274 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.505283 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.505296 4809 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.505305 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.505315 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.505627 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.505992 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.507176 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.507442 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.507494 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.507318 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.507612 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.507946 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.509035 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.519665 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.520299 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.520400 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.520816 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.521055 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.521265 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.521312 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.521451 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.521687 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.521920 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.522084 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.522094 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.522349 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.522514 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.522946 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.523236 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.524557 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.524730 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.525366 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.525834 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: E1206 05:51:35.526125 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 05:51:35 crc kubenswrapper[4809]: E1206 05:51:35.526165 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 05:51:35 crc kubenswrapper[4809]: E1206 05:51:35.526187 4809 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:51:35 crc kubenswrapper[4809]: E1206 05:51:35.526266 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 05:51:36.026235412 +0000 UTC m=+20.915218354 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:51:35 crc kubenswrapper[4809]: E1206 05:51:35.526298 4809 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-apiserver-crc\" already exists" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.526406 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.526725 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.526742 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.526921 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.527113 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.527426 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.527448 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.527477 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.527542 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.527576 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.527630 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.528064 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.528202 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.528627 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.530922 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a48de155-7389-48db-8f86-d175b7db70c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9t7pq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.536431 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.540386 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.541061 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.542253 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.542691 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.543067 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.543255 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.544985 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.545225 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.548582 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.549150 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.551353 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.555651 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cb4vc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed42d686-41af-470e-910d-a1fabbec66b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2nwpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cb4vc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.561203 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.566218 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa4dd3f8-c119-46de-ad4e-dd29bb11a32c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa59e4ec9867cf280b308ee3adb5e303eaf94455aca588f5eb74a2b6263d57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5720f7f2977d9af6ed5324a30d08d34a53d55460951b3875b7d7d753d202f349\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dfd977b049a8ff81192d4041adb265366ab88475ee84c5750305d061d21c8db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d5801b61a87c95c9a623c3106e305920199c640efa8a32cf57698245697e54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.567888 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.573683 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.577358 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.594401 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5xvvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.604460 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cb4vc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed42d686-41af-470e-910d-a1fabbec66b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2nwpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cb4vc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.609799 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-run-ovn\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.609851 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.609870 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.609886 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/ed42d686-41af-470e-910d-a1fabbec66b0-host-var-lib-kubelet\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.609912 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4z265\" (UniqueName: \"kubernetes.io/projected/a48de155-7389-48db-8f86-d175b7db70c8-kube-api-access-4z265\") pod \"multus-additional-cni-plugins-9t7pq\" (UID: \"a48de155-7389-48db-8f86-d175b7db70c8\") " pod="openshift-multus/multus-additional-cni-plugins-9t7pq" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.609953 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4mgv\" (UniqueName: \"kubernetes.io/projected/e0075197-9392-4109-b8a7-39507b15cc17-kube-api-access-q4mgv\") pod \"node-resolver-rnzq4\" (UID: \"e0075197-9392-4109-b8a7-39507b15cc17\") " pod="openshift-dns/node-resolver-rnzq4" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.609971 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a48de155-7389-48db-8f86-d175b7db70c8-cni-binary-copy\") pod \"multus-additional-cni-plugins-9t7pq\" (UID: \"a48de155-7389-48db-8f86-d175b7db70c8\") " pod="openshift-multus/multus-additional-cni-plugins-9t7pq" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.609989 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-host-run-ovn-kubernetes\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.610004 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ed42d686-41af-470e-910d-a1fabbec66b0-system-cni-dir\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.610018 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ed42d686-41af-470e-910d-a1fabbec66b0-cni-binary-copy\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.610033 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/ed42d686-41af-470e-910d-a1fabbec66b0-multus-daemon-config\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.610049 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/ed42d686-41af-470e-910d-a1fabbec66b0-host-run-multus-certs\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.610068 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.610085 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/ed42d686-41af-470e-910d-a1fabbec66b0-multus-socket-dir-parent\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.610140 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ed42d686-41af-470e-910d-a1fabbec66b0-host-var-lib-cni-bin\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.610155 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a48de155-7389-48db-8f86-d175b7db70c8-os-release\") pod \"multus-additional-cni-plugins-9t7pq\" (UID: \"a48de155-7389-48db-8f86-d175b7db70c8\") " pod="openshift-multus/multus-additional-cni-plugins-9t7pq" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.610171 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-host-run-netns\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.610188 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-env-overrides\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.610205 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/e0075197-9392-4109-b8a7-39507b15cc17-hosts-file\") pod \"node-resolver-rnzq4\" (UID: \"e0075197-9392-4109-b8a7-39507b15cc17\") " pod="openshift-dns/node-resolver-rnzq4" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.610221 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a48de155-7389-48db-8f86-d175b7db70c8-tuning-conf-dir\") pod \"multus-additional-cni-plugins-9t7pq\" (UID: \"a48de155-7389-48db-8f86-d175b7db70c8\") " pod="openshift-multus/multus-additional-cni-plugins-9t7pq" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.610235 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-host-slash\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.610249 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-run-openvswitch\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.610263 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-host-cni-bin\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.610282 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-var-lib-openvswitch\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.610301 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-ovnkube-config\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.610324 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mqln\" (UniqueName: \"kubernetes.io/projected/cbf4cf62-024e-4703-a8b8-9aecda9cd26a-kube-api-access-4mqln\") pod \"machine-config-daemon-npms2\" (UID: \"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\") " pod="openshift-machine-config-operator/machine-config-daemon-npms2" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.610339 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/ed42d686-41af-470e-910d-a1fabbec66b0-host-run-k8s-cni-cncf-io\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.610361 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ed42d686-41af-470e-910d-a1fabbec66b0-etc-kubernetes\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.610376 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-etc-openvswitch\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.610391 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-ovnkube-script-lib\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.610407 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ed42d686-41af-470e-910d-a1fabbec66b0-host-run-netns\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.610423 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/ed42d686-41af-470e-910d-a1fabbec66b0-host-var-lib-cni-multus\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.610438 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-host-cni-netd\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.610452 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/ed42d686-41af-470e-910d-a1fabbec66b0-hostroot\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.610465 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nwpx\" (UniqueName: \"kubernetes.io/projected/ed42d686-41af-470e-910d-a1fabbec66b0-kube-api-access-2nwpx\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.610481 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-host-kubelet\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.610495 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-systemd-units\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.610511 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a48de155-7389-48db-8f86-d175b7db70c8-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-9t7pq\" (UID: \"a48de155-7389-48db-8f86-d175b7db70c8\") " pod="openshift-multus/multus-additional-cni-plugins-9t7pq" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.610531 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/cbf4cf62-024e-4703-a8b8-9aecda9cd26a-proxy-tls\") pod \"machine-config-daemon-npms2\" (UID: \"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\") " pod="openshift-machine-config-operator/machine-config-daemon-npms2" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.610581 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ed42d686-41af-470e-910d-a1fabbec66b0-multus-conf-dir\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.610604 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-log-socket\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.610624 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-227vc\" (UniqueName: \"kubernetes.io/projected/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-kube-api-access-227vc\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.610644 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cbf4cf62-024e-4703-a8b8-9aecda9cd26a-mcd-auth-proxy-config\") pod \"machine-config-daemon-npms2\" (UID: \"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\") " pod="openshift-machine-config-operator/machine-config-daemon-npms2" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.610662 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ed42d686-41af-470e-910d-a1fabbec66b0-multus-cni-dir\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.610681 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ed42d686-41af-470e-910d-a1fabbec66b0-cnibin\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.610700 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a48de155-7389-48db-8f86-d175b7db70c8-system-cni-dir\") pod \"multus-additional-cni-plugins-9t7pq\" (UID: \"a48de155-7389-48db-8f86-d175b7db70c8\") " pod="openshift-multus/multus-additional-cni-plugins-9t7pq" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.610721 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-run-systemd\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.610714 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/e0075197-9392-4109-b8a7-39507b15cc17-hosts-file\") pod \"node-resolver-rnzq4\" (UID: \"e0075197-9392-4109-b8a7-39507b15cc17\") " pod="openshift-dns/node-resolver-rnzq4" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.610825 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ed42d686-41af-470e-910d-a1fabbec66b0-os-release\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.610876 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-run-ovn\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.611050 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-host-slash\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.611097 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-run-openvswitch\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.611169 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-host-cni-bin\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.611198 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-var-lib-openvswitch\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.611622 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-host-kubelet\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.611663 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ed42d686-41af-470e-910d-a1fabbec66b0-cni-binary-copy\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.611732 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.611734 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-host-cni-netd\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.611769 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.611798 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/ed42d686-41af-470e-910d-a1fabbec66b0-host-var-lib-kubelet\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.611807 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ed42d686-41af-470e-910d-a1fabbec66b0-host-run-netns\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.611857 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/ed42d686-41af-470e-910d-a1fabbec66b0-host-var-lib-cni-multus\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.612074 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-ovnkube-config\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.612145 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a48de155-7389-48db-8f86-d175b7db70c8-tuning-conf-dir\") pod \"multus-additional-cni-plugins-9t7pq\" (UID: \"a48de155-7389-48db-8f86-d175b7db70c8\") " pod="openshift-multus/multus-additional-cni-plugins-9t7pq" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.612203 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-host-run-ovn-kubernetes\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.612551 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cbf4cf62-024e-4703-a8b8-9aecda9cd26a-mcd-auth-proxy-config\") pod \"machine-config-daemon-npms2\" (UID: \"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\") " pod="openshift-machine-config-operator/machine-config-daemon-npms2" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.612704 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ed42d686-41af-470e-910d-a1fabbec66b0-system-cni-dir\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.612835 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ed42d686-41af-470e-910d-a1fabbec66b0-host-var-lib-cni-bin\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.612912 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ed42d686-41af-470e-910d-a1fabbec66b0-multus-cni-dir\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.612992 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ed42d686-41af-470e-910d-a1fabbec66b0-cnibin\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.613023 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a48de155-7389-48db-8f86-d175b7db70c8-system-cni-dir\") pod \"multus-additional-cni-plugins-9t7pq\" (UID: \"a48de155-7389-48db-8f86-d175b7db70c8\") " pod="openshift-multus/multus-additional-cni-plugins-9t7pq" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.613047 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-run-systemd\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.613086 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/ed42d686-41af-470e-910d-a1fabbec66b0-hostroot\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.613112 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-host-run-netns\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.613159 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a48de155-7389-48db-8f86-d175b7db70c8-os-release\") pod \"multus-additional-cni-plugins-9t7pq\" (UID: \"a48de155-7389-48db-8f86-d175b7db70c8\") " pod="openshift-multus/multus-additional-cni-plugins-9t7pq" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.612954 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/ed42d686-41af-470e-910d-a1fabbec66b0-host-run-multus-certs\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.613291 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.613384 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/ed42d686-41af-470e-910d-a1fabbec66b0-multus-socket-dir-parent\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.613465 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ed42d686-41af-470e-910d-a1fabbec66b0-multus-conf-dir\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.613519 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/ed42d686-41af-470e-910d-a1fabbec66b0-host-run-k8s-cni-cncf-io\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.613487 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ed42d686-41af-470e-910d-a1fabbec66b0-etc-kubernetes\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.613580 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-systemd-units\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.613607 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-etc-openvswitch\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.613717 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-log-socket\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.610741 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ed42d686-41af-470e-910d-a1fabbec66b0-os-release\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614008 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a48de155-7389-48db-8f86-d175b7db70c8-cnibin\") pod \"multus-additional-cni-plugins-9t7pq\" (UID: \"a48de155-7389-48db-8f86-d175b7db70c8\") " pod="openshift-multus/multus-additional-cni-plugins-9t7pq" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614046 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-node-log\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614128 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-ovn-node-metrics-cert\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614169 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/cbf4cf62-024e-4703-a8b8-9aecda9cd26a-rootfs\") pod \"machine-config-daemon-npms2\" (UID: \"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\") " pod="openshift-machine-config-operator/machine-config-daemon-npms2" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614260 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614284 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614299 4809 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614313 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614325 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614339 4809 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614352 4809 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614365 4809 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614379 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614391 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614405 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614417 4809 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614433 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614446 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614442 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-ovnkube-script-lib\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614482 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/cbf4cf62-024e-4703-a8b8-9aecda9cd26a-rootfs\") pod \"machine-config-daemon-npms2\" (UID: \"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\") " pod="openshift-machine-config-operator/machine-config-daemon-npms2" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614457 4809 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614514 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-node-log\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614519 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614589 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614603 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614616 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614629 4809 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614641 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614656 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614669 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614682 4809 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614694 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614706 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614718 4809 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614732 4809 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614738 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a48de155-7389-48db-8f86-d175b7db70c8-cnibin\") pod \"multus-additional-cni-plugins-9t7pq\" (UID: \"a48de155-7389-48db-8f86-d175b7db70c8\") " pod="openshift-multus/multus-additional-cni-plugins-9t7pq" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614748 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614783 4809 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614819 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614832 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614843 4809 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614861 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614874 4809 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614889 4809 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614902 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614918 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614952 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614968 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614980 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.614992 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.615007 4809 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.615019 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.615032 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.615046 4809 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.615049 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-env-overrides\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.615058 4809 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.615100 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.615115 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.615131 4809 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.615146 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.615160 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.615173 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.615187 4809 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.615200 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.615513 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/ed42d686-41af-470e-910d-a1fabbec66b0-multus-daemon-config\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.618486 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/cbf4cf62-024e-4703-a8b8-9aecda9cd26a-proxy-tls\") pod \"machine-config-daemon-npms2\" (UID: \"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\") " pod="openshift-machine-config-operator/machine-config-daemon-npms2" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.618528 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-ovn-node-metrics-cert\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.621776 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa4dd3f8-c119-46de-ad4e-dd29bb11a32c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa59e4ec9867cf280b308ee3adb5e303eaf94455aca588f5eb74a2b6263d57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5720f7f2977d9af6ed5324a30d08d34a53d55460951b3875b7d7d753d202f349\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dfd977b049a8ff81192d4041adb265366ab88475ee84c5750305d061d21c8db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d5801b61a87c95c9a623c3106e305920199c640efa8a32cf57698245697e54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.625887 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a48de155-7389-48db-8f86-d175b7db70c8-cni-binary-copy\") pod \"multus-additional-cni-plugins-9t7pq\" (UID: \"a48de155-7389-48db-8f86-d175b7db70c8\") " pod="openshift-multus/multus-additional-cni-plugins-9t7pq" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.625975 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a48de155-7389-48db-8f86-d175b7db70c8-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-9t7pq\" (UID: \"a48de155-7389-48db-8f86-d175b7db70c8\") " pod="openshift-multus/multus-additional-cni-plugins-9t7pq" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.629145 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4z265\" (UniqueName: \"kubernetes.io/projected/a48de155-7389-48db-8f86-d175b7db70c8-kube-api-access-4z265\") pod \"multus-additional-cni-plugins-9t7pq\" (UID: \"a48de155-7389-48db-8f86-d175b7db70c8\") " pod="openshift-multus/multus-additional-cni-plugins-9t7pq" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.632131 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4mgv\" (UniqueName: \"kubernetes.io/projected/e0075197-9392-4109-b8a7-39507b15cc17-kube-api-access-q4mgv\") pod \"node-resolver-rnzq4\" (UID: \"e0075197-9392-4109-b8a7-39507b15cc17\") " pod="openshift-dns/node-resolver-rnzq4" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.633616 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.633850 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-227vc\" (UniqueName: \"kubernetes.io/projected/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-kube-api-access-227vc\") pod \"ovnkube-node-5xvvb\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.635006 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mqln\" (UniqueName: \"kubernetes.io/projected/cbf4cf62-024e-4703-a8b8-9aecda9cd26a-kube-api-access-4mqln\") pod \"machine-config-daemon-npms2\" (UID: \"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\") " pod="openshift-machine-config-operator/machine-config-daemon-npms2" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.635432 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nwpx\" (UniqueName: \"kubernetes.io/projected/ed42d686-41af-470e-910d-a1fabbec66b0-kube-api-access-2nwpx\") pod \"multus-cb4vc\" (UID: \"ed42d686-41af-470e-910d-a1fabbec66b0\") " pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.649232 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5xvvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.658407 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.661475 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5636dd78-3d61-4c6a-b8fd-6d2457e19234\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.667679 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.672377 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.675159 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-cb4vc" Dec 06 05:51:35 crc kubenswrapper[4809]: W1206 05:51:35.679215 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-3a1ab400b7cf86a4c94c62e2766d31134b38cc1ddf64b79117b996589aecd9c6 WatchSource:0}: Error finding container 3a1ab400b7cf86a4c94c62e2766d31134b38cc1ddf64b79117b996589aecd9c6: Status 404 returned error can't find the container with id 3a1ab400b7cf86a4c94c62e2766d31134b38cc1ddf64b79117b996589aecd9c6 Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.682622 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-npms2" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.687337 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.689648 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.697992 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.700334 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.712306 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-rnzq4" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.716179 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.716872 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.743366 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-npms2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.775009 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rnzq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0075197-9392-4109-b8a7-39507b15cc17\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4mgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rnzq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.793151 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:51:35 crc kubenswrapper[4809]: W1206 05:51:35.797966 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcbf4cf62_024e_4703_a8b8_9aecda9cd26a.slice/crio-66d8a935a218d0c3987600f6afb595e4880a80657b1e6493b7460d3c24fe9627 WatchSource:0}: Error finding container 66d8a935a218d0c3987600f6afb595e4880a80657b1e6493b7460d3c24fe9627: Status 404 returned error can't find the container with id 66d8a935a218d0c3987600f6afb595e4880a80657b1e6493b7460d3c24fe9627 Dec 06 05:51:35 crc kubenswrapper[4809]: I1206 05:51:35.806393 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a48de155-7389-48db-8f86-d175b7db70c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9t7pq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:51:35 crc kubenswrapper[4809]: W1206 05:51:35.818557 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode0075197_9392_4109_b8a7_39507b15cc17.slice/crio-c6a321e97c4cd1c73ae00fe23ff50c76c773d0a1d0fa91b7b3d61245b8d02787 WatchSource:0}: Error finding container c6a321e97c4cd1c73ae00fe23ff50c76c773d0a1d0fa91b7b3d61245b8d02787: Status 404 returned error can't find the container with id c6a321e97c4cd1c73ae00fe23ff50c76c773d0a1d0fa91b7b3d61245b8d02787 Dec 06 05:51:35 crc kubenswrapper[4809]: W1206 05:51:35.821206 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podffb3a4f1_1e22_4220_ac80_3b2d69d2db99.slice/crio-91e57bd1e4bba7f733c10d4f92b20420e5365a2ab7815db7bba835c242b62f49 WatchSource:0}: Error finding container 91e57bd1e4bba7f733c10d4f92b20420e5365a2ab7815db7bba835c242b62f49: Status 404 returned error can't find the container with id 91e57bd1e4bba7f733c10d4f92b20420e5365a2ab7815db7bba835c242b62f49 Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.019760 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.019878 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.019978 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:51:36 crc kubenswrapper[4809]: E1206 05:51:36.020263 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 05:51:36 crc kubenswrapper[4809]: E1206 05:51:36.020684 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 05:51:36 crc kubenswrapper[4809]: E1206 05:51:36.020702 4809 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:51:36 crc kubenswrapper[4809]: E1206 05:51:36.020366 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:51:37.020345911 +0000 UTC m=+21.909328903 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:51:36 crc kubenswrapper[4809]: E1206 05:51:36.020405 4809 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 05:51:36 crc kubenswrapper[4809]: E1206 05:51:36.020785 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 05:51:37.020773102 +0000 UTC m=+21.909756124 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 05:51:36 crc kubenswrapper[4809]: E1206 05:51:36.020846 4809 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 05:51:36 crc kubenswrapper[4809]: E1206 05:51:36.020912 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 05:51:37.020869234 +0000 UTC m=+21.909852176 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:51:36 crc kubenswrapper[4809]: E1206 05:51:36.020970 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 05:51:37.020962776 +0000 UTC m=+21.909945718 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.020774 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.121860 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:51:36 crc kubenswrapper[4809]: E1206 05:51:36.122098 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 05:51:36 crc kubenswrapper[4809]: E1206 05:51:36.122276 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 05:51:36 crc kubenswrapper[4809]: E1206 05:51:36.122290 4809 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:51:36 crc kubenswrapper[4809]: E1206 05:51:36.122334 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 05:51:37.122320704 +0000 UTC m=+22.011303646 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.388319 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:51:36 crc kubenswrapper[4809]: E1206 05:51:36.388477 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.511794 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-cb4vc" event={"ID":"ed42d686-41af-470e-910d-a1fabbec66b0","Type":"ContainerStarted","Data":"73f629b20d9f61227553520b30d0ca046750433926a8d31501e5fb77467426cc"} Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.511839 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-cb4vc" event={"ID":"ed42d686-41af-470e-910d-a1fabbec66b0","Type":"ContainerStarted","Data":"31bf20b07b99f5ff0620ec6937bac871209fa1a4bf587675e94005200db70c0d"} Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.513210 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"fd45866c51410e910e87d43870bf531b0d05541e9164e700150fd37987e284af"} Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.513230 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"a3cc205fdd244e10caccfe4786dbc3cf45182993d829a352c6155ff1a1b08e4a"} Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.514545 4809 generic.go:334] "Generic (PLEG): container finished" podID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerID="3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b" exitCode=0 Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.514590 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" event={"ID":"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99","Type":"ContainerDied","Data":"3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b"} Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.514605 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" event={"ID":"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99","Type":"ContainerStarted","Data":"91e57bd1e4bba7f733c10d4f92b20420e5365a2ab7815db7bba835c242b62f49"} Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.516724 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"f28d0a3aec8a36273fa4480c552d80a5ee23210203861811e37f27bccda1147e"} Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.516752 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"40791d93876595c4093814146d269b8fe70efb64ef2888e03c25aebef0e9295e"} Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.516761 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"e6d43fb9aa36210c270943f2ddadcd893b188e76f2ff8d2b2209c0a407d46972"} Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.523216 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" event={"ID":"cbf4cf62-024e-4703-a8b8-9aecda9cd26a","Type":"ContainerStarted","Data":"947d89be9d04856af8d09facb8d4583f91898bd03a36d3ce8b9b74bc3bb795e6"} Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.523264 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" event={"ID":"cbf4cf62-024e-4703-a8b8-9aecda9cd26a","Type":"ContainerStarted","Data":"930788ec9b4d97fdc39353a93111b84707037ce445f1930aa7737a20a44e1d37"} Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.523275 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" event={"ID":"cbf4cf62-024e-4703-a8b8-9aecda9cd26a","Type":"ContainerStarted","Data":"66d8a935a218d0c3987600f6afb595e4880a80657b1e6493b7460d3c24fe9627"} Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.524248 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"3a1ab400b7cf86a4c94c62e2766d31134b38cc1ddf64b79117b996589aecd9c6"} Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.526206 4809 generic.go:334] "Generic (PLEG): container finished" podID="a48de155-7389-48db-8f86-d175b7db70c8" containerID="d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66" exitCode=0 Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.526289 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" event={"ID":"a48de155-7389-48db-8f86-d175b7db70c8","Type":"ContainerDied","Data":"d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66"} Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.526334 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" event={"ID":"a48de155-7389-48db-8f86-d175b7db70c8","Type":"ContainerStarted","Data":"749ac5355069d6aecc18d28e2caab86e2bc6ec3e007198d45af3d9e2bebcbe0a"} Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.527824 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-rnzq4" event={"ID":"e0075197-9392-4109-b8a7-39507b15cc17","Type":"ContainerStarted","Data":"e53a00b17138b236dc8517cbd4ee71ea3ed7e8885247d06eb2e8840b8b145783"} Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.527872 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-rnzq4" event={"ID":"e0075197-9392-4109-b8a7-39507b15cc17","Type":"ContainerStarted","Data":"c6a321e97c4cd1c73ae00fe23ff50c76c773d0a1d0fa91b7b3d61245b8d02787"} Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.532280 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:36Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.545068 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:36Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.557745 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:36Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.581650 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:36Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.595475 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-npms2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:36Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.613665 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5636dd78-3d61-4c6a-b8fd-6d2457e19234\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:36Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.631954 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rnzq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0075197-9392-4109-b8a7-39507b15cc17\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4mgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rnzq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:36Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.651582 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a48de155-7389-48db-8f86-d175b7db70c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9t7pq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:36Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.665049 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:36Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.677208 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cb4vc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed42d686-41af-470e-910d-a1fabbec66b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f629b20d9f61227553520b30d0ca046750433926a8d31501e5fb77467426cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2nwpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cb4vc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:36Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.693617 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa4dd3f8-c119-46de-ad4e-dd29bb11a32c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa59e4ec9867cf280b308ee3adb5e303eaf94455aca588f5eb74a2b6263d57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5720f7f2977d9af6ed5324a30d08d34a53d55460951b3875b7d7d753d202f349\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dfd977b049a8ff81192d4041adb265366ab88475ee84c5750305d061d21c8db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d5801b61a87c95c9a623c3106e305920199c640efa8a32cf57698245697e54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:36Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.704591 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:36Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.722156 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5xvvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:36Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.733686 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947d89be9d04856af8d09facb8d4583f91898bd03a36d3ce8b9b74bc3bb795e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://930788ec9b4d97fdc39353a93111b84707037ce445f1930aa7737a20a44e1d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-npms2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:36Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.755444 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5636dd78-3d61-4c6a-b8fd-6d2457e19234\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:36Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.769050 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd45866c51410e910e87d43870bf531b0d05541e9164e700150fd37987e284af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:36Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.789148 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:36Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.802755 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:36Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.817156 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:36Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.831121 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rnzq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0075197-9392-4109-b8a7-39507b15cc17\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53a00b17138b236dc8517cbd4ee71ea3ed7e8885247d06eb2e8840b8b145783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4mgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rnzq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:36Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.870012 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f28d0a3aec8a36273fa4480c552d80a5ee23210203861811e37f27bccda1147e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40791d93876595c4093814146d269b8fe70efb64ef2888e03c25aebef0e9295e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:36Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.885594 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a48de155-7389-48db-8f86-d175b7db70c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9t7pq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:36Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.900098 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cb4vc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed42d686-41af-470e-910d-a1fabbec66b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f629b20d9f61227553520b30d0ca046750433926a8d31501e5fb77467426cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2nwpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cb4vc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:36Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.913344 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa4dd3f8-c119-46de-ad4e-dd29bb11a32c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa59e4ec9867cf280b308ee3adb5e303eaf94455aca588f5eb74a2b6263d57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5720f7f2977d9af6ed5324a30d08d34a53d55460951b3875b7d7d753d202f349\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dfd977b049a8ff81192d4041adb265366ab88475ee84c5750305d061d21c8db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d5801b61a87c95c9a623c3106e305920199c640efa8a32cf57698245697e54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:36Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.924782 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:36Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.951091 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5xvvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:36Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.967970 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-hmlmc"] Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.968372 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-hmlmc" Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.970630 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.970731 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.971054 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.972083 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 06 05:51:36 crc kubenswrapper[4809]: I1206 05:51:36.997274 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f28d0a3aec8a36273fa4480c552d80a5ee23210203861811e37f27bccda1147e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40791d93876595c4093814146d269b8fe70efb64ef2888e03c25aebef0e9295e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:36Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.013526 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a48de155-7389-48db-8f86-d175b7db70c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9t7pq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:37Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.026175 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cb4vc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed42d686-41af-470e-910d-a1fabbec66b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f629b20d9f61227553520b30d0ca046750433926a8d31501e5fb77467426cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2nwpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cb4vc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:37Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.033916 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.034234 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.034273 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.034306 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:51:37 crc kubenswrapper[4809]: E1206 05:51:37.034438 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 05:51:37 crc kubenswrapper[4809]: E1206 05:51:37.034463 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 05:51:37 crc kubenswrapper[4809]: E1206 05:51:37.034474 4809 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:51:37 crc kubenswrapper[4809]: E1206 05:51:37.034517 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 05:51:39.034501392 +0000 UTC m=+23.923484334 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:51:37 crc kubenswrapper[4809]: E1206 05:51:37.034588 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:51:39.034581214 +0000 UTC m=+23.923564156 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:51:37 crc kubenswrapper[4809]: E1206 05:51:37.034645 4809 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 05:51:37 crc kubenswrapper[4809]: E1206 05:51:37.034673 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 05:51:39.034666116 +0000 UTC m=+23.923649058 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 05:51:37 crc kubenswrapper[4809]: E1206 05:51:37.034738 4809 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 05:51:37 crc kubenswrapper[4809]: E1206 05:51:37.034763 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 05:51:39.034755858 +0000 UTC m=+23.923738800 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.040419 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hmlmc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c7558c-89b1-432d-893c-ccb2b3aba183\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm7wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hmlmc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:37Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.054073 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa4dd3f8-c119-46de-ad4e-dd29bb11a32c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa59e4ec9867cf280b308ee3adb5e303eaf94455aca588f5eb74a2b6263d57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5720f7f2977d9af6ed5324a30d08d34a53d55460951b3875b7d7d753d202f349\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dfd977b049a8ff81192d4041adb265366ab88475ee84c5750305d061d21c8db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d5801b61a87c95c9a623c3106e305920199c640efa8a32cf57698245697e54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:37Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.067639 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:37Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.090115 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5xvvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:37Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.101987 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:37Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.115295 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947d89be9d04856af8d09facb8d4583f91898bd03a36d3ce8b9b74bc3bb795e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://930788ec9b4d97fdc39353a93111b84707037ce445f1930aa7737a20a44e1d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-npms2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:37Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.126760 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5636dd78-3d61-4c6a-b8fd-6d2457e19234\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:37Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.134850 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/78c7558c-89b1-432d-893c-ccb2b3aba183-serviceca\") pod \"node-ca-hmlmc\" (UID: \"78c7558c-89b1-432d-893c-ccb2b3aba183\") " pod="openshift-image-registry/node-ca-hmlmc" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.134899 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/78c7558c-89b1-432d-893c-ccb2b3aba183-host\") pod \"node-ca-hmlmc\" (UID: \"78c7558c-89b1-432d-893c-ccb2b3aba183\") " pod="openshift-image-registry/node-ca-hmlmc" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.134948 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tm7wf\" (UniqueName: \"kubernetes.io/projected/78c7558c-89b1-432d-893c-ccb2b3aba183-kube-api-access-tm7wf\") pod \"node-ca-hmlmc\" (UID: \"78c7558c-89b1-432d-893c-ccb2b3aba183\") " pod="openshift-image-registry/node-ca-hmlmc" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.134973 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:51:37 crc kubenswrapper[4809]: E1206 05:51:37.135225 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 05:51:37 crc kubenswrapper[4809]: E1206 05:51:37.135271 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 05:51:37 crc kubenswrapper[4809]: E1206 05:51:37.135287 4809 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:51:37 crc kubenswrapper[4809]: E1206 05:51:37.135344 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 05:51:39.135328588 +0000 UTC m=+24.024311520 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.141455 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd45866c51410e910e87d43870bf531b0d05541e9164e700150fd37987e284af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:37Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.158587 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:37Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.184576 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:37Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.209888 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rnzq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0075197-9392-4109-b8a7-39507b15cc17\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53a00b17138b236dc8517cbd4ee71ea3ed7e8885247d06eb2e8840b8b145783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4mgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rnzq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:37Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.236464 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/78c7558c-89b1-432d-893c-ccb2b3aba183-serviceca\") pod \"node-ca-hmlmc\" (UID: \"78c7558c-89b1-432d-893c-ccb2b3aba183\") " pod="openshift-image-registry/node-ca-hmlmc" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.236504 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/78c7558c-89b1-432d-893c-ccb2b3aba183-host\") pod \"node-ca-hmlmc\" (UID: \"78c7558c-89b1-432d-893c-ccb2b3aba183\") " pod="openshift-image-registry/node-ca-hmlmc" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.236537 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tm7wf\" (UniqueName: \"kubernetes.io/projected/78c7558c-89b1-432d-893c-ccb2b3aba183-kube-api-access-tm7wf\") pod \"node-ca-hmlmc\" (UID: \"78c7558c-89b1-432d-893c-ccb2b3aba183\") " pod="openshift-image-registry/node-ca-hmlmc" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.237605 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/78c7558c-89b1-432d-893c-ccb2b3aba183-serviceca\") pod \"node-ca-hmlmc\" (UID: \"78c7558c-89b1-432d-893c-ccb2b3aba183\") " pod="openshift-image-registry/node-ca-hmlmc" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.237651 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/78c7558c-89b1-432d-893c-ccb2b3aba183-host\") pod \"node-ca-hmlmc\" (UID: \"78c7558c-89b1-432d-893c-ccb2b3aba183\") " pod="openshift-image-registry/node-ca-hmlmc" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.277922 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tm7wf\" (UniqueName: \"kubernetes.io/projected/78c7558c-89b1-432d-893c-ccb2b3aba183-kube-api-access-tm7wf\") pod \"node-ca-hmlmc\" (UID: \"78c7558c-89b1-432d-893c-ccb2b3aba183\") " pod="openshift-image-registry/node-ca-hmlmc" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.296110 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-hmlmc" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.388896 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:51:37 crc kubenswrapper[4809]: E1206 05:51:37.389417 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.389577 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:51:37 crc kubenswrapper[4809]: E1206 05:51:37.389647 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:51:37 crc kubenswrapper[4809]: W1206 05:51:37.390031 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod78c7558c_89b1_432d_893c_ccb2b3aba183.slice/crio-db780deedd5157a468ffacf697279b854a8e8d175e7036684200804a880696ba WatchSource:0}: Error finding container db780deedd5157a468ffacf697279b854a8e8d175e7036684200804a880696ba: Status 404 returned error can't find the container with id db780deedd5157a468ffacf697279b854a8e8d175e7036684200804a880696ba Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.392925 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.393839 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.395964 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.396780 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.398144 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.398839 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.399861 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.401210 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.402054 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.403226 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.403775 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.405040 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.405550 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.406073 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.407353 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.408767 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.410272 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.410718 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.417185 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.417959 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.419239 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.419796 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.420265 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.422211 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.422653 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.423764 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.424590 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.425459 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.426077 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.427088 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.427679 4809 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.427864 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.431333 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.435529 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.436253 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.438134 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.439274 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.439821 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.441298 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.442468 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.443147 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.444278 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.445507 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.446214 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.447156 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.447679 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.448724 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.449476 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.450369 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.450963 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.451481 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.452790 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.453811 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.455317 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.534487 4809 generic.go:334] "Generic (PLEG): container finished" podID="a48de155-7389-48db-8f86-d175b7db70c8" containerID="a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d" exitCode=0 Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.534560 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" event={"ID":"a48de155-7389-48db-8f86-d175b7db70c8","Type":"ContainerDied","Data":"a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d"} Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.540968 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" event={"ID":"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99","Type":"ContainerStarted","Data":"83d58d489f532b78aca59f5b92584aad1dc292e6700c112c531cbeb3209fd90c"} Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.541005 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" event={"ID":"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99","Type":"ContainerStarted","Data":"c981561d51870deffeb446e8ca769b7ac0a1a16edbe4e982e0f2c84a45043db9"} Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.541016 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" event={"ID":"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99","Type":"ContainerStarted","Data":"0a3e4453935871bb84d68bc4a8b6c98f299b7c8094ba82291bc90e258e1fec55"} Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.541024 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" event={"ID":"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99","Type":"ContainerStarted","Data":"5be86696b7a6518c5df930accb2c6890d58cfc37104fd34a359724b9e30231b1"} Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.542218 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-hmlmc" event={"ID":"78c7558c-89b1-432d-893c-ccb2b3aba183","Type":"ContainerStarted","Data":"db780deedd5157a468ffacf697279b854a8e8d175e7036684200804a880696ba"} Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.552274 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cb4vc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed42d686-41af-470e-910d-a1fabbec66b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f629b20d9f61227553520b30d0ca046750433926a8d31501e5fb77467426cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2nwpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cb4vc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:37Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.571079 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hmlmc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c7558c-89b1-432d-893c-ccb2b3aba183\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm7wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hmlmc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:37Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.591656 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa4dd3f8-c119-46de-ad4e-dd29bb11a32c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa59e4ec9867cf280b308ee3adb5e303eaf94455aca588f5eb74a2b6263d57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5720f7f2977d9af6ed5324a30d08d34a53d55460951b3875b7d7d753d202f349\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dfd977b049a8ff81192d4041adb265366ab88475ee84c5750305d061d21c8db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d5801b61a87c95c9a623c3106e305920199c640efa8a32cf57698245697e54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:37Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.606567 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:37Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.626038 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5xvvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:37Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.638425 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:37Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.651299 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:37Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.662644 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947d89be9d04856af8d09facb8d4583f91898bd03a36d3ce8b9b74bc3bb795e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://930788ec9b4d97fdc39353a93111b84707037ce445f1930aa7737a20a44e1d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-npms2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:37Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.676748 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5636dd78-3d61-4c6a-b8fd-6d2457e19234\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:37Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.695831 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd45866c51410e910e87d43870bf531b0d05541e9164e700150fd37987e284af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:37Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.711667 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:37Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.726099 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rnzq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0075197-9392-4109-b8a7-39507b15cc17\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53a00b17138b236dc8517cbd4ee71ea3ed7e8885247d06eb2e8840b8b145783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4mgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rnzq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:37Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.743812 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f28d0a3aec8a36273fa4480c552d80a5ee23210203861811e37f27bccda1147e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40791d93876595c4093814146d269b8fe70efb64ef2888e03c25aebef0e9295e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:37Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:37 crc kubenswrapper[4809]: I1206 05:51:37.757606 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a48de155-7389-48db-8f86-d175b7db70c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9t7pq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:37Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.388365 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:51:38 crc kubenswrapper[4809]: E1206 05:51:38.389029 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.546971 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.549459 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.549505 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.549519 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.549594 4809 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.553148 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" event={"ID":"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99","Type":"ContainerStarted","Data":"3dcd421b38e5f9ba0c542a09f0047bde6c043942a096d11575afb7423ab2552c"} Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.553221 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" event={"ID":"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99","Type":"ContainerStarted","Data":"ab9b3bd4c8a6d6af7939b06fe01b8a0a50d81bb8c669bc6a27f486953c234e11"} Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.556436 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-hmlmc" event={"ID":"78c7558c-89b1-432d-893c-ccb2b3aba183","Type":"ContainerStarted","Data":"1c2f8876171399b61abd26285afba0547dac3c24c957bd6d202721b04c912aa1"} Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.558121 4809 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.558528 4809 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.558642 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"010c2681c08dc21bc3a6fb82c04ab3dc2c26dcb76ec12b143c22845cac04ed25"} Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.559496 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.559541 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.559548 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.559562 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.559573 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:38Z","lastTransitionTime":"2025-12-06T05:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.562811 4809 generic.go:334] "Generic (PLEG): container finished" podID="a48de155-7389-48db-8f86-d175b7db70c8" containerID="bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f" exitCode=0 Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.562895 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" event={"ID":"a48de155-7389-48db-8f86-d175b7db70c8","Type":"ContainerDied","Data":"bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f"} Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.571641 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cb4vc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed42d686-41af-470e-910d-a1fabbec66b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f629b20d9f61227553520b30d0ca046750433926a8d31501e5fb77467426cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2nwpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cb4vc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:38Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:38 crc kubenswrapper[4809]: E1206 05:51:38.583665 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"88a28b6b-9c85-4932-b5bc-36346303aee5\\\",\\\"systemUUID\\\":\\\"5931df91-c1df-4f78-9e5f-391a01a76127\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:38Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.589319 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hmlmc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c7558c-89b1-432d-893c-ccb2b3aba183\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c2f8876171399b61abd26285afba0547dac3c24c957bd6d202721b04c912aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm7wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hmlmc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:38Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.590925 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.590991 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.591005 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.591027 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.591040 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:38Z","lastTransitionTime":"2025-12-06T05:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:38 crc kubenswrapper[4809]: E1206 05:51:38.605329 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"88a28b6b-9c85-4932-b5bc-36346303aee5\\\",\\\"systemUUID\\\":\\\"5931df91-c1df-4f78-9e5f-391a01a76127\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:38Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.606400 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5xvvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:38Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.608806 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.608837 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.608846 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.608860 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.608869 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:38Z","lastTransitionTime":"2025-12-06T05:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.627235 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa4dd3f8-c119-46de-ad4e-dd29bb11a32c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa59e4ec9867cf280b308ee3adb5e303eaf94455aca588f5eb74a2b6263d57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5720f7f2977d9af6ed5324a30d08d34a53d55460951b3875b7d7d753d202f349\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dfd977b049a8ff81192d4041adb265366ab88475ee84c5750305d061d21c8db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d5801b61a87c95c9a623c3106e305920199c640efa8a32cf57698245697e54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:38Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:38 crc kubenswrapper[4809]: E1206 05:51:38.631761 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"88a28b6b-9c85-4932-b5bc-36346303aee5\\\",\\\"systemUUID\\\":\\\"5931df91-c1df-4f78-9e5f-391a01a76127\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:38Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.635613 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.635646 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.635655 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.635668 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.635677 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:38Z","lastTransitionTime":"2025-12-06T05:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.638708 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:38Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:38 crc kubenswrapper[4809]: E1206 05:51:38.648411 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"88a28b6b-9c85-4932-b5bc-36346303aee5\\\",\\\"systemUUID\\\":\\\"5931df91-c1df-4f78-9e5f-391a01a76127\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:38Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.651020 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd45866c51410e910e87d43870bf531b0d05541e9164e700150fd37987e284af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:38Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.652798 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.652828 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.652837 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.652851 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.652862 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:38Z","lastTransitionTime":"2025-12-06T05:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.664128 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:38Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:38 crc kubenswrapper[4809]: E1206 05:51:38.665493 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"88a28b6b-9c85-4932-b5bc-36346303aee5\\\",\\\"systemUUID\\\":\\\"5931df91-c1df-4f78-9e5f-391a01a76127\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:38Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:38 crc kubenswrapper[4809]: E1206 05:51:38.665598 4809 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.667113 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.667163 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.667178 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.667198 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.667209 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:38Z","lastTransitionTime":"2025-12-06T05:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.677100 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:38Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.692040 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:38Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.705821 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947d89be9d04856af8d09facb8d4583f91898bd03a36d3ce8b9b74bc3bb795e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://930788ec9b4d97fdc39353a93111b84707037ce445f1930aa7737a20a44e1d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-npms2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:38Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.719213 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5636dd78-3d61-4c6a-b8fd-6d2457e19234\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:38Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.737452 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rnzq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0075197-9392-4109-b8a7-39507b15cc17\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53a00b17138b236dc8517cbd4ee71ea3ed7e8885247d06eb2e8840b8b145783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4mgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rnzq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:38Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.750481 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f28d0a3aec8a36273fa4480c552d80a5ee23210203861811e37f27bccda1147e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40791d93876595c4093814146d269b8fe70efb64ef2888e03c25aebef0e9295e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:38Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.765197 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a48de155-7389-48db-8f86-d175b7db70c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9t7pq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:38Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.774705 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.774747 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.774757 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.774774 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.774783 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:38Z","lastTransitionTime":"2025-12-06T05:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.783055 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:38Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.796346 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:38Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.809016 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:38Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.829508 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947d89be9d04856af8d09facb8d4583f91898bd03a36d3ce8b9b74bc3bb795e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://930788ec9b4d97fdc39353a93111b84707037ce445f1930aa7737a20a44e1d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-npms2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:38Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.858650 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5636dd78-3d61-4c6a-b8fd-6d2457e19234\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:38Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.876779 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.876820 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.876830 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.876849 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.876861 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:38Z","lastTransitionTime":"2025-12-06T05:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.887226 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd45866c51410e910e87d43870bf531b0d05541e9164e700150fd37987e284af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:38Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.897698 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rnzq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0075197-9392-4109-b8a7-39507b15cc17\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53a00b17138b236dc8517cbd4ee71ea3ed7e8885247d06eb2e8840b8b145783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4mgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rnzq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:38Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.909594 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f28d0a3aec8a36273fa4480c552d80a5ee23210203861811e37f27bccda1147e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40791d93876595c4093814146d269b8fe70efb64ef2888e03c25aebef0e9295e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:38Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.925260 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a48de155-7389-48db-8f86-d175b7db70c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9t7pq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:38Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.937470 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cb4vc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed42d686-41af-470e-910d-a1fabbec66b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f629b20d9f61227553520b30d0ca046750433926a8d31501e5fb77467426cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2nwpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cb4vc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:38Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.947163 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hmlmc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c7558c-89b1-432d-893c-ccb2b3aba183\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c2f8876171399b61abd26285afba0547dac3c24c957bd6d202721b04c912aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm7wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hmlmc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:38Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.959475 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa4dd3f8-c119-46de-ad4e-dd29bb11a32c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa59e4ec9867cf280b308ee3adb5e303eaf94455aca588f5eb74a2b6263d57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5720f7f2977d9af6ed5324a30d08d34a53d55460951b3875b7d7d753d202f349\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dfd977b049a8ff81192d4041adb265366ab88475ee84c5750305d061d21c8db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d5801b61a87c95c9a623c3106e305920199c640efa8a32cf57698245697e54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:38Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.971050 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://010c2681c08dc21bc3a6fb82c04ab3dc2c26dcb76ec12b143c22845cac04ed25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:38Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.978966 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.979202 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.979214 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.979228 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.979238 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:38Z","lastTransitionTime":"2025-12-06T05:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:38 crc kubenswrapper[4809]: I1206 05:51:38.989167 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5xvvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:38Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.053247 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.053355 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:51:39 crc kubenswrapper[4809]: E1206 05:51:39.053396 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:51:43.053370498 +0000 UTC m=+27.942353440 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.053438 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.053496 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:51:39 crc kubenswrapper[4809]: E1206 05:51:39.053444 4809 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 05:51:39 crc kubenswrapper[4809]: E1206 05:51:39.053581 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 05:51:43.053565562 +0000 UTC m=+27.942548574 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 05:51:39 crc kubenswrapper[4809]: E1206 05:51:39.053529 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 05:51:39 crc kubenswrapper[4809]: E1206 05:51:39.053600 4809 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 05:51:39 crc kubenswrapper[4809]: E1206 05:51:39.053610 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 05:51:39 crc kubenswrapper[4809]: E1206 05:51:39.053623 4809 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:51:39 crc kubenswrapper[4809]: E1206 05:51:39.053636 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 05:51:43.053629154 +0000 UTC m=+27.942612096 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 05:51:39 crc kubenswrapper[4809]: E1206 05:51:39.053648 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 05:51:43.053641354 +0000 UTC m=+27.942624386 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.082180 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.082212 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.082221 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.082233 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.082242 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:39Z","lastTransitionTime":"2025-12-06T05:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.154008 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:51:39 crc kubenswrapper[4809]: E1206 05:51:39.154281 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 05:51:39 crc kubenswrapper[4809]: E1206 05:51:39.154558 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 05:51:39 crc kubenswrapper[4809]: E1206 05:51:39.154626 4809 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:51:39 crc kubenswrapper[4809]: E1206 05:51:39.154748 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 05:51:43.154727486 +0000 UTC m=+28.043710428 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.184687 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.184726 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.184738 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.184755 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.184775 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:39Z","lastTransitionTime":"2025-12-06T05:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.287191 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.287227 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.287238 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.287253 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.287264 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:39Z","lastTransitionTime":"2025-12-06T05:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.388438 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:51:39 crc kubenswrapper[4809]: E1206 05:51:39.388805 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.388461 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:51:39 crc kubenswrapper[4809]: E1206 05:51:39.389559 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.389840 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.389991 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.390086 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.390177 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.390282 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:39Z","lastTransitionTime":"2025-12-06T05:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.492270 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.492530 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.492690 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.492761 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.492819 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:39Z","lastTransitionTime":"2025-12-06T05:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.569608 4809 generic.go:334] "Generic (PLEG): container finished" podID="a48de155-7389-48db-8f86-d175b7db70c8" containerID="f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2" exitCode=0 Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.569853 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" event={"ID":"a48de155-7389-48db-8f86-d175b7db70c8","Type":"ContainerDied","Data":"f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2"} Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.585988 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f28d0a3aec8a36273fa4480c552d80a5ee23210203861811e37f27bccda1147e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40791d93876595c4093814146d269b8fe70efb64ef2888e03c25aebef0e9295e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:39Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.596309 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.596335 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.596365 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.596379 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.596389 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:39Z","lastTransitionTime":"2025-12-06T05:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.606838 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a48de155-7389-48db-8f86-d175b7db70c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9t7pq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:39Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.623158 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cb4vc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed42d686-41af-470e-910d-a1fabbec66b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f629b20d9f61227553520b30d0ca046750433926a8d31501e5fb77467426cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2nwpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cb4vc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:39Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.634253 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hmlmc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c7558c-89b1-432d-893c-ccb2b3aba183\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c2f8876171399b61abd26285afba0547dac3c24c957bd6d202721b04c912aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm7wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hmlmc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:39Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.648542 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa4dd3f8-c119-46de-ad4e-dd29bb11a32c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa59e4ec9867cf280b308ee3adb5e303eaf94455aca588f5eb74a2b6263d57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5720f7f2977d9af6ed5324a30d08d34a53d55460951b3875b7d7d753d202f349\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dfd977b049a8ff81192d4041adb265366ab88475ee84c5750305d061d21c8db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d5801b61a87c95c9a623c3106e305920199c640efa8a32cf57698245697e54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:39Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.663761 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://010c2681c08dc21bc3a6fb82c04ab3dc2c26dcb76ec12b143c22845cac04ed25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:39Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.687246 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5xvvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:39Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.699842 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.699882 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.699893 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.699910 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.699921 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:39Z","lastTransitionTime":"2025-12-06T05:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.700024 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:39Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.714710 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:39Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.729077 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947d89be9d04856af8d09facb8d4583f91898bd03a36d3ce8b9b74bc3bb795e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://930788ec9b4d97fdc39353a93111b84707037ce445f1930aa7737a20a44e1d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-npms2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:39Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.744501 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5636dd78-3d61-4c6a-b8fd-6d2457e19234\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:39Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.758909 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd45866c51410e910e87d43870bf531b0d05541e9164e700150fd37987e284af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:39Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.774452 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:39Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.787074 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rnzq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0075197-9392-4109-b8a7-39507b15cc17\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53a00b17138b236dc8517cbd4ee71ea3ed7e8885247d06eb2e8840b8b145783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4mgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rnzq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:39Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.801609 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.801640 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.801652 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.801668 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.801681 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:39Z","lastTransitionTime":"2025-12-06T05:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.904580 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.904622 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.904637 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.904662 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:39 crc kubenswrapper[4809]: I1206 05:51:39.904678 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:39Z","lastTransitionTime":"2025-12-06T05:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.006731 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.007007 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.007097 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.007193 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.007253 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:40Z","lastTransitionTime":"2025-12-06T05:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.109827 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.109878 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.109891 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.109913 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.109943 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:40Z","lastTransitionTime":"2025-12-06T05:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.213678 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.213739 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.213752 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.213775 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.213787 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:40Z","lastTransitionTime":"2025-12-06T05:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.318506 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.318573 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.318589 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.318615 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.318631 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:40Z","lastTransitionTime":"2025-12-06T05:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.388186 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:51:40 crc kubenswrapper[4809]: E1206 05:51:40.388323 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.422266 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.422312 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.422338 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.422353 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.422364 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:40Z","lastTransitionTime":"2025-12-06T05:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.524681 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.524725 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.524740 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.524756 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.524768 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:40Z","lastTransitionTime":"2025-12-06T05:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.575470 4809 generic.go:334] "Generic (PLEG): container finished" podID="a48de155-7389-48db-8f86-d175b7db70c8" containerID="f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61" exitCode=0 Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.575543 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" event={"ID":"a48de155-7389-48db-8f86-d175b7db70c8","Type":"ContainerDied","Data":"f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61"} Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.586634 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" event={"ID":"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99","Type":"ContainerStarted","Data":"dee8b965be3b2486f1f532f2bdfbbcbd84f9b3b711a11c25e86dae1284314d4c"} Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.596010 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cb4vc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed42d686-41af-470e-910d-a1fabbec66b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f629b20d9f61227553520b30d0ca046750433926a8d31501e5fb77467426cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2nwpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cb4vc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.609148 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hmlmc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c7558c-89b1-432d-893c-ccb2b3aba183\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c2f8876171399b61abd26285afba0547dac3c24c957bd6d202721b04c912aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm7wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hmlmc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.622653 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://010c2681c08dc21bc3a6fb82c04ab3dc2c26dcb76ec12b143c22845cac04ed25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.628339 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.628403 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.628420 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.628445 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.628463 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:40Z","lastTransitionTime":"2025-12-06T05:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.642275 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5xvvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.659424 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa4dd3f8-c119-46de-ad4e-dd29bb11a32c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa59e4ec9867cf280b308ee3adb5e303eaf94455aca588f5eb74a2b6263d57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5720f7f2977d9af6ed5324a30d08d34a53d55460951b3875b7d7d753d202f349\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dfd977b049a8ff81192d4041adb265366ab88475ee84c5750305d061d21c8db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d5801b61a87c95c9a623c3106e305920199c640efa8a32cf57698245697e54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.676294 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd45866c51410e910e87d43870bf531b0d05541e9164e700150fd37987e284af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.689731 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.703342 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.712521 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.724980 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.730696 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.730726 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.730736 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.730751 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.730763 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:40Z","lastTransitionTime":"2025-12-06T05:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.730841 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.745952 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947d89be9d04856af8d09facb8d4583f91898bd03a36d3ce8b9b74bc3bb795e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://930788ec9b4d97fdc39353a93111b84707037ce445f1930aa7737a20a44e1d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-npms2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.761175 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5636dd78-3d61-4c6a-b8fd-6d2457e19234\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.773230 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rnzq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0075197-9392-4109-b8a7-39507b15cc17\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53a00b17138b236dc8517cbd4ee71ea3ed7e8885247d06eb2e8840b8b145783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4mgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rnzq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.789568 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f28d0a3aec8a36273fa4480c552d80a5ee23210203861811e37f27bccda1147e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40791d93876595c4093814146d269b8fe70efb64ef2888e03c25aebef0e9295e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.806493 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a48de155-7389-48db-8f86-d175b7db70c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9t7pq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.824772 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa4dd3f8-c119-46de-ad4e-dd29bb11a32c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa59e4ec9867cf280b308ee3adb5e303eaf94455aca588f5eb74a2b6263d57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5720f7f2977d9af6ed5324a30d08d34a53d55460951b3875b7d7d753d202f349\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dfd977b049a8ff81192d4041adb265366ab88475ee84c5750305d061d21c8db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d5801b61a87c95c9a623c3106e305920199c640efa8a32cf57698245697e54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.834678 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.834730 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.834747 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.834769 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.834781 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:40Z","lastTransitionTime":"2025-12-06T05:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.838358 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://010c2681c08dc21bc3a6fb82c04ab3dc2c26dcb76ec12b143c22845cac04ed25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.857082 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5xvvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.870691 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5636dd78-3d61-4c6a-b8fd-6d2457e19234\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.885299 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd45866c51410e910e87d43870bf531b0d05541e9164e700150fd37987e284af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.901758 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.914227 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.927324 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.939428 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.939471 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.939487 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.939505 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.939516 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:40Z","lastTransitionTime":"2025-12-06T05:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.940309 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947d89be9d04856af8d09facb8d4583f91898bd03a36d3ce8b9b74bc3bb795e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://930788ec9b4d97fdc39353a93111b84707037ce445f1930aa7737a20a44e1d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-npms2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.950878 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rnzq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0075197-9392-4109-b8a7-39507b15cc17\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53a00b17138b236dc8517cbd4ee71ea3ed7e8885247d06eb2e8840b8b145783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4mgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rnzq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.963352 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f28d0a3aec8a36273fa4480c552d80a5ee23210203861811e37f27bccda1147e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40791d93876595c4093814146d269b8fe70efb64ef2888e03c25aebef0e9295e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.976851 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a48de155-7389-48db-8f86-d175b7db70c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9t7pq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:40 crc kubenswrapper[4809]: I1206 05:51:40.999051 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cb4vc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed42d686-41af-470e-910d-a1fabbec66b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f629b20d9f61227553520b30d0ca046750433926a8d31501e5fb77467426cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2nwpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cb4vc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.011383 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hmlmc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c7558c-89b1-432d-893c-ccb2b3aba183\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c2f8876171399b61abd26285afba0547dac3c24c957bd6d202721b04c912aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm7wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hmlmc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.045417 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.045455 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.045581 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.045687 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.045708 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:41Z","lastTransitionTime":"2025-12-06T05:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.148518 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.148554 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.148563 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.148579 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.148590 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:41Z","lastTransitionTime":"2025-12-06T05:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.251774 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.251828 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.251843 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.251863 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.251878 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:41Z","lastTransitionTime":"2025-12-06T05:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.354837 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.354893 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.354908 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.354967 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.354985 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:41Z","lastTransitionTime":"2025-12-06T05:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.388386 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:51:41 crc kubenswrapper[4809]: E1206 05:51:41.388536 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.388607 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:51:41 crc kubenswrapper[4809]: E1206 05:51:41.388797 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.458732 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.458779 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.458796 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.458818 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.458835 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:41Z","lastTransitionTime":"2025-12-06T05:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.560846 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.560908 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.560919 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.560972 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.560986 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:41Z","lastTransitionTime":"2025-12-06T05:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.593486 4809 generic.go:334] "Generic (PLEG): container finished" podID="a48de155-7389-48db-8f86-d175b7db70c8" containerID="a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5" exitCode=0 Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.593633 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" event={"ID":"a48de155-7389-48db-8f86-d175b7db70c8","Type":"ContainerDied","Data":"a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5"} Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.608370 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hmlmc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c7558c-89b1-432d-893c-ccb2b3aba183\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c2f8876171399b61abd26285afba0547dac3c24c957bd6d202721b04c912aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm7wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hmlmc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.623293 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cb4vc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed42d686-41af-470e-910d-a1fabbec66b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f629b20d9f61227553520b30d0ca046750433926a8d31501e5fb77467426cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2nwpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cb4vc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.635758 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa4dd3f8-c119-46de-ad4e-dd29bb11a32c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa59e4ec9867cf280b308ee3adb5e303eaf94455aca588f5eb74a2b6263d57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5720f7f2977d9af6ed5324a30d08d34a53d55460951b3875b7d7d753d202f349\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dfd977b049a8ff81192d4041adb265366ab88475ee84c5750305d061d21c8db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d5801b61a87c95c9a623c3106e305920199c640efa8a32cf57698245697e54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.648098 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://010c2681c08dc21bc3a6fb82c04ab3dc2c26dcb76ec12b143c22845cac04ed25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.664154 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.664208 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.664220 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.664238 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.664249 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:41Z","lastTransitionTime":"2025-12-06T05:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.667889 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5xvvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.681429 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd45866c51410e910e87d43870bf531b0d05541e9164e700150fd37987e284af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.694123 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.705664 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.723954 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.737166 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947d89be9d04856af8d09facb8d4583f91898bd03a36d3ce8b9b74bc3bb795e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://930788ec9b4d97fdc39353a93111b84707037ce445f1930aa7737a20a44e1d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-npms2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.752487 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5636dd78-3d61-4c6a-b8fd-6d2457e19234\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.766466 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.766493 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.766502 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.766518 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.766527 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:41Z","lastTransitionTime":"2025-12-06T05:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.766805 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rnzq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0075197-9392-4109-b8a7-39507b15cc17\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53a00b17138b236dc8517cbd4ee71ea3ed7e8885247d06eb2e8840b8b145783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4mgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rnzq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.785453 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a48de155-7389-48db-8f86-d175b7db70c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9t7pq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.800037 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f28d0a3aec8a36273fa4480c552d80a5ee23210203861811e37f27bccda1147e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40791d93876595c4093814146d269b8fe70efb64ef2888e03c25aebef0e9295e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.868574 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.868638 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.868660 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.868688 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.868710 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:41Z","lastTransitionTime":"2025-12-06T05:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.971524 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.971567 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.971582 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.971601 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:41 crc kubenswrapper[4809]: I1206 05:51:41.971619 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:41Z","lastTransitionTime":"2025-12-06T05:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.074436 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.074487 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.074498 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.074522 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.074534 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:42Z","lastTransitionTime":"2025-12-06T05:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.177070 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.177103 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.177111 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.177125 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.177134 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:42Z","lastTransitionTime":"2025-12-06T05:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.279609 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.279642 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.279651 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.279663 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.279671 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:42Z","lastTransitionTime":"2025-12-06T05:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.382518 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.382617 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.382643 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.382678 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.382701 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:42Z","lastTransitionTime":"2025-12-06T05:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.387872 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:51:42 crc kubenswrapper[4809]: E1206 05:51:42.388078 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.485216 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.485256 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.485268 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.485297 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.485309 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:42Z","lastTransitionTime":"2025-12-06T05:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.587734 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.588312 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.588326 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.588345 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.588357 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:42Z","lastTransitionTime":"2025-12-06T05:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.600488 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" event={"ID":"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99","Type":"ContainerStarted","Data":"9ef6d4c42523249187a0998e293a702feb4e1aec2f31cade2e6f5a3cbfcb3ce5"} Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.600789 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.605311 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" event={"ID":"a48de155-7389-48db-8f86-d175b7db70c8","Type":"ContainerStarted","Data":"701274e49f68efbc0a939bc08771288c50662cef9c59da73170f7f114e222337"} Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.614333 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cb4vc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed42d686-41af-470e-910d-a1fabbec66b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f629b20d9f61227553520b30d0ca046750433926a8d31501e5fb77467426cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2nwpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cb4vc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:42Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.626315 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hmlmc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c7558c-89b1-432d-893c-ccb2b3aba183\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c2f8876171399b61abd26285afba0547dac3c24c957bd6d202721b04c912aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm7wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hmlmc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:42Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.627061 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.636638 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa4dd3f8-c119-46de-ad4e-dd29bb11a32c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa59e4ec9867cf280b308ee3adb5e303eaf94455aca588f5eb74a2b6263d57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5720f7f2977d9af6ed5324a30d08d34a53d55460951b3875b7d7d753d202f349\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dfd977b049a8ff81192d4041adb265366ab88475ee84c5750305d061d21c8db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d5801b61a87c95c9a623c3106e305920199c640efa8a32cf57698245697e54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:42Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.647648 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://010c2681c08dc21bc3a6fb82c04ab3dc2c26dcb76ec12b143c22845cac04ed25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:42Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.665375 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c981561d51870deffeb446e8ca769b7ac0a1a16edbe4e982e0f2c84a45043db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d58d489f532b78aca59f5b92584aad1dc292e6700c112c531cbeb3209fd90c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dcd421b38e5f9ba0c542a09f0047bde6c043942a096d11575afb7423ab2552c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab9b3bd4c8a6d6af7939b06fe01b8a0a50d81bb8c669bc6a27f486953c234e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a3e4453935871bb84d68bc4a8b6c98f299b7c8094ba82291bc90e258e1fec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5be86696b7a6518c5df930accb2c6890d58cfc37104fd34a359724b9e30231b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ef6d4c42523249187a0998e293a702feb4e1aec2f31cade2e6f5a3cbfcb3ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dee8b965be3b2486f1f532f2bdfbbcbd84f9b3b711a11c25e86dae1284314d4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5xvvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:42Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.681223 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5636dd78-3d61-4c6a-b8fd-6d2457e19234\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:42Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.690677 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.690710 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.690719 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.690732 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.690741 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:42Z","lastTransitionTime":"2025-12-06T05:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.693661 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd45866c51410e910e87d43870bf531b0d05541e9164e700150fd37987e284af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:42Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.707609 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:42Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.720916 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:42Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.732505 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:42Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.744792 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947d89be9d04856af8d09facb8d4583f91898bd03a36d3ce8b9b74bc3bb795e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://930788ec9b4d97fdc39353a93111b84707037ce445f1930aa7737a20a44e1d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-npms2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:42Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.754588 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rnzq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0075197-9392-4109-b8a7-39507b15cc17\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53a00b17138b236dc8517cbd4ee71ea3ed7e8885247d06eb2e8840b8b145783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4mgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rnzq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:42Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.765643 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f28d0a3aec8a36273fa4480c552d80a5ee23210203861811e37f27bccda1147e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40791d93876595c4093814146d269b8fe70efb64ef2888e03c25aebef0e9295e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:42Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.778499 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a48de155-7389-48db-8f86-d175b7db70c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9t7pq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:42Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.790655 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa4dd3f8-c119-46de-ad4e-dd29bb11a32c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa59e4ec9867cf280b308ee3adb5e303eaf94455aca588f5eb74a2b6263d57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5720f7f2977d9af6ed5324a30d08d34a53d55460951b3875b7d7d753d202f349\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dfd977b049a8ff81192d4041adb265366ab88475ee84c5750305d061d21c8db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d5801b61a87c95c9a623c3106e305920199c640efa8a32cf57698245697e54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:42Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.792285 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.792329 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.792350 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.792369 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.792380 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:42Z","lastTransitionTime":"2025-12-06T05:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.801137 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://010c2681c08dc21bc3a6fb82c04ab3dc2c26dcb76ec12b143c22845cac04ed25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:42Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.817946 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c981561d51870deffeb446e8ca769b7ac0a1a16edbe4e982e0f2c84a45043db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d58d489f532b78aca59f5b92584aad1dc292e6700c112c531cbeb3209fd90c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dcd421b38e5f9ba0c542a09f0047bde6c043942a096d11575afb7423ab2552c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab9b3bd4c8a6d6af7939b06fe01b8a0a50d81bb8c669bc6a27f486953c234e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a3e4453935871bb84d68bc4a8b6c98f299b7c8094ba82291bc90e258e1fec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5be86696b7a6518c5df930accb2c6890d58cfc37104fd34a359724b9e30231b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ef6d4c42523249187a0998e293a702feb4e1aec2f31cade2e6f5a3cbfcb3ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dee8b965be3b2486f1f532f2bdfbbcbd84f9b3b711a11c25e86dae1284314d4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5xvvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:42Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.829793 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd45866c51410e910e87d43870bf531b0d05541e9164e700150fd37987e284af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:42Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.842612 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:42Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.854259 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:42Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.865258 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:42Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.876440 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947d89be9d04856af8d09facb8d4583f91898bd03a36d3ce8b9b74bc3bb795e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://930788ec9b4d97fdc39353a93111b84707037ce445f1930aa7737a20a44e1d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-npms2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:42Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.889036 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5636dd78-3d61-4c6a-b8fd-6d2457e19234\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:42Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.894317 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.894374 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.894389 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.894408 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.894419 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:42Z","lastTransitionTime":"2025-12-06T05:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.898375 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rnzq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0075197-9392-4109-b8a7-39507b15cc17\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53a00b17138b236dc8517cbd4ee71ea3ed7e8885247d06eb2e8840b8b145783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4mgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rnzq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:42Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.912230 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a48de155-7389-48db-8f86-d175b7db70c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://701274e49f68efbc0a939bc08771288c50662cef9c59da73170f7f114e222337\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9t7pq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:42Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.924290 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f28d0a3aec8a36273fa4480c552d80a5ee23210203861811e37f27bccda1147e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40791d93876595c4093814146d269b8fe70efb64ef2888e03c25aebef0e9295e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:42Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.933489 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hmlmc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c7558c-89b1-432d-893c-ccb2b3aba183\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c2f8876171399b61abd26285afba0547dac3c24c957bd6d202721b04c912aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm7wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hmlmc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:42Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.945243 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cb4vc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed42d686-41af-470e-910d-a1fabbec66b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f629b20d9f61227553520b30d0ca046750433926a8d31501e5fb77467426cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2nwpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cb4vc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:42Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.996792 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.996853 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.996872 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.996898 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:42 crc kubenswrapper[4809]: I1206 05:51:42.996919 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:42Z","lastTransitionTime":"2025-12-06T05:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.095848 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.096030 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.096070 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:51:43 crc kubenswrapper[4809]: E1206 05:51:43.096160 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:51:51.096122318 +0000 UTC m=+35.985105310 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:51:43 crc kubenswrapper[4809]: E1206 05:51:43.096192 4809 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 05:51:43 crc kubenswrapper[4809]: E1206 05:51:43.096236 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 05:51:43 crc kubenswrapper[4809]: E1206 05:51:43.096263 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 05:51:43 crc kubenswrapper[4809]: E1206 05:51:43.096267 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 05:51:51.096247831 +0000 UTC m=+35.985230773 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 05:51:43 crc kubenswrapper[4809]: E1206 05:51:43.096284 4809 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.096293 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:51:43 crc kubenswrapper[4809]: E1206 05:51:43.096337 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 05:51:51.096324763 +0000 UTC m=+35.985307895 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:51:43 crc kubenswrapper[4809]: E1206 05:51:43.096359 4809 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 05:51:43 crc kubenswrapper[4809]: E1206 05:51:43.096393 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 05:51:51.096385315 +0000 UTC m=+35.985368257 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.099489 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.099539 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.099552 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.099567 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.099577 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:43Z","lastTransitionTime":"2025-12-06T05:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.197641 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:51:43 crc kubenswrapper[4809]: E1206 05:51:43.197962 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 05:51:43 crc kubenswrapper[4809]: E1206 05:51:43.198006 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 05:51:43 crc kubenswrapper[4809]: E1206 05:51:43.198022 4809 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:51:43 crc kubenswrapper[4809]: E1206 05:51:43.198101 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 05:51:51.198074492 +0000 UTC m=+36.087057494 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.202449 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.202525 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.202545 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.202572 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.202591 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:43Z","lastTransitionTime":"2025-12-06T05:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.304615 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.304664 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.304677 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.304694 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.304708 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:43Z","lastTransitionTime":"2025-12-06T05:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.388376 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:51:43 crc kubenswrapper[4809]: E1206 05:51:43.388533 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.388613 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:51:43 crc kubenswrapper[4809]: E1206 05:51:43.388795 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.407382 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.407604 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.407688 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.407758 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.407818 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:43Z","lastTransitionTime":"2025-12-06T05:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.511523 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.511613 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.511640 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.511680 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.511702 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:43Z","lastTransitionTime":"2025-12-06T05:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.609421 4809 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.609985 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.614072 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.614116 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.614128 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.614151 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.614165 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:43Z","lastTransitionTime":"2025-12-06T05:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.640020 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.655276 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f28d0a3aec8a36273fa4480c552d80a5ee23210203861811e37f27bccda1147e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40791d93876595c4093814146d269b8fe70efb64ef2888e03c25aebef0e9295e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.675039 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a48de155-7389-48db-8f86-d175b7db70c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://701274e49f68efbc0a939bc08771288c50662cef9c59da73170f7f114e222337\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9t7pq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.714782 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cb4vc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed42d686-41af-470e-910d-a1fabbec66b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f629b20d9f61227553520b30d0ca046750433926a8d31501e5fb77467426cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2nwpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cb4vc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.716353 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.716394 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.716406 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.716423 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.716434 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:43Z","lastTransitionTime":"2025-12-06T05:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.739586 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hmlmc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c7558c-89b1-432d-893c-ccb2b3aba183\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c2f8876171399b61abd26285afba0547dac3c24c957bd6d202721b04c912aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm7wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hmlmc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.758725 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://010c2681c08dc21bc3a6fb82c04ab3dc2c26dcb76ec12b143c22845cac04ed25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.777079 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c981561d51870deffeb446e8ca769b7ac0a1a16edbe4e982e0f2c84a45043db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d58d489f532b78aca59f5b92584aad1dc292e6700c112c531cbeb3209fd90c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dcd421b38e5f9ba0c542a09f0047bde6c043942a096d11575afb7423ab2552c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab9b3bd4c8a6d6af7939b06fe01b8a0a50d81bb8c669bc6a27f486953c234e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a3e4453935871bb84d68bc4a8b6c98f299b7c8094ba82291bc90e258e1fec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5be86696b7a6518c5df930accb2c6890d58cfc37104fd34a359724b9e30231b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ef6d4c42523249187a0998e293a702feb4e1aec2f31cade2e6f5a3cbfcb3ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dee8b965be3b2486f1f532f2bdfbbcbd84f9b3b711a11c25e86dae1284314d4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5xvvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.788563 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa4dd3f8-c119-46de-ad4e-dd29bb11a32c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa59e4ec9867cf280b308ee3adb5e303eaf94455aca588f5eb74a2b6263d57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5720f7f2977d9af6ed5324a30d08d34a53d55460951b3875b7d7d753d202f349\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dfd977b049a8ff81192d4041adb265366ab88475ee84c5750305d061d21c8db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d5801b61a87c95c9a623c3106e305920199c640efa8a32cf57698245697e54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.801815 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd45866c51410e910e87d43870bf531b0d05541e9164e700150fd37987e284af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.814955 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.819781 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.819812 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.819827 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.819844 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.819855 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:43Z","lastTransitionTime":"2025-12-06T05:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.863223 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.888857 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.900296 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947d89be9d04856af8d09facb8d4583f91898bd03a36d3ce8b9b74bc3bb795e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://930788ec9b4d97fdc39353a93111b84707037ce445f1930aa7737a20a44e1d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-npms2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.913945 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5636dd78-3d61-4c6a-b8fd-6d2457e19234\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.922398 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.922473 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.922491 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.922517 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.922532 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:43Z","lastTransitionTime":"2025-12-06T05:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:43 crc kubenswrapper[4809]: I1206 05:51:43.931457 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rnzq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0075197-9392-4109-b8a7-39507b15cc17\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53a00b17138b236dc8517cbd4ee71ea3ed7e8885247d06eb2e8840b8b145783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4mgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rnzq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.024381 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.024419 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.024428 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.024442 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.024451 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:44Z","lastTransitionTime":"2025-12-06T05:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.130910 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.130977 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.130990 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.131012 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.131025 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:44Z","lastTransitionTime":"2025-12-06T05:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.233795 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.233844 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.233854 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.233869 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.233881 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:44Z","lastTransitionTime":"2025-12-06T05:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.336544 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.336764 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.336820 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.336916 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.337005 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:44Z","lastTransitionTime":"2025-12-06T05:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.388083 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:51:44 crc kubenswrapper[4809]: E1206 05:51:44.388238 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.439555 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.439851 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.440015 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.440107 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.440207 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:44Z","lastTransitionTime":"2025-12-06T05:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.543276 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.543332 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.543341 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.543383 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.543393 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:44Z","lastTransitionTime":"2025-12-06T05:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.612281 4809 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.646401 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.646454 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.646462 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.646476 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.646485 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:44Z","lastTransitionTime":"2025-12-06T05:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.748613 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.748644 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.748652 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.748664 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.748673 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:44Z","lastTransitionTime":"2025-12-06T05:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.850487 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.850532 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.850544 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.850564 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.850584 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:44Z","lastTransitionTime":"2025-12-06T05:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.953161 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.953199 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.953211 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.953227 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:44 crc kubenswrapper[4809]: I1206 05:51:44.953238 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:44Z","lastTransitionTime":"2025-12-06T05:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.055116 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.055151 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.055160 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.055172 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.055181 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:45Z","lastTransitionTime":"2025-12-06T05:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.157694 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.157754 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.157765 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.157776 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.157785 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:45Z","lastTransitionTime":"2025-12-06T05:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.259789 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.259848 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.259860 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.259876 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.259887 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:45Z","lastTransitionTime":"2025-12-06T05:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.362454 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.362490 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.362501 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.362516 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.362524 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:45Z","lastTransitionTime":"2025-12-06T05:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.387746 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.387751 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:51:45 crc kubenswrapper[4809]: E1206 05:51:45.387854 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:51:45 crc kubenswrapper[4809]: E1206 05:51:45.387907 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.400726 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd45866c51410e910e87d43870bf531b0d05541e9164e700150fd37987e284af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.412678 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.423505 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.436341 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.449491 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947d89be9d04856af8d09facb8d4583f91898bd03a36d3ce8b9b74bc3bb795e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://930788ec9b4d97fdc39353a93111b84707037ce445f1930aa7737a20a44e1d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-npms2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.463267 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5636dd78-3d61-4c6a-b8fd-6d2457e19234\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.464478 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.464516 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.464533 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.464552 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.464567 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:45Z","lastTransitionTime":"2025-12-06T05:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.476881 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rnzq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0075197-9392-4109-b8a7-39507b15cc17\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53a00b17138b236dc8517cbd4ee71ea3ed7e8885247d06eb2e8840b8b145783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4mgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rnzq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.491621 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a48de155-7389-48db-8f86-d175b7db70c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://701274e49f68efbc0a939bc08771288c50662cef9c59da73170f7f114e222337\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9t7pq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.510624 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f28d0a3aec8a36273fa4480c552d80a5ee23210203861811e37f27bccda1147e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40791d93876595c4093814146d269b8fe70efb64ef2888e03c25aebef0e9295e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.519614 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hmlmc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c7558c-89b1-432d-893c-ccb2b3aba183\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c2f8876171399b61abd26285afba0547dac3c24c957bd6d202721b04c912aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm7wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hmlmc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.532154 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cb4vc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed42d686-41af-470e-910d-a1fabbec66b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f629b20d9f61227553520b30d0ca046750433926a8d31501e5fb77467426cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2nwpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cb4vc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.543780 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa4dd3f8-c119-46de-ad4e-dd29bb11a32c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa59e4ec9867cf280b308ee3adb5e303eaf94455aca588f5eb74a2b6263d57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5720f7f2977d9af6ed5324a30d08d34a53d55460951b3875b7d7d753d202f349\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dfd977b049a8ff81192d4041adb265366ab88475ee84c5750305d061d21c8db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d5801b61a87c95c9a623c3106e305920199c640efa8a32cf57698245697e54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.553787 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://010c2681c08dc21bc3a6fb82c04ab3dc2c26dcb76ec12b143c22845cac04ed25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.566417 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.566479 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.566493 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.566511 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.566523 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:45Z","lastTransitionTime":"2025-12-06T05:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.578667 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c981561d51870deffeb446e8ca769b7ac0a1a16edbe4e982e0f2c84a45043db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d58d489f532b78aca59f5b92584aad1dc292e6700c112c531cbeb3209fd90c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dcd421b38e5f9ba0c542a09f0047bde6c043942a096d11575afb7423ab2552c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab9b3bd4c8a6d6af7939b06fe01b8a0a50d81bb8c669bc6a27f486953c234e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a3e4453935871bb84d68bc4a8b6c98f299b7c8094ba82291bc90e258e1fec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5be86696b7a6518c5df930accb2c6890d58cfc37104fd34a359724b9e30231b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ef6d4c42523249187a0998e293a702feb4e1aec2f31cade2e6f5a3cbfcb3ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dee8b965be3b2486f1f532f2bdfbbcbd84f9b3b711a11c25e86dae1284314d4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5xvvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.618789 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5xvvb_ffb3a4f1-1e22-4220-ac80-3b2d69d2db99/ovnkube-controller/0.log" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.621254 4809 generic.go:334] "Generic (PLEG): container finished" podID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerID="9ef6d4c42523249187a0998e293a702feb4e1aec2f31cade2e6f5a3cbfcb3ce5" exitCode=1 Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.621306 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" event={"ID":"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99","Type":"ContainerDied","Data":"9ef6d4c42523249187a0998e293a702feb4e1aec2f31cade2e6f5a3cbfcb3ce5"} Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.621947 4809 scope.go:117] "RemoveContainer" containerID="9ef6d4c42523249187a0998e293a702feb4e1aec2f31cade2e6f5a3cbfcb3ce5" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.637780 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.653793 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947d89be9d04856af8d09facb8d4583f91898bd03a36d3ce8b9b74bc3bb795e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://930788ec9b4d97fdc39353a93111b84707037ce445f1930aa7737a20a44e1d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-npms2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.669370 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5636dd78-3d61-4c6a-b8fd-6d2457e19234\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.669688 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.669717 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.669728 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.669744 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.669756 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:45Z","lastTransitionTime":"2025-12-06T05:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.682392 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd45866c51410e910e87d43870bf531b0d05541e9164e700150fd37987e284af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.695312 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.711508 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.722368 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rnzq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0075197-9392-4109-b8a7-39507b15cc17\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53a00b17138b236dc8517cbd4ee71ea3ed7e8885247d06eb2e8840b8b145783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4mgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rnzq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.739193 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f28d0a3aec8a36273fa4480c552d80a5ee23210203861811e37f27bccda1147e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40791d93876595c4093814146d269b8fe70efb64ef2888e03c25aebef0e9295e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.754927 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a48de155-7389-48db-8f86-d175b7db70c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://701274e49f68efbc0a939bc08771288c50662cef9c59da73170f7f114e222337\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9t7pq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.770520 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cb4vc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed42d686-41af-470e-910d-a1fabbec66b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f629b20d9f61227553520b30d0ca046750433926a8d31501e5fb77467426cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2nwpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cb4vc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.772146 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.772190 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.772203 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.772226 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.772242 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:45Z","lastTransitionTime":"2025-12-06T05:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.782298 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hmlmc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c7558c-89b1-432d-893c-ccb2b3aba183\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c2f8876171399b61abd26285afba0547dac3c24c957bd6d202721b04c912aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm7wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hmlmc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.796564 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa4dd3f8-c119-46de-ad4e-dd29bb11a32c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa59e4ec9867cf280b308ee3adb5e303eaf94455aca588f5eb74a2b6263d57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5720f7f2977d9af6ed5324a30d08d34a53d55460951b3875b7d7d753d202f349\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dfd977b049a8ff81192d4041adb265366ab88475ee84c5750305d061d21c8db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d5801b61a87c95c9a623c3106e305920199c640efa8a32cf57698245697e54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.808770 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://010c2681c08dc21bc3a6fb82c04ab3dc2c26dcb76ec12b143c22845cac04ed25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.827913 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c981561d51870deffeb446e8ca769b7ac0a1a16edbe4e982e0f2c84a45043db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d58d489f532b78aca59f5b92584aad1dc292e6700c112c531cbeb3209fd90c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dcd421b38e5f9ba0c542a09f0047bde6c043942a096d11575afb7423ab2552c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab9b3bd4c8a6d6af7939b06fe01b8a0a50d81bb8c669bc6a27f486953c234e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a3e4453935871bb84d68bc4a8b6c98f299b7c8094ba82291bc90e258e1fec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5be86696b7a6518c5df930accb2c6890d58cfc37104fd34a359724b9e30231b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ef6d4c42523249187a0998e293a702feb4e1aec2f31cade2e6f5a3cbfcb3ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ef6d4c42523249187a0998e293a702feb4e1aec2f31cade2e6f5a3cbfcb3ce5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:51:45Z\\\",\\\"message\\\":\\\"(0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 05:51:44.868917 6104 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 05:51:44.869052 6104 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 05:51:44.869150 6104 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 05:51:44.870031 6104 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 05:51:44.870070 6104 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 05:51:44.870083 6104 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 05:51:44.870088 6104 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 05:51:44.870102 6104 factory.go:656] Stopping watch factory\\\\nI1206 05:51:44.870114 6104 ovnkube.go:599] Stopped ovnkube\\\\nI1206 05:51:44.870135 6104 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 05:51:44.870134 6104 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 05:51:44.870148 6104 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dee8b965be3b2486f1f532f2bdfbbcbd84f9b3b711a11c25e86dae1284314d4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5xvvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.874382 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.874428 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.874438 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.874458 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.874471 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:45Z","lastTransitionTime":"2025-12-06T05:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.976614 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.976660 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.976672 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.976687 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:45 crc kubenswrapper[4809]: I1206 05:51:45.976699 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:45Z","lastTransitionTime":"2025-12-06T05:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.078574 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.078952 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.078968 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.078988 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.079004 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:46Z","lastTransitionTime":"2025-12-06T05:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.181254 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.181311 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.181325 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.181345 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.181361 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:46Z","lastTransitionTime":"2025-12-06T05:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.285134 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.285191 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.285201 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.285217 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.285227 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:46Z","lastTransitionTime":"2025-12-06T05:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.387550 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:51:46 crc kubenswrapper[4809]: E1206 05:51:46.387754 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.387903 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.387962 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.387971 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.387987 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.388001 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:46Z","lastTransitionTime":"2025-12-06T05:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.490263 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.490298 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.490310 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.490325 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.490335 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:46Z","lastTransitionTime":"2025-12-06T05:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.592903 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.592969 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.592980 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.592998 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.593009 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:46Z","lastTransitionTime":"2025-12-06T05:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.626643 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5xvvb_ffb3a4f1-1e22-4220-ac80-3b2d69d2db99/ovnkube-controller/0.log" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.629566 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" event={"ID":"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99","Type":"ContainerStarted","Data":"f973788e6cfc8d6032d0a82dd6dceec13bb5ebdd481636c561d03f469b8746dd"} Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.629774 4809 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.651097 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa4dd3f8-c119-46de-ad4e-dd29bb11a32c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa59e4ec9867cf280b308ee3adb5e303eaf94455aca588f5eb74a2b6263d57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5720f7f2977d9af6ed5324a30d08d34a53d55460951b3875b7d7d753d202f349\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dfd977b049a8ff81192d4041adb265366ab88475ee84c5750305d061d21c8db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d5801b61a87c95c9a623c3106e305920199c640efa8a32cf57698245697e54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:46Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.665767 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://010c2681c08dc21bc3a6fb82c04ab3dc2c26dcb76ec12b143c22845cac04ed25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:46Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.683611 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c981561d51870deffeb446e8ca769b7ac0a1a16edbe4e982e0f2c84a45043db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d58d489f532b78aca59f5b92584aad1dc292e6700c112c531cbeb3209fd90c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dcd421b38e5f9ba0c542a09f0047bde6c043942a096d11575afb7423ab2552c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab9b3bd4c8a6d6af7939b06fe01b8a0a50d81bb8c669bc6a27f486953c234e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a3e4453935871bb84d68bc4a8b6c98f299b7c8094ba82291bc90e258e1fec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5be86696b7a6518c5df930accb2c6890d58cfc37104fd34a359724b9e30231b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f973788e6cfc8d6032d0a82dd6dceec13bb5ebdd481636c561d03f469b8746dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ef6d4c42523249187a0998e293a702feb4e1aec2f31cade2e6f5a3cbfcb3ce5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:51:45Z\\\",\\\"message\\\":\\\"(0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 05:51:44.868917 6104 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 05:51:44.869052 6104 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 05:51:44.869150 6104 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 05:51:44.870031 6104 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 05:51:44.870070 6104 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 05:51:44.870083 6104 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 05:51:44.870088 6104 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 05:51:44.870102 6104 factory.go:656] Stopping watch factory\\\\nI1206 05:51:44.870114 6104 ovnkube.go:599] Stopped ovnkube\\\\nI1206 05:51:44.870135 6104 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 05:51:44.870134 6104 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 05:51:44.870148 6104 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dee8b965be3b2486f1f532f2bdfbbcbd84f9b3b711a11c25e86dae1284314d4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5xvvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:46Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.695363 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.695396 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.695406 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.695424 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.695435 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:46Z","lastTransitionTime":"2025-12-06T05:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.697584 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd45866c51410e910e87d43870bf531b0d05541e9164e700150fd37987e284af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:46Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.712251 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:46Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.727040 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:46Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.741484 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:46Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.752618 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947d89be9d04856af8d09facb8d4583f91898bd03a36d3ce8b9b74bc3bb795e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://930788ec9b4d97fdc39353a93111b84707037ce445f1930aa7737a20a44e1d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-npms2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:46Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.767826 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5636dd78-3d61-4c6a-b8fd-6d2457e19234\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:46Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.781649 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rnzq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0075197-9392-4109-b8a7-39507b15cc17\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53a00b17138b236dc8517cbd4ee71ea3ed7e8885247d06eb2e8840b8b145783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4mgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rnzq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:46Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.795010 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a48de155-7389-48db-8f86-d175b7db70c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://701274e49f68efbc0a939bc08771288c50662cef9c59da73170f7f114e222337\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9t7pq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:46Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.797713 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.797751 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.797765 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.797786 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.797799 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:46Z","lastTransitionTime":"2025-12-06T05:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.807475 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f28d0a3aec8a36273fa4480c552d80a5ee23210203861811e37f27bccda1147e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40791d93876595c4093814146d269b8fe70efb64ef2888e03c25aebef0e9295e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:46Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.816582 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hmlmc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c7558c-89b1-432d-893c-ccb2b3aba183\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c2f8876171399b61abd26285afba0547dac3c24c957bd6d202721b04c912aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm7wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hmlmc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:46Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.831093 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cb4vc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed42d686-41af-470e-910d-a1fabbec66b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f629b20d9f61227553520b30d0ca046750433926a8d31501e5fb77467426cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2nwpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cb4vc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:46Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.899850 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.899899 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.899914 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.899969 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:46 crc kubenswrapper[4809]: I1206 05:51:46.899991 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:46Z","lastTransitionTime":"2025-12-06T05:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.002588 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.002619 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.002627 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.002642 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.002650 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:47Z","lastTransitionTime":"2025-12-06T05:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.106702 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.106754 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.106767 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.106786 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.106798 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:47Z","lastTransitionTime":"2025-12-06T05:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.209576 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.209630 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.209645 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.209661 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.209674 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:47Z","lastTransitionTime":"2025-12-06T05:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.312271 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.312309 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.312326 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.312343 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.312362 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:47Z","lastTransitionTime":"2025-12-06T05:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.388550 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.388633 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:51:47 crc kubenswrapper[4809]: E1206 05:51:47.388676 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:51:47 crc kubenswrapper[4809]: E1206 05:51:47.388765 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.415006 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.415053 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.415064 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.415080 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.415093 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:47Z","lastTransitionTime":"2025-12-06T05:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.517529 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.517573 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.517584 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.517601 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.517612 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:47Z","lastTransitionTime":"2025-12-06T05:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.624580 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.624625 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.624636 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.624657 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.624668 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:47Z","lastTransitionTime":"2025-12-06T05:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.633044 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5xvvb_ffb3a4f1-1e22-4220-ac80-3b2d69d2db99/ovnkube-controller/1.log" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.633523 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5xvvb_ffb3a4f1-1e22-4220-ac80-3b2d69d2db99/ovnkube-controller/0.log" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.635717 4809 generic.go:334] "Generic (PLEG): container finished" podID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerID="f973788e6cfc8d6032d0a82dd6dceec13bb5ebdd481636c561d03f469b8746dd" exitCode=1 Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.635751 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" event={"ID":"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99","Type":"ContainerDied","Data":"f973788e6cfc8d6032d0a82dd6dceec13bb5ebdd481636c561d03f469b8746dd"} Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.635792 4809 scope.go:117] "RemoveContainer" containerID="9ef6d4c42523249187a0998e293a702feb4e1aec2f31cade2e6f5a3cbfcb3ce5" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.636539 4809 scope.go:117] "RemoveContainer" containerID="f973788e6cfc8d6032d0a82dd6dceec13bb5ebdd481636c561d03f469b8746dd" Dec 06 05:51:47 crc kubenswrapper[4809]: E1206 05:51:47.636683 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5xvvb_openshift-ovn-kubernetes(ffb3a4f1-1e22-4220-ac80-3b2d69d2db99)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.649145 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cb4vc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed42d686-41af-470e-910d-a1fabbec66b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f629b20d9f61227553520b30d0ca046750433926a8d31501e5fb77467426cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2nwpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cb4vc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:47Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.660213 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hmlmc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c7558c-89b1-432d-893c-ccb2b3aba183\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c2f8876171399b61abd26285afba0547dac3c24c957bd6d202721b04c912aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm7wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hmlmc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:47Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.671865 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa4dd3f8-c119-46de-ad4e-dd29bb11a32c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa59e4ec9867cf280b308ee3adb5e303eaf94455aca588f5eb74a2b6263d57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5720f7f2977d9af6ed5324a30d08d34a53d55460951b3875b7d7d753d202f349\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dfd977b049a8ff81192d4041adb265366ab88475ee84c5750305d061d21c8db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d5801b61a87c95c9a623c3106e305920199c640efa8a32cf57698245697e54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:47Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.682392 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://010c2681c08dc21bc3a6fb82c04ab3dc2c26dcb76ec12b143c22845cac04ed25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:47Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.700018 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c981561d51870deffeb446e8ca769b7ac0a1a16edbe4e982e0f2c84a45043db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d58d489f532b78aca59f5b92584aad1dc292e6700c112c531cbeb3209fd90c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dcd421b38e5f9ba0c542a09f0047bde6c043942a096d11575afb7423ab2552c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab9b3bd4c8a6d6af7939b06fe01b8a0a50d81bb8c669bc6a27f486953c234e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a3e4453935871bb84d68bc4a8b6c98f299b7c8094ba82291bc90e258e1fec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5be86696b7a6518c5df930accb2c6890d58cfc37104fd34a359724b9e30231b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f973788e6cfc8d6032d0a82dd6dceec13bb5ebdd481636c561d03f469b8746dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ef6d4c42523249187a0998e293a702feb4e1aec2f31cade2e6f5a3cbfcb3ce5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:51:45Z\\\",\\\"message\\\":\\\"(0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 05:51:44.868917 6104 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 05:51:44.869052 6104 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 05:51:44.869150 6104 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 05:51:44.870031 6104 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 05:51:44.870070 6104 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 05:51:44.870083 6104 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 05:51:44.870088 6104 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 05:51:44.870102 6104 factory.go:656] Stopping watch factory\\\\nI1206 05:51:44.870114 6104 ovnkube.go:599] Stopped ovnkube\\\\nI1206 05:51:44.870135 6104 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 05:51:44.870134 6104 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 05:51:44.870148 6104 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f973788e6cfc8d6032d0a82dd6dceec13bb5ebdd481636c561d03f469b8746dd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:51:47Z\\\",\\\"message\\\":\\\"ackage-server-manager-metrics for network=default are: map[]\\\\nI1206 05:51:46.633808 6223 services_controller.go:443] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.110\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1206 05:51:46.633876 6223 services_controller.go:444] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1206 05:51:46.633966 6223 services_controller.go:445] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nI1206 05:51:46.633753 6223 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-operators\\\\\\\"}\\\\nI1206 05:51:46.634043 6223 services_controller.go:360] Finished syncing service redhat-operators on namespace openshift-marketplace for network=default : 1.127407ms\\\\nF1206 05:51:46.633754 6223 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dee8b965be3b2486f1f532f2bdfbbcbd84f9b3b711a11c25e86dae1284314d4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5xvvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:47Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.709529 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947d89be9d04856af8d09facb8d4583f91898bd03a36d3ce8b9b74bc3bb795e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://930788ec9b4d97fdc39353a93111b84707037ce445f1930aa7737a20a44e1d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-npms2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:47Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.722192 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5636dd78-3d61-4c6a-b8fd-6d2457e19234\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:47Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.727304 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.727373 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.727395 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.727423 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.727443 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:47Z","lastTransitionTime":"2025-12-06T05:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.733767 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd45866c51410e910e87d43870bf531b0d05541e9164e700150fd37987e284af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:47Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.747312 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:47Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.760821 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:47Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.774312 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:47Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.784177 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rnzq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0075197-9392-4109-b8a7-39507b15cc17\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53a00b17138b236dc8517cbd4ee71ea3ed7e8885247d06eb2e8840b8b145783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4mgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rnzq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:47Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.797620 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f28d0a3aec8a36273fa4480c552d80a5ee23210203861811e37f27bccda1147e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40791d93876595c4093814146d269b8fe70efb64ef2888e03c25aebef0e9295e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:47Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.810922 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a48de155-7389-48db-8f86-d175b7db70c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://701274e49f68efbc0a939bc08771288c50662cef9c59da73170f7f114e222337\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9t7pq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:47Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.830204 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.830267 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.830277 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.830297 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.830327 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:47Z","lastTransitionTime":"2025-12-06T05:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.933210 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.933289 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.933318 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.933346 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:47 crc kubenswrapper[4809]: I1206 05:51:47.933369 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:47Z","lastTransitionTime":"2025-12-06T05:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.037080 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.037118 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.037130 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.037146 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.037160 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:48Z","lastTransitionTime":"2025-12-06T05:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.139165 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.139199 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.139208 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.139224 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.139233 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:48Z","lastTransitionTime":"2025-12-06T05:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.209422 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nndh4"] Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.210284 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nndh4" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.213792 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.213817 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.224235 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa4dd3f8-c119-46de-ad4e-dd29bb11a32c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa59e4ec9867cf280b308ee3adb5e303eaf94455aca588f5eb74a2b6263d57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5720f7f2977d9af6ed5324a30d08d34a53d55460951b3875b7d7d753d202f349\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dfd977b049a8ff81192d4041adb265366ab88475ee84c5750305d061d21c8db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d5801b61a87c95c9a623c3106e305920199c640efa8a32cf57698245697e54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:48Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.242438 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.242518 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.242617 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.242642 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.242653 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:48Z","lastTransitionTime":"2025-12-06T05:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.243790 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://010c2681c08dc21bc3a6fb82c04ab3dc2c26dcb76ec12b143c22845cac04ed25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:48Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.250636 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a80ba5e1-90cc-44ce-be19-fdf0c007d5b8-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-nndh4\" (UID: \"a80ba5e1-90cc-44ce-be19-fdf0c007d5b8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nndh4" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.250714 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a80ba5e1-90cc-44ce-be19-fdf0c007d5b8-env-overrides\") pod \"ovnkube-control-plane-749d76644c-nndh4\" (UID: \"a80ba5e1-90cc-44ce-be19-fdf0c007d5b8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nndh4" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.250764 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a80ba5e1-90cc-44ce-be19-fdf0c007d5b8-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-nndh4\" (UID: \"a80ba5e1-90cc-44ce-be19-fdf0c007d5b8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nndh4" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.250803 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2c22k\" (UniqueName: \"kubernetes.io/projected/a80ba5e1-90cc-44ce-be19-fdf0c007d5b8-kube-api-access-2c22k\") pod \"ovnkube-control-plane-749d76644c-nndh4\" (UID: \"a80ba5e1-90cc-44ce-be19-fdf0c007d5b8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nndh4" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.263688 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c981561d51870deffeb446e8ca769b7ac0a1a16edbe4e982e0f2c84a45043db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d58d489f532b78aca59f5b92584aad1dc292e6700c112c531cbeb3209fd90c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dcd421b38e5f9ba0c542a09f0047bde6c043942a096d11575afb7423ab2552c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab9b3bd4c8a6d6af7939b06fe01b8a0a50d81bb8c669bc6a27f486953c234e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a3e4453935871bb84d68bc4a8b6c98f299b7c8094ba82291bc90e258e1fec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5be86696b7a6518c5df930accb2c6890d58cfc37104fd34a359724b9e30231b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f973788e6cfc8d6032d0a82dd6dceec13bb5ebdd481636c561d03f469b8746dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ef6d4c42523249187a0998e293a702feb4e1aec2f31cade2e6f5a3cbfcb3ce5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:51:45Z\\\",\\\"message\\\":\\\"(0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 05:51:44.868917 6104 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 05:51:44.869052 6104 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 05:51:44.869150 6104 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 05:51:44.870031 6104 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 05:51:44.870070 6104 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 05:51:44.870083 6104 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 05:51:44.870088 6104 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 05:51:44.870102 6104 factory.go:656] Stopping watch factory\\\\nI1206 05:51:44.870114 6104 ovnkube.go:599] Stopped ovnkube\\\\nI1206 05:51:44.870135 6104 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 05:51:44.870134 6104 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 05:51:44.870148 6104 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f973788e6cfc8d6032d0a82dd6dceec13bb5ebdd481636c561d03f469b8746dd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:51:47Z\\\",\\\"message\\\":\\\"ackage-server-manager-metrics for network=default are: map[]\\\\nI1206 05:51:46.633808 6223 services_controller.go:443] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.110\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1206 05:51:46.633876 6223 services_controller.go:444] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1206 05:51:46.633966 6223 services_controller.go:445] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nI1206 05:51:46.633753 6223 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-operators\\\\\\\"}\\\\nI1206 05:51:46.634043 6223 services_controller.go:360] Finished syncing service redhat-operators on namespace openshift-marketplace for network=default : 1.127407ms\\\\nF1206 05:51:46.633754 6223 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dee8b965be3b2486f1f532f2bdfbbcbd84f9b3b711a11c25e86dae1284314d4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5xvvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:48Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.275964 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:48Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.286540 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:48Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.295061 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947d89be9d04856af8d09facb8d4583f91898bd03a36d3ce8b9b74bc3bb795e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://930788ec9b4d97fdc39353a93111b84707037ce445f1930aa7737a20a44e1d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-npms2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:48Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.306300 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5636dd78-3d61-4c6a-b8fd-6d2457e19234\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:48Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.317278 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd45866c51410e910e87d43870bf531b0d05541e9164e700150fd37987e284af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:48Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.328456 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:48Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.336914 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rnzq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0075197-9392-4109-b8a7-39507b15cc17\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53a00b17138b236dc8517cbd4ee71ea3ed7e8885247d06eb2e8840b8b145783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4mgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rnzq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:48Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.345472 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.345536 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.345549 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.345570 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.345582 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:48Z","lastTransitionTime":"2025-12-06T05:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.346848 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f28d0a3aec8a36273fa4480c552d80a5ee23210203861811e37f27bccda1147e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40791d93876595c4093814146d269b8fe70efb64ef2888e03c25aebef0e9295e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:48Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.351873 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a80ba5e1-90cc-44ce-be19-fdf0c007d5b8-env-overrides\") pod \"ovnkube-control-plane-749d76644c-nndh4\" (UID: \"a80ba5e1-90cc-44ce-be19-fdf0c007d5b8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nndh4" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.351955 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a80ba5e1-90cc-44ce-be19-fdf0c007d5b8-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-nndh4\" (UID: \"a80ba5e1-90cc-44ce-be19-fdf0c007d5b8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nndh4" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.351995 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2c22k\" (UniqueName: \"kubernetes.io/projected/a80ba5e1-90cc-44ce-be19-fdf0c007d5b8-kube-api-access-2c22k\") pod \"ovnkube-control-plane-749d76644c-nndh4\" (UID: \"a80ba5e1-90cc-44ce-be19-fdf0c007d5b8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nndh4" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.352023 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a80ba5e1-90cc-44ce-be19-fdf0c007d5b8-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-nndh4\" (UID: \"a80ba5e1-90cc-44ce-be19-fdf0c007d5b8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nndh4" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.353252 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a80ba5e1-90cc-44ce-be19-fdf0c007d5b8-env-overrides\") pod \"ovnkube-control-plane-749d76644c-nndh4\" (UID: \"a80ba5e1-90cc-44ce-be19-fdf0c007d5b8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nndh4" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.353514 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a80ba5e1-90cc-44ce-be19-fdf0c007d5b8-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-nndh4\" (UID: \"a80ba5e1-90cc-44ce-be19-fdf0c007d5b8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nndh4" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.360513 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a80ba5e1-90cc-44ce-be19-fdf0c007d5b8-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-nndh4\" (UID: \"a80ba5e1-90cc-44ce-be19-fdf0c007d5b8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nndh4" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.366639 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2c22k\" (UniqueName: \"kubernetes.io/projected/a80ba5e1-90cc-44ce-be19-fdf0c007d5b8-kube-api-access-2c22k\") pod \"ovnkube-control-plane-749d76644c-nndh4\" (UID: \"a80ba5e1-90cc-44ce-be19-fdf0c007d5b8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nndh4" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.366978 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a48de155-7389-48db-8f86-d175b7db70c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://701274e49f68efbc0a939bc08771288c50662cef9c59da73170f7f114e222337\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9t7pq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:48Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.378774 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nndh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a80ba5e1-90cc-44ce-be19-fdf0c007d5b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2c22k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2c22k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nndh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:48Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.388078 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:51:48 crc kubenswrapper[4809]: E1206 05:51:48.388182 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.390639 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cb4vc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed42d686-41af-470e-910d-a1fabbec66b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f629b20d9f61227553520b30d0ca046750433926a8d31501e5fb77467426cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2nwpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cb4vc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:48Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.399398 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hmlmc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c7558c-89b1-432d-893c-ccb2b3aba183\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c2f8876171399b61abd26285afba0547dac3c24c957bd6d202721b04c912aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm7wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hmlmc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:48Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.448373 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.448440 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.448452 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.448468 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.448478 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:48Z","lastTransitionTime":"2025-12-06T05:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.525373 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nndh4" Dec 06 05:51:48 crc kubenswrapper[4809]: W1206 05:51:48.539245 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda80ba5e1_90cc_44ce_be19_fdf0c007d5b8.slice/crio-0685e6a7f185709cc307d75b9dea14fb900b6f9051bbdd9ecf10ec22f35ddccd WatchSource:0}: Error finding container 0685e6a7f185709cc307d75b9dea14fb900b6f9051bbdd9ecf10ec22f35ddccd: Status 404 returned error can't find the container with id 0685e6a7f185709cc307d75b9dea14fb900b6f9051bbdd9ecf10ec22f35ddccd Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.550576 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.550633 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.550649 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.550671 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.550684 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:48Z","lastTransitionTime":"2025-12-06T05:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.639753 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nndh4" event={"ID":"a80ba5e1-90cc-44ce-be19-fdf0c007d5b8","Type":"ContainerStarted","Data":"0685e6a7f185709cc307d75b9dea14fb900b6f9051bbdd9ecf10ec22f35ddccd"} Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.640591 4809 scope.go:117] "RemoveContainer" containerID="f973788e6cfc8d6032d0a82dd6dceec13bb5ebdd481636c561d03f469b8746dd" Dec 06 05:51:48 crc kubenswrapper[4809]: E1206 05:51:48.640731 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5xvvb_openshift-ovn-kubernetes(ffb3a4f1-1e22-4220-ac80-3b2d69d2db99)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.655007 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cb4vc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed42d686-41af-470e-910d-a1fabbec66b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f629b20d9f61227553520b30d0ca046750433926a8d31501e5fb77467426cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2nwpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cb4vc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:48Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.655277 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.655512 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.655520 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.655534 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.655560 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:48Z","lastTransitionTime":"2025-12-06T05:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.666887 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hmlmc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c7558c-89b1-432d-893c-ccb2b3aba183\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c2f8876171399b61abd26285afba0547dac3c24c957bd6d202721b04c912aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm7wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hmlmc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:48Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.679990 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa4dd3f8-c119-46de-ad4e-dd29bb11a32c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa59e4ec9867cf280b308ee3adb5e303eaf94455aca588f5eb74a2b6263d57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5720f7f2977d9af6ed5324a30d08d34a53d55460951b3875b7d7d753d202f349\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dfd977b049a8ff81192d4041adb265366ab88475ee84c5750305d061d21c8db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d5801b61a87c95c9a623c3106e305920199c640efa8a32cf57698245697e54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:48Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.702360 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://010c2681c08dc21bc3a6fb82c04ab3dc2c26dcb76ec12b143c22845cac04ed25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:48Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.720912 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c981561d51870deffeb446e8ca769b7ac0a1a16edbe4e982e0f2c84a45043db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d58d489f532b78aca59f5b92584aad1dc292e6700c112c531cbeb3209fd90c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dcd421b38e5f9ba0c542a09f0047bde6c043942a096d11575afb7423ab2552c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab9b3bd4c8a6d6af7939b06fe01b8a0a50d81bb8c669bc6a27f486953c234e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a3e4453935871bb84d68bc4a8b6c98f299b7c8094ba82291bc90e258e1fec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5be86696b7a6518c5df930accb2c6890d58cfc37104fd34a359724b9e30231b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f973788e6cfc8d6032d0a82dd6dceec13bb5ebdd481636c561d03f469b8746dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f973788e6cfc8d6032d0a82dd6dceec13bb5ebdd481636c561d03f469b8746dd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:51:47Z\\\",\\\"message\\\":\\\"ackage-server-manager-metrics for network=default are: map[]\\\\nI1206 05:51:46.633808 6223 services_controller.go:443] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.110\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1206 05:51:46.633876 6223 services_controller.go:444] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1206 05:51:46.633966 6223 services_controller.go:445] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nI1206 05:51:46.633753 6223 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-operators\\\\\\\"}\\\\nI1206 05:51:46.634043 6223 services_controller.go:360] Finished syncing service redhat-operators on namespace openshift-marketplace for network=default : 1.127407ms\\\\nF1206 05:51:46.633754 6223 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5xvvb_openshift-ovn-kubernetes(ffb3a4f1-1e22-4220-ac80-3b2d69d2db99)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dee8b965be3b2486f1f532f2bdfbbcbd84f9b3b711a11c25e86dae1284314d4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5xvvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:48Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.735039 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:48Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.746647 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:48Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.757233 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.757269 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.757277 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.757297 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.757306 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:48Z","lastTransitionTime":"2025-12-06T05:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.759028 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:48Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.768674 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947d89be9d04856af8d09facb8d4583f91898bd03a36d3ce8b9b74bc3bb795e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://930788ec9b4d97fdc39353a93111b84707037ce445f1930aa7737a20a44e1d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-npms2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:48Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.781879 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5636dd78-3d61-4c6a-b8fd-6d2457e19234\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:48Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.792528 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd45866c51410e910e87d43870bf531b0d05541e9164e700150fd37987e284af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:48Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.801725 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rnzq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0075197-9392-4109-b8a7-39507b15cc17\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53a00b17138b236dc8517cbd4ee71ea3ed7e8885247d06eb2e8840b8b145783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4mgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rnzq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:48Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.813969 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f28d0a3aec8a36273fa4480c552d80a5ee23210203861811e37f27bccda1147e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40791d93876595c4093814146d269b8fe70efb64ef2888e03c25aebef0e9295e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:48Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.826816 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a48de155-7389-48db-8f86-d175b7db70c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://701274e49f68efbc0a939bc08771288c50662cef9c59da73170f7f114e222337\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9t7pq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:48Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.836858 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nndh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a80ba5e1-90cc-44ce-be19-fdf0c007d5b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2c22k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2c22k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nndh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:48Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.859432 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.859473 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.859482 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.859496 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.859505 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:48Z","lastTransitionTime":"2025-12-06T05:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.864637 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.864689 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.864702 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.864720 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.864731 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:48Z","lastTransitionTime":"2025-12-06T05:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:48 crc kubenswrapper[4809]: E1206 05:51:48.876785 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"88a28b6b-9c85-4932-b5bc-36346303aee5\\\",\\\"systemUUID\\\":\\\"5931df91-c1df-4f78-9e5f-391a01a76127\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:48Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.880023 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.880063 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.880073 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.880088 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.880097 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:48Z","lastTransitionTime":"2025-12-06T05:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:48 crc kubenswrapper[4809]: E1206 05:51:48.893436 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"88a28b6b-9c85-4932-b5bc-36346303aee5\\\",\\\"systemUUID\\\":\\\"5931df91-c1df-4f78-9e5f-391a01a76127\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:48Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.896979 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.897003 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.897012 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.897024 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.897034 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:48Z","lastTransitionTime":"2025-12-06T05:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:48 crc kubenswrapper[4809]: E1206 05:51:48.912032 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"88a28b6b-9c85-4932-b5bc-36346303aee5\\\",\\\"systemUUID\\\":\\\"5931df91-c1df-4f78-9e5f-391a01a76127\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:48Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.915392 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.915426 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.915439 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.915459 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.915471 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:48Z","lastTransitionTime":"2025-12-06T05:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:48 crc kubenswrapper[4809]: E1206 05:51:48.930552 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"88a28b6b-9c85-4932-b5bc-36346303aee5\\\",\\\"systemUUID\\\":\\\"5931df91-c1df-4f78-9e5f-391a01a76127\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:48Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.933758 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.933823 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.933839 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.933860 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.933872 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:48Z","lastTransitionTime":"2025-12-06T05:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:48 crc kubenswrapper[4809]: E1206 05:51:48.948234 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"88a28b6b-9c85-4932-b5bc-36346303aee5\\\",\\\"systemUUID\\\":\\\"5931df91-c1df-4f78-9e5f-391a01a76127\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:48Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:48 crc kubenswrapper[4809]: E1206 05:51:48.948457 4809 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.951048 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-9k8zx"] Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.951896 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:51:48 crc kubenswrapper[4809]: E1206 05:51:48.952045 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.970431 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.970667 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.970760 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.970842 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.970915 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:48Z","lastTransitionTime":"2025-12-06T05:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.978806 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:48Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:48 crc kubenswrapper[4809]: I1206 05:51:48.991016 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947d89be9d04856af8d09facb8d4583f91898bd03a36d3ce8b9b74bc3bb795e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://930788ec9b4d97fdc39353a93111b84707037ce445f1930aa7737a20a44e1d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-npms2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:48Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.009625 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5636dd78-3d61-4c6a-b8fd-6d2457e19234\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:49Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.026680 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd45866c51410e910e87d43870bf531b0d05541e9164e700150fd37987e284af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:49Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.047239 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:49Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.059212 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpdh8\" (UniqueName: \"kubernetes.io/projected/a4f75fd3-03e8-41e7-8926-f6ff21a5b681-kube-api-access-kpdh8\") pod \"network-metrics-daemon-9k8zx\" (UID: \"a4f75fd3-03e8-41e7-8926-f6ff21a5b681\") " pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.059318 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a4f75fd3-03e8-41e7-8926-f6ff21a5b681-metrics-certs\") pod \"network-metrics-daemon-9k8zx\" (UID: \"a4f75fd3-03e8-41e7-8926-f6ff21a5b681\") " pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.065118 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:49Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.074368 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.074406 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.074418 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.074434 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.074445 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:49Z","lastTransitionTime":"2025-12-06T05:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.078207 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rnzq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0075197-9392-4109-b8a7-39507b15cc17\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53a00b17138b236dc8517cbd4ee71ea3ed7e8885247d06eb2e8840b8b145783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4mgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rnzq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:49Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.094769 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9k8zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4f75fd3-03e8-41e7-8926-f6ff21a5b681\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpdh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpdh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9k8zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:49Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.110083 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f28d0a3aec8a36273fa4480c552d80a5ee23210203861811e37f27bccda1147e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40791d93876595c4093814146d269b8fe70efb64ef2888e03c25aebef0e9295e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:49Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.131071 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a48de155-7389-48db-8f86-d175b7db70c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://701274e49f68efbc0a939bc08771288c50662cef9c59da73170f7f114e222337\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9t7pq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:49Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.147237 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nndh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a80ba5e1-90cc-44ce-be19-fdf0c007d5b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2c22k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2c22k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nndh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:49Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.161530 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpdh8\" (UniqueName: \"kubernetes.io/projected/a4f75fd3-03e8-41e7-8926-f6ff21a5b681-kube-api-access-kpdh8\") pod \"network-metrics-daemon-9k8zx\" (UID: \"a4f75fd3-03e8-41e7-8926-f6ff21a5b681\") " pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.162839 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a4f75fd3-03e8-41e7-8926-f6ff21a5b681-metrics-certs\") pod \"network-metrics-daemon-9k8zx\" (UID: \"a4f75fd3-03e8-41e7-8926-f6ff21a5b681\") " pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:51:49 crc kubenswrapper[4809]: E1206 05:51:49.164045 4809 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 05:51:49 crc kubenswrapper[4809]: E1206 05:51:49.164349 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a4f75fd3-03e8-41e7-8926-f6ff21a5b681-metrics-certs podName:a4f75fd3-03e8-41e7-8926-f6ff21a5b681 nodeName:}" failed. No retries permitted until 2025-12-06 05:51:49.664320362 +0000 UTC m=+34.553303304 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a4f75fd3-03e8-41e7-8926-f6ff21a5b681-metrics-certs") pod "network-metrics-daemon-9k8zx" (UID: "a4f75fd3-03e8-41e7-8926-f6ff21a5b681") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.170081 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cb4vc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed42d686-41af-470e-910d-a1fabbec66b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f629b20d9f61227553520b30d0ca046750433926a8d31501e5fb77467426cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2nwpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cb4vc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:49Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.179453 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.179503 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.179511 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.179523 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.179532 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:49Z","lastTransitionTime":"2025-12-06T05:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.180686 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hmlmc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c7558c-89b1-432d-893c-ccb2b3aba183\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c2f8876171399b61abd26285afba0547dac3c24c957bd6d202721b04c912aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm7wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hmlmc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:49Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.183598 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpdh8\" (UniqueName: \"kubernetes.io/projected/a4f75fd3-03e8-41e7-8926-f6ff21a5b681-kube-api-access-kpdh8\") pod \"network-metrics-daemon-9k8zx\" (UID: \"a4f75fd3-03e8-41e7-8926-f6ff21a5b681\") " pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.192209 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa4dd3f8-c119-46de-ad4e-dd29bb11a32c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa59e4ec9867cf280b308ee3adb5e303eaf94455aca588f5eb74a2b6263d57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5720f7f2977d9af6ed5324a30d08d34a53d55460951b3875b7d7d753d202f349\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dfd977b049a8ff81192d4041adb265366ab88475ee84c5750305d061d21c8db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d5801b61a87c95c9a623c3106e305920199c640efa8a32cf57698245697e54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:49Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.203208 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://010c2681c08dc21bc3a6fb82c04ab3dc2c26dcb76ec12b143c22845cac04ed25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:49Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.219604 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c981561d51870deffeb446e8ca769b7ac0a1a16edbe4e982e0f2c84a45043db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d58d489f532b78aca59f5b92584aad1dc292e6700c112c531cbeb3209fd90c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dcd421b38e5f9ba0c542a09f0047bde6c043942a096d11575afb7423ab2552c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab9b3bd4c8a6d6af7939b06fe01b8a0a50d81bb8c669bc6a27f486953c234e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a3e4453935871bb84d68bc4a8b6c98f299b7c8094ba82291bc90e258e1fec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5be86696b7a6518c5df930accb2c6890d58cfc37104fd34a359724b9e30231b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f973788e6cfc8d6032d0a82dd6dceec13bb5ebdd481636c561d03f469b8746dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f973788e6cfc8d6032d0a82dd6dceec13bb5ebdd481636c561d03f469b8746dd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:51:47Z\\\",\\\"message\\\":\\\"ackage-server-manager-metrics for network=default are: map[]\\\\nI1206 05:51:46.633808 6223 services_controller.go:443] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.110\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1206 05:51:46.633876 6223 services_controller.go:444] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1206 05:51:46.633966 6223 services_controller.go:445] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nI1206 05:51:46.633753 6223 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-operators\\\\\\\"}\\\\nI1206 05:51:46.634043 6223 services_controller.go:360] Finished syncing service redhat-operators on namespace openshift-marketplace for network=default : 1.127407ms\\\\nF1206 05:51:46.633754 6223 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5xvvb_openshift-ovn-kubernetes(ffb3a4f1-1e22-4220-ac80-3b2d69d2db99)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dee8b965be3b2486f1f532f2bdfbbcbd84f9b3b711a11c25e86dae1284314d4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5xvvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:49Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.282232 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.282294 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.282306 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.282323 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.282635 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:49Z","lastTransitionTime":"2025-12-06T05:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.384835 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.384907 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.384920 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.384963 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.384976 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:49Z","lastTransitionTime":"2025-12-06T05:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.388254 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.388331 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:51:49 crc kubenswrapper[4809]: E1206 05:51:49.388409 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:51:49 crc kubenswrapper[4809]: E1206 05:51:49.388870 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.489071 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.489132 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.489150 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.489174 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.489192 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:49Z","lastTransitionTime":"2025-12-06T05:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.591213 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.591252 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.591262 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.591275 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.591287 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:49Z","lastTransitionTime":"2025-12-06T05:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.669618 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a4f75fd3-03e8-41e7-8926-f6ff21a5b681-metrics-certs\") pod \"network-metrics-daemon-9k8zx\" (UID: \"a4f75fd3-03e8-41e7-8926-f6ff21a5b681\") " pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:51:49 crc kubenswrapper[4809]: E1206 05:51:49.669841 4809 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 05:51:49 crc kubenswrapper[4809]: E1206 05:51:49.669923 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a4f75fd3-03e8-41e7-8926-f6ff21a5b681-metrics-certs podName:a4f75fd3-03e8-41e7-8926-f6ff21a5b681 nodeName:}" failed. No retries permitted until 2025-12-06 05:51:50.669900777 +0000 UTC m=+35.558883749 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a4f75fd3-03e8-41e7-8926-f6ff21a5b681-metrics-certs") pod "network-metrics-daemon-9k8zx" (UID: "a4f75fd3-03e8-41e7-8926-f6ff21a5b681") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.693909 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.693997 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.694015 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.694035 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.694052 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:49Z","lastTransitionTime":"2025-12-06T05:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.795922 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.796197 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.796208 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.796223 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.796232 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:49Z","lastTransitionTime":"2025-12-06T05:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.898840 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.898871 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.898879 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.898893 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:49 crc kubenswrapper[4809]: I1206 05:51:49.898904 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:49Z","lastTransitionTime":"2025-12-06T05:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.001093 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.001133 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.001144 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.001159 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.001170 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:50Z","lastTransitionTime":"2025-12-06T05:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.103279 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.103343 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.103356 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.103372 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.103384 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:50Z","lastTransitionTime":"2025-12-06T05:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.205809 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.205840 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.205850 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.205864 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.205872 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:50Z","lastTransitionTime":"2025-12-06T05:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.308881 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.308955 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.308967 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.308985 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.308998 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:50Z","lastTransitionTime":"2025-12-06T05:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.388223 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:51:50 crc kubenswrapper[4809]: E1206 05:51:50.388353 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.388495 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:51:50 crc kubenswrapper[4809]: E1206 05:51:50.388668 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.411796 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.411839 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.411850 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.411864 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.411877 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:50Z","lastTransitionTime":"2025-12-06T05:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.515029 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.515095 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.515118 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.515150 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.515176 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:50Z","lastTransitionTime":"2025-12-06T05:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.618800 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.618850 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.618860 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.618891 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.618902 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:50Z","lastTransitionTime":"2025-12-06T05:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.650534 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5xvvb_ffb3a4f1-1e22-4220-ac80-3b2d69d2db99/ovnkube-controller/1.log" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.654767 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nndh4" event={"ID":"a80ba5e1-90cc-44ce-be19-fdf0c007d5b8","Type":"ContainerStarted","Data":"b31d74665ae370eed09b8a3195bd3d122ad64c6e33bf55fc42cd40a862671b8c"} Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.654802 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nndh4" event={"ID":"a80ba5e1-90cc-44ce-be19-fdf0c007d5b8","Type":"ContainerStarted","Data":"d3684c4e5e839ce117a1772571ae080042e0187ea1aaca635edca07ac10d94c6"} Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.667194 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f28d0a3aec8a36273fa4480c552d80a5ee23210203861811e37f27bccda1147e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40791d93876595c4093814146d269b8fe70efb64ef2888e03c25aebef0e9295e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.682027 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a48de155-7389-48db-8f86-d175b7db70c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://701274e49f68efbc0a939bc08771288c50662cef9c59da73170f7f114e222337\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9t7pq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.682261 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a4f75fd3-03e8-41e7-8926-f6ff21a5b681-metrics-certs\") pod \"network-metrics-daemon-9k8zx\" (UID: \"a4f75fd3-03e8-41e7-8926-f6ff21a5b681\") " pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:51:50 crc kubenswrapper[4809]: E1206 05:51:50.682580 4809 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 05:51:50 crc kubenswrapper[4809]: E1206 05:51:50.682743 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a4f75fd3-03e8-41e7-8926-f6ff21a5b681-metrics-certs podName:a4f75fd3-03e8-41e7-8926-f6ff21a5b681 nodeName:}" failed. No retries permitted until 2025-12-06 05:51:52.682708416 +0000 UTC m=+37.571691418 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a4f75fd3-03e8-41e7-8926-f6ff21a5b681-metrics-certs") pod "network-metrics-daemon-9k8zx" (UID: "a4f75fd3-03e8-41e7-8926-f6ff21a5b681") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.703063 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nndh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a80ba5e1-90cc-44ce-be19-fdf0c007d5b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3684c4e5e839ce117a1772571ae080042e0187ea1aaca635edca07ac10d94c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2c22k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31d74665ae370eed09b8a3195bd3d122ad64c6e33bf55fc42cd40a862671b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2c22k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nndh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.716051 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cb4vc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed42d686-41af-470e-910d-a1fabbec66b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f629b20d9f61227553520b30d0ca046750433926a8d31501e5fb77467426cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2nwpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cb4vc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.720767 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.720808 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.720819 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.720835 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.720847 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:50Z","lastTransitionTime":"2025-12-06T05:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.727746 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hmlmc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c7558c-89b1-432d-893c-ccb2b3aba183\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c2f8876171399b61abd26285afba0547dac3c24c957bd6d202721b04c912aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm7wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hmlmc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.738582 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa4dd3f8-c119-46de-ad4e-dd29bb11a32c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa59e4ec9867cf280b308ee3adb5e303eaf94455aca588f5eb74a2b6263d57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5720f7f2977d9af6ed5324a30d08d34a53d55460951b3875b7d7d753d202f349\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dfd977b049a8ff81192d4041adb265366ab88475ee84c5750305d061d21c8db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d5801b61a87c95c9a623c3106e305920199c640efa8a32cf57698245697e54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.748407 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://010c2681c08dc21bc3a6fb82c04ab3dc2c26dcb76ec12b143c22845cac04ed25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.763755 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c981561d51870deffeb446e8ca769b7ac0a1a16edbe4e982e0f2c84a45043db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d58d489f532b78aca59f5b92584aad1dc292e6700c112c531cbeb3209fd90c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dcd421b38e5f9ba0c542a09f0047bde6c043942a096d11575afb7423ab2552c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab9b3bd4c8a6d6af7939b06fe01b8a0a50d81bb8c669bc6a27f486953c234e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a3e4453935871bb84d68bc4a8b6c98f299b7c8094ba82291bc90e258e1fec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5be86696b7a6518c5df930accb2c6890d58cfc37104fd34a359724b9e30231b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f973788e6cfc8d6032d0a82dd6dceec13bb5ebdd481636c561d03f469b8746dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f973788e6cfc8d6032d0a82dd6dceec13bb5ebdd481636c561d03f469b8746dd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:51:47Z\\\",\\\"message\\\":\\\"ackage-server-manager-metrics for network=default are: map[]\\\\nI1206 05:51:46.633808 6223 services_controller.go:443] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.110\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1206 05:51:46.633876 6223 services_controller.go:444] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1206 05:51:46.633966 6223 services_controller.go:445] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nI1206 05:51:46.633753 6223 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-operators\\\\\\\"}\\\\nI1206 05:51:46.634043 6223 services_controller.go:360] Finished syncing service redhat-operators on namespace openshift-marketplace for network=default : 1.127407ms\\\\nF1206 05:51:46.633754 6223 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5xvvb_openshift-ovn-kubernetes(ffb3a4f1-1e22-4220-ac80-3b2d69d2db99)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dee8b965be3b2486f1f532f2bdfbbcbd84f9b3b711a11c25e86dae1284314d4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5xvvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.774915 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.788846 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.801112 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947d89be9d04856af8d09facb8d4583f91898bd03a36d3ce8b9b74bc3bb795e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://930788ec9b4d97fdc39353a93111b84707037ce445f1930aa7737a20a44e1d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-npms2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.818683 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5636dd78-3d61-4c6a-b8fd-6d2457e19234\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.823504 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.823536 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.823549 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.823564 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.823577 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:50Z","lastTransitionTime":"2025-12-06T05:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.832472 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd45866c51410e910e87d43870bf531b0d05541e9164e700150fd37987e284af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.844273 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.852059 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rnzq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0075197-9392-4109-b8a7-39507b15cc17\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53a00b17138b236dc8517cbd4ee71ea3ed7e8885247d06eb2e8840b8b145783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4mgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rnzq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.859900 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9k8zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4f75fd3-03e8-41e7-8926-f6ff21a5b681\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpdh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpdh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9k8zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.927560 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.927632 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.927657 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.927685 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:50 crc kubenswrapper[4809]: I1206 05:51:50.927702 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:50Z","lastTransitionTime":"2025-12-06T05:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.030656 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.030723 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.030744 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.030769 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.030787 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:51Z","lastTransitionTime":"2025-12-06T05:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.057263 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.058479 4809 scope.go:117] "RemoveContainer" containerID="f973788e6cfc8d6032d0a82dd6dceec13bb5ebdd481636c561d03f469b8746dd" Dec 06 05:51:51 crc kubenswrapper[4809]: E1206 05:51:51.058723 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5xvvb_openshift-ovn-kubernetes(ffb3a4f1-1e22-4220-ac80-3b2d69d2db99)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.133779 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.133837 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.133860 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.133888 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.133915 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:51Z","lastTransitionTime":"2025-12-06T05:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.186539 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:51:51 crc kubenswrapper[4809]: E1206 05:51:51.186730 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:52:07.186692022 +0000 UTC m=+52.075674994 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.186797 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.186913 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:51:51 crc kubenswrapper[4809]: E1206 05:51:51.187031 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.187079 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:51:51 crc kubenswrapper[4809]: E1206 05:51:51.187094 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 05:51:51 crc kubenswrapper[4809]: E1206 05:51:51.187181 4809 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 05:51:51 crc kubenswrapper[4809]: E1206 05:51:51.187193 4809 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:51:51 crc kubenswrapper[4809]: E1206 05:51:51.187103 4809 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 05:51:51 crc kubenswrapper[4809]: E1206 05:51:51.187260 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 05:52:07.187235175 +0000 UTC m=+52.076218157 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 05:51:51 crc kubenswrapper[4809]: E1206 05:51:51.187296 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 05:52:07.187275906 +0000 UTC m=+52.076258888 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 05:51:51 crc kubenswrapper[4809]: E1206 05:51:51.187327 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 05:52:07.187312467 +0000 UTC m=+52.076295449 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.236607 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.236911 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.237194 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.237390 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.237529 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:51Z","lastTransitionTime":"2025-12-06T05:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.288101 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:51:51 crc kubenswrapper[4809]: E1206 05:51:51.288284 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 05:51:51 crc kubenswrapper[4809]: E1206 05:51:51.288796 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 05:51:51 crc kubenswrapper[4809]: E1206 05:51:51.288969 4809 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:51:51 crc kubenswrapper[4809]: E1206 05:51:51.289165 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 05:52:07.289140337 +0000 UTC m=+52.178123309 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.340766 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.340826 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.340843 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.340865 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.340883 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:51Z","lastTransitionTime":"2025-12-06T05:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.388694 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.388714 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:51:51 crc kubenswrapper[4809]: E1206 05:51:51.389187 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:51:51 crc kubenswrapper[4809]: E1206 05:51:51.389376 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.444344 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.444384 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.444397 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.444418 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.444433 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:51Z","lastTransitionTime":"2025-12-06T05:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.546639 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.546692 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.546702 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.546717 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.546728 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:51Z","lastTransitionTime":"2025-12-06T05:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.649059 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.649098 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.649108 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.649120 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.649130 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:51Z","lastTransitionTime":"2025-12-06T05:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.751634 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.751661 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.751669 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.751683 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.751691 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:51Z","lastTransitionTime":"2025-12-06T05:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.854124 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.854159 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.854168 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.854180 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.854189 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:51Z","lastTransitionTime":"2025-12-06T05:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.956896 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.956945 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.956958 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.956973 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:51 crc kubenswrapper[4809]: I1206 05:51:51.956987 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:51Z","lastTransitionTime":"2025-12-06T05:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.059455 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.059706 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.059964 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.060157 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.060351 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:52Z","lastTransitionTime":"2025-12-06T05:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.162215 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.162489 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.162549 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.162674 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.162738 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:52Z","lastTransitionTime":"2025-12-06T05:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.265244 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.265529 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.265595 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.265696 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.265778 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:52Z","lastTransitionTime":"2025-12-06T05:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.368870 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.368917 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.368950 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.368966 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.368977 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:52Z","lastTransitionTime":"2025-12-06T05:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.388619 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:51:52 crc kubenswrapper[4809]: E1206 05:51:52.388796 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.388633 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:51:52 crc kubenswrapper[4809]: E1206 05:51:52.388922 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.472490 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.472539 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.472550 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.472568 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.472581 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:52Z","lastTransitionTime":"2025-12-06T05:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.575856 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.576044 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.576089 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.576122 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.576158 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:52Z","lastTransitionTime":"2025-12-06T05:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.698187 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.698215 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.698224 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.698236 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.698245 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:52Z","lastTransitionTime":"2025-12-06T05:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.702313 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a4f75fd3-03e8-41e7-8926-f6ff21a5b681-metrics-certs\") pod \"network-metrics-daemon-9k8zx\" (UID: \"a4f75fd3-03e8-41e7-8926-f6ff21a5b681\") " pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:51:52 crc kubenswrapper[4809]: E1206 05:51:52.702455 4809 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 05:51:52 crc kubenswrapper[4809]: E1206 05:51:52.702525 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a4f75fd3-03e8-41e7-8926-f6ff21a5b681-metrics-certs podName:a4f75fd3-03e8-41e7-8926-f6ff21a5b681 nodeName:}" failed. No retries permitted until 2025-12-06 05:51:56.702504333 +0000 UTC m=+41.591487285 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a4f75fd3-03e8-41e7-8926-f6ff21a5b681-metrics-certs") pod "network-metrics-daemon-9k8zx" (UID: "a4f75fd3-03e8-41e7-8926-f6ff21a5b681") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.800448 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.800496 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.800516 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.800538 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.800556 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:52Z","lastTransitionTime":"2025-12-06T05:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.902781 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.902811 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.902818 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.902831 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:52 crc kubenswrapper[4809]: I1206 05:51:52.902839 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:52Z","lastTransitionTime":"2025-12-06T05:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.005631 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.005703 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.005718 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.005734 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.005746 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:53Z","lastTransitionTime":"2025-12-06T05:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.108267 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.108358 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.108394 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.108424 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.108448 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:53Z","lastTransitionTime":"2025-12-06T05:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.210287 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.210329 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.210340 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.210356 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.210369 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:53Z","lastTransitionTime":"2025-12-06T05:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.313305 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.313393 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.313423 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.313453 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.313475 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:53Z","lastTransitionTime":"2025-12-06T05:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.387654 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:51:53 crc kubenswrapper[4809]: E1206 05:51:53.387819 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.387681 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:51:53 crc kubenswrapper[4809]: E1206 05:51:53.388310 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.415489 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.415528 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.415538 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.415553 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.415564 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:53Z","lastTransitionTime":"2025-12-06T05:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.517788 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.517836 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.517845 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.517857 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.517865 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:53Z","lastTransitionTime":"2025-12-06T05:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.621698 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.621759 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.621769 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.621790 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.621802 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:53Z","lastTransitionTime":"2025-12-06T05:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.725114 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.725163 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.725175 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.725194 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.725207 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:53Z","lastTransitionTime":"2025-12-06T05:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.827509 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.827563 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.827580 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.827602 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.827618 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:53Z","lastTransitionTime":"2025-12-06T05:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.929844 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.929904 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.929920 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.929972 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:53 crc kubenswrapper[4809]: I1206 05:51:53.929989 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:53Z","lastTransitionTime":"2025-12-06T05:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.032975 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.033049 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.033074 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.033102 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.033124 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:54Z","lastTransitionTime":"2025-12-06T05:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.136706 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.136766 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.136785 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.136810 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.136826 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:54Z","lastTransitionTime":"2025-12-06T05:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.239781 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.239823 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.239832 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.239846 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.239855 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:54Z","lastTransitionTime":"2025-12-06T05:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.342509 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.342599 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.342625 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.342656 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.342681 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:54Z","lastTransitionTime":"2025-12-06T05:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.388214 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.388353 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:51:54 crc kubenswrapper[4809]: E1206 05:51:54.388423 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:51:54 crc kubenswrapper[4809]: E1206 05:51:54.388755 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.444900 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.445025 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.445053 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.445090 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.445112 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:54Z","lastTransitionTime":"2025-12-06T05:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.547628 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.547666 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.547674 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.547687 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.547696 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:54Z","lastTransitionTime":"2025-12-06T05:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.650217 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.650272 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.650284 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.650301 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.650312 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:54Z","lastTransitionTime":"2025-12-06T05:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.753626 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.753693 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.753714 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.753739 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.753757 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:54Z","lastTransitionTime":"2025-12-06T05:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.856362 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.856411 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.856435 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.856452 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.856463 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:54Z","lastTransitionTime":"2025-12-06T05:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.959362 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.959436 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.959461 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.959490 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:54 crc kubenswrapper[4809]: I1206 05:51:54.959513 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:54Z","lastTransitionTime":"2025-12-06T05:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.062079 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.062142 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.062160 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.062180 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.062198 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:55Z","lastTransitionTime":"2025-12-06T05:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.164870 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.164926 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.164977 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.165002 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.165021 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:55Z","lastTransitionTime":"2025-12-06T05:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.267053 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.267395 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.267521 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.267670 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.267788 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:55Z","lastTransitionTime":"2025-12-06T05:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.370928 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.371314 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.371526 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.371755 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.371996 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:55Z","lastTransitionTime":"2025-12-06T05:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.388446 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:51:55 crc kubenswrapper[4809]: E1206 05:51:55.390001 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.388497 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:51:55 crc kubenswrapper[4809]: E1206 05:51:55.390470 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.409114 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd45866c51410e910e87d43870bf531b0d05541e9164e700150fd37987e284af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:55Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.427179 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:55Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.447421 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:55Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.467249 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:55Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.474777 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.475123 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.475331 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.475522 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.475686 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:55Z","lastTransitionTime":"2025-12-06T05:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.484381 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947d89be9d04856af8d09facb8d4583f91898bd03a36d3ce8b9b74bc3bb795e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://930788ec9b4d97fdc39353a93111b84707037ce445f1930aa7737a20a44e1d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-npms2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:55Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.501654 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5636dd78-3d61-4c6a-b8fd-6d2457e19234\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:55Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.516639 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rnzq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0075197-9392-4109-b8a7-39507b15cc17\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53a00b17138b236dc8517cbd4ee71ea3ed7e8885247d06eb2e8840b8b145783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4mgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rnzq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:55Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.531731 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9k8zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4f75fd3-03e8-41e7-8926-f6ff21a5b681\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpdh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpdh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9k8zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:55Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.544873 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nndh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a80ba5e1-90cc-44ce-be19-fdf0c007d5b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3684c4e5e839ce117a1772571ae080042e0187ea1aaca635edca07ac10d94c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2c22k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31d74665ae370eed09b8a3195bd3d122ad64c6e33bf55fc42cd40a862671b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2c22k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nndh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:55Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.565330 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f28d0a3aec8a36273fa4480c552d80a5ee23210203861811e37f27bccda1147e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40791d93876595c4093814146d269b8fe70efb64ef2888e03c25aebef0e9295e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:55Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.578206 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.578531 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.578619 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.578717 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.578796 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:55Z","lastTransitionTime":"2025-12-06T05:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.581461 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a48de155-7389-48db-8f86-d175b7db70c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://701274e49f68efbc0a939bc08771288c50662cef9c59da73170f7f114e222337\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9t7pq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:55Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.596653 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cb4vc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed42d686-41af-470e-910d-a1fabbec66b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f629b20d9f61227553520b30d0ca046750433926a8d31501e5fb77467426cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2nwpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cb4vc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:55Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.607772 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hmlmc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c7558c-89b1-432d-893c-ccb2b3aba183\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c2f8876171399b61abd26285afba0547dac3c24c957bd6d202721b04c912aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm7wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hmlmc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:55Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.624906 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://010c2681c08dc21bc3a6fb82c04ab3dc2c26dcb76ec12b143c22845cac04ed25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:55Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.648259 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c981561d51870deffeb446e8ca769b7ac0a1a16edbe4e982e0f2c84a45043db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d58d489f532b78aca59f5b92584aad1dc292e6700c112c531cbeb3209fd90c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dcd421b38e5f9ba0c542a09f0047bde6c043942a096d11575afb7423ab2552c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab9b3bd4c8a6d6af7939b06fe01b8a0a50d81bb8c669bc6a27f486953c234e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a3e4453935871bb84d68bc4a8b6c98f299b7c8094ba82291bc90e258e1fec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5be86696b7a6518c5df930accb2c6890d58cfc37104fd34a359724b9e30231b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f973788e6cfc8d6032d0a82dd6dceec13bb5ebdd481636c561d03f469b8746dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f973788e6cfc8d6032d0a82dd6dceec13bb5ebdd481636c561d03f469b8746dd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:51:47Z\\\",\\\"message\\\":\\\"ackage-server-manager-metrics for network=default are: map[]\\\\nI1206 05:51:46.633808 6223 services_controller.go:443] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.110\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1206 05:51:46.633876 6223 services_controller.go:444] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1206 05:51:46.633966 6223 services_controller.go:445] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nI1206 05:51:46.633753 6223 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-operators\\\\\\\"}\\\\nI1206 05:51:46.634043 6223 services_controller.go:360] Finished syncing service redhat-operators on namespace openshift-marketplace for network=default : 1.127407ms\\\\nF1206 05:51:46.633754 6223 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5xvvb_openshift-ovn-kubernetes(ffb3a4f1-1e22-4220-ac80-3b2d69d2db99)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dee8b965be3b2486f1f532f2bdfbbcbd84f9b3b711a11c25e86dae1284314d4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5xvvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:55Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.660379 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa4dd3f8-c119-46de-ad4e-dd29bb11a32c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa59e4ec9867cf280b308ee3adb5e303eaf94455aca588f5eb74a2b6263d57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5720f7f2977d9af6ed5324a30d08d34a53d55460951b3875b7d7d753d202f349\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dfd977b049a8ff81192d4041adb265366ab88475ee84c5750305d061d21c8db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d5801b61a87c95c9a623c3106e305920199c640efa8a32cf57698245697e54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:55Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.681279 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.681309 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.681317 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.681331 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.681341 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:55Z","lastTransitionTime":"2025-12-06T05:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.784980 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.785016 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.785027 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.785045 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.785057 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:55Z","lastTransitionTime":"2025-12-06T05:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.887897 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.887955 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.887964 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.887980 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.887989 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:55Z","lastTransitionTime":"2025-12-06T05:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.990723 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.991053 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.991065 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.991083 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:55 crc kubenswrapper[4809]: I1206 05:51:55.991096 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:55Z","lastTransitionTime":"2025-12-06T05:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.094528 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.094770 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.094780 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.094798 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.094809 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:56Z","lastTransitionTime":"2025-12-06T05:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.198396 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.198463 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.198487 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.198518 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.198541 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:56Z","lastTransitionTime":"2025-12-06T05:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.302291 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.302365 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.302383 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.302408 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.302428 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:56Z","lastTransitionTime":"2025-12-06T05:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.387732 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.387732 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:51:56 crc kubenswrapper[4809]: E1206 05:51:56.387908 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:51:56 crc kubenswrapper[4809]: E1206 05:51:56.388079 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.405518 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.405562 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.405571 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.405585 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.405597 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:56Z","lastTransitionTime":"2025-12-06T05:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.508655 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.508712 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.508730 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.508751 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.508770 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:56Z","lastTransitionTime":"2025-12-06T05:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.611524 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.611597 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.611617 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.611643 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.611662 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:56Z","lastTransitionTime":"2025-12-06T05:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.713531 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.713583 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.713599 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.713621 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.713638 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:56Z","lastTransitionTime":"2025-12-06T05:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.770234 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a4f75fd3-03e8-41e7-8926-f6ff21a5b681-metrics-certs\") pod \"network-metrics-daemon-9k8zx\" (UID: \"a4f75fd3-03e8-41e7-8926-f6ff21a5b681\") " pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:51:56 crc kubenswrapper[4809]: E1206 05:51:56.770441 4809 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 05:51:56 crc kubenswrapper[4809]: E1206 05:51:56.770529 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a4f75fd3-03e8-41e7-8926-f6ff21a5b681-metrics-certs podName:a4f75fd3-03e8-41e7-8926-f6ff21a5b681 nodeName:}" failed. No retries permitted until 2025-12-06 05:52:04.770507 +0000 UTC m=+49.659490042 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a4f75fd3-03e8-41e7-8926-f6ff21a5b681-metrics-certs") pod "network-metrics-daemon-9k8zx" (UID: "a4f75fd3-03e8-41e7-8926-f6ff21a5b681") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.816477 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.816542 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.816558 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.816582 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.816599 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:56Z","lastTransitionTime":"2025-12-06T05:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.920036 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.920102 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.920119 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.920173 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:56 crc kubenswrapper[4809]: I1206 05:51:56.920191 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:56Z","lastTransitionTime":"2025-12-06T05:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.023746 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.023832 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.023854 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.023882 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.023907 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:57Z","lastTransitionTime":"2025-12-06T05:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.127729 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.127774 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.127792 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.127813 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.127828 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:57Z","lastTransitionTime":"2025-12-06T05:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.230044 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.230089 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.230102 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.230117 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.230126 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:57Z","lastTransitionTime":"2025-12-06T05:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.333215 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.333283 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.333301 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.333323 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.333353 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:57Z","lastTransitionTime":"2025-12-06T05:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.388498 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:51:57 crc kubenswrapper[4809]: E1206 05:51:57.388620 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.388653 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:51:57 crc kubenswrapper[4809]: E1206 05:51:57.388818 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.436374 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.436425 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.436434 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.436448 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.436457 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:57Z","lastTransitionTime":"2025-12-06T05:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.539383 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.539469 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.539500 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.539530 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.539549 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:57Z","lastTransitionTime":"2025-12-06T05:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.642895 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.643591 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.643908 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.644360 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.644580 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:57Z","lastTransitionTime":"2025-12-06T05:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.746712 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.746766 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.746781 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.746805 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.746820 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:57Z","lastTransitionTime":"2025-12-06T05:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.869371 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.869626 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.869713 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.869808 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.869961 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:57Z","lastTransitionTime":"2025-12-06T05:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.973743 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.973802 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.973813 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.973851 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:57 crc kubenswrapper[4809]: I1206 05:51:57.973865 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:57Z","lastTransitionTime":"2025-12-06T05:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:58 crc kubenswrapper[4809]: I1206 05:51:58.076376 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:58 crc kubenswrapper[4809]: I1206 05:51:58.076418 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:58 crc kubenswrapper[4809]: I1206 05:51:58.076434 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:58 crc kubenswrapper[4809]: I1206 05:51:58.076497 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:58 crc kubenswrapper[4809]: I1206 05:51:58.076512 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:58Z","lastTransitionTime":"2025-12-06T05:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:58 crc kubenswrapper[4809]: I1206 05:51:58.179075 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:58 crc kubenswrapper[4809]: I1206 05:51:58.179150 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:58 crc kubenswrapper[4809]: I1206 05:51:58.179170 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:58 crc kubenswrapper[4809]: I1206 05:51:58.179194 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:58 crc kubenswrapper[4809]: I1206 05:51:58.179212 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:58Z","lastTransitionTime":"2025-12-06T05:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:58 crc kubenswrapper[4809]: I1206 05:51:58.281461 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:58 crc kubenswrapper[4809]: I1206 05:51:58.281495 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:58 crc kubenswrapper[4809]: I1206 05:51:58.281504 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:58 crc kubenswrapper[4809]: I1206 05:51:58.281519 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:58 crc kubenswrapper[4809]: I1206 05:51:58.281529 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:58Z","lastTransitionTime":"2025-12-06T05:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:58 crc kubenswrapper[4809]: I1206 05:51:58.384626 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:58 crc kubenswrapper[4809]: I1206 05:51:58.384872 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:58 crc kubenswrapper[4809]: I1206 05:51:58.384967 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:58 crc kubenswrapper[4809]: I1206 05:51:58.385077 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:58 crc kubenswrapper[4809]: I1206 05:51:58.385142 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:58Z","lastTransitionTime":"2025-12-06T05:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:58 crc kubenswrapper[4809]: I1206 05:51:58.387926 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:51:58 crc kubenswrapper[4809]: I1206 05:51:58.387971 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:51:58 crc kubenswrapper[4809]: E1206 05:51:58.388128 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:51:58 crc kubenswrapper[4809]: E1206 05:51:58.388267 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:51:58 crc kubenswrapper[4809]: I1206 05:51:58.487706 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:58 crc kubenswrapper[4809]: I1206 05:51:58.487739 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:58 crc kubenswrapper[4809]: I1206 05:51:58.487749 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:58 crc kubenswrapper[4809]: I1206 05:51:58.487764 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:58 crc kubenswrapper[4809]: I1206 05:51:58.487775 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:58Z","lastTransitionTime":"2025-12-06T05:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:58 crc kubenswrapper[4809]: I1206 05:51:58.590517 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:58 crc kubenswrapper[4809]: I1206 05:51:58.590626 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:58 crc kubenswrapper[4809]: I1206 05:51:58.590655 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:58 crc kubenswrapper[4809]: I1206 05:51:58.590685 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:58 crc kubenswrapper[4809]: I1206 05:51:58.590704 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:58Z","lastTransitionTime":"2025-12-06T05:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:58 crc kubenswrapper[4809]: I1206 05:51:58.693250 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:58 crc kubenswrapper[4809]: I1206 05:51:58.693288 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:58 crc kubenswrapper[4809]: I1206 05:51:58.693296 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:58 crc kubenswrapper[4809]: I1206 05:51:58.693309 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:58 crc kubenswrapper[4809]: I1206 05:51:58.693317 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:58Z","lastTransitionTime":"2025-12-06T05:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:58 crc kubenswrapper[4809]: I1206 05:51:58.796124 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:58 crc kubenswrapper[4809]: I1206 05:51:58.796407 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:58 crc kubenswrapper[4809]: I1206 05:51:58.796473 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:58 crc kubenswrapper[4809]: I1206 05:51:58.796548 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:58 crc kubenswrapper[4809]: I1206 05:51:58.796611 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:58Z","lastTransitionTime":"2025-12-06T05:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:58 crc kubenswrapper[4809]: I1206 05:51:58.899231 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:58 crc kubenswrapper[4809]: I1206 05:51:58.899275 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:58 crc kubenswrapper[4809]: I1206 05:51:58.899288 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:58 crc kubenswrapper[4809]: I1206 05:51:58.899304 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:58 crc kubenswrapper[4809]: I1206 05:51:58.899314 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:58Z","lastTransitionTime":"2025-12-06T05:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.000865 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.000907 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.000917 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.000952 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.000969 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:59Z","lastTransitionTime":"2025-12-06T05:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.103109 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.103154 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.103163 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.103177 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.103186 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:59Z","lastTransitionTime":"2025-12-06T05:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.206152 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.206192 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.206203 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.206219 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.206233 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:59Z","lastTransitionTime":"2025-12-06T05:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.272598 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.272641 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.272652 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.272667 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.272681 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:59Z","lastTransitionTime":"2025-12-06T05:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:59 crc kubenswrapper[4809]: E1206 05:51:59.286292 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"88a28b6b-9c85-4932-b5bc-36346303aee5\\\",\\\"systemUUID\\\":\\\"5931df91-c1df-4f78-9e5f-391a01a76127\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:59Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.289635 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.289690 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.289709 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.289735 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.289752 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:59Z","lastTransitionTime":"2025-12-06T05:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:59 crc kubenswrapper[4809]: E1206 05:51:59.303529 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"88a28b6b-9c85-4932-b5bc-36346303aee5\\\",\\\"systemUUID\\\":\\\"5931df91-c1df-4f78-9e5f-391a01a76127\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:59Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.307823 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.307874 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.307885 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.307898 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.307910 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:59Z","lastTransitionTime":"2025-12-06T05:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:59 crc kubenswrapper[4809]: E1206 05:51:59.321144 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"88a28b6b-9c85-4932-b5bc-36346303aee5\\\",\\\"systemUUID\\\":\\\"5931df91-c1df-4f78-9e5f-391a01a76127\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:59Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.328341 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.328391 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.328410 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.328427 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.328439 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:59Z","lastTransitionTime":"2025-12-06T05:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:59 crc kubenswrapper[4809]: E1206 05:51:59.341407 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"88a28b6b-9c85-4932-b5bc-36346303aee5\\\",\\\"systemUUID\\\":\\\"5931df91-c1df-4f78-9e5f-391a01a76127\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:59Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.346619 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.346658 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.346675 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.346693 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.346708 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:59Z","lastTransitionTime":"2025-12-06T05:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:59 crc kubenswrapper[4809]: E1206 05:51:59.364325 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:51:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"88a28b6b-9c85-4932-b5bc-36346303aee5\\\",\\\"systemUUID\\\":\\\"5931df91-c1df-4f78-9e5f-391a01a76127\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:51:59Z is after 2025-08-24T17:21:41Z" Dec 06 05:51:59 crc kubenswrapper[4809]: E1206 05:51:59.364480 4809 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.366160 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.366249 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.366266 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.366286 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.366300 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:59Z","lastTransitionTime":"2025-12-06T05:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.388535 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.388535 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:51:59 crc kubenswrapper[4809]: E1206 05:51:59.388739 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:51:59 crc kubenswrapper[4809]: E1206 05:51:59.388859 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.468340 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.468369 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.468412 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.468431 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.468442 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:59Z","lastTransitionTime":"2025-12-06T05:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.571285 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.571335 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.571347 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.571366 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.571377 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:59Z","lastTransitionTime":"2025-12-06T05:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.673950 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.673987 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.673999 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.674015 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.674026 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:59Z","lastTransitionTime":"2025-12-06T05:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.776624 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.776685 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.776706 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.776732 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.776749 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:59Z","lastTransitionTime":"2025-12-06T05:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.879337 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.879380 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.879393 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.879409 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.879422 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:59Z","lastTransitionTime":"2025-12-06T05:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.982294 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.982339 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.982350 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.982367 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:51:59 crc kubenswrapper[4809]: I1206 05:51:59.982377 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:51:59Z","lastTransitionTime":"2025-12-06T05:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:00 crc kubenswrapper[4809]: I1206 05:52:00.085382 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:00 crc kubenswrapper[4809]: I1206 05:52:00.085448 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:00 crc kubenswrapper[4809]: I1206 05:52:00.085465 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:00 crc kubenswrapper[4809]: I1206 05:52:00.085488 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:00 crc kubenswrapper[4809]: I1206 05:52:00.085505 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:00Z","lastTransitionTime":"2025-12-06T05:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:00 crc kubenswrapper[4809]: I1206 05:52:00.188116 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:00 crc kubenswrapper[4809]: I1206 05:52:00.188155 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:00 crc kubenswrapper[4809]: I1206 05:52:00.188163 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:00 crc kubenswrapper[4809]: I1206 05:52:00.188178 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:00 crc kubenswrapper[4809]: I1206 05:52:00.188187 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:00Z","lastTransitionTime":"2025-12-06T05:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:00 crc kubenswrapper[4809]: I1206 05:52:00.290705 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:00 crc kubenswrapper[4809]: I1206 05:52:00.290744 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:00 crc kubenswrapper[4809]: I1206 05:52:00.290756 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:00 crc kubenswrapper[4809]: I1206 05:52:00.290772 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:00 crc kubenswrapper[4809]: I1206 05:52:00.290784 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:00Z","lastTransitionTime":"2025-12-06T05:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:00 crc kubenswrapper[4809]: I1206 05:52:00.388575 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:52:00 crc kubenswrapper[4809]: I1206 05:52:00.388734 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:52:00 crc kubenswrapper[4809]: E1206 05:52:00.389006 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:52:00 crc kubenswrapper[4809]: E1206 05:52:00.389266 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:52:00 crc kubenswrapper[4809]: I1206 05:52:00.393959 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:00 crc kubenswrapper[4809]: I1206 05:52:00.394014 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:00 crc kubenswrapper[4809]: I1206 05:52:00.394031 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:00 crc kubenswrapper[4809]: I1206 05:52:00.394057 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:00 crc kubenswrapper[4809]: I1206 05:52:00.394074 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:00Z","lastTransitionTime":"2025-12-06T05:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:00 crc kubenswrapper[4809]: I1206 05:52:00.496715 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:00 crc kubenswrapper[4809]: I1206 05:52:00.496772 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:00 crc kubenswrapper[4809]: I1206 05:52:00.496790 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:00 crc kubenswrapper[4809]: I1206 05:52:00.496806 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:00 crc kubenswrapper[4809]: I1206 05:52:00.496817 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:00Z","lastTransitionTime":"2025-12-06T05:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:00 crc kubenswrapper[4809]: I1206 05:52:00.599896 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:00 crc kubenswrapper[4809]: I1206 05:52:00.599984 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:00 crc kubenswrapper[4809]: I1206 05:52:00.599998 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:00 crc kubenswrapper[4809]: I1206 05:52:00.600016 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:00 crc kubenswrapper[4809]: I1206 05:52:00.600028 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:00Z","lastTransitionTime":"2025-12-06T05:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:00 crc kubenswrapper[4809]: I1206 05:52:00.702811 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:00 crc kubenswrapper[4809]: I1206 05:52:00.702882 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:00 crc kubenswrapper[4809]: I1206 05:52:00.702892 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:00 crc kubenswrapper[4809]: I1206 05:52:00.702914 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:00 crc kubenswrapper[4809]: I1206 05:52:00.702929 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:00Z","lastTransitionTime":"2025-12-06T05:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:00 crc kubenswrapper[4809]: I1206 05:52:00.805907 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:00 crc kubenswrapper[4809]: I1206 05:52:00.805994 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:00 crc kubenswrapper[4809]: I1206 05:52:00.806012 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:00 crc kubenswrapper[4809]: I1206 05:52:00.806041 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:00 crc kubenswrapper[4809]: I1206 05:52:00.806058 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:00Z","lastTransitionTime":"2025-12-06T05:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:00 crc kubenswrapper[4809]: I1206 05:52:00.908291 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:00 crc kubenswrapper[4809]: I1206 05:52:00.908413 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:00 crc kubenswrapper[4809]: I1206 05:52:00.908440 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:00 crc kubenswrapper[4809]: I1206 05:52:00.908471 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:00 crc kubenswrapper[4809]: I1206 05:52:00.908494 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:00Z","lastTransitionTime":"2025-12-06T05:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.010637 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.010689 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.010702 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.010719 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.010735 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:01Z","lastTransitionTime":"2025-12-06T05:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.113036 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.113121 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.113138 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.113153 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.113164 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:01Z","lastTransitionTime":"2025-12-06T05:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.216152 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.216206 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.216222 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.216243 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.216260 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:01Z","lastTransitionTime":"2025-12-06T05:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.320392 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.320484 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.320508 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.320541 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.321208 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:01Z","lastTransitionTime":"2025-12-06T05:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.388231 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.388247 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:52:01 crc kubenswrapper[4809]: E1206 05:52:01.388496 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:52:01 crc kubenswrapper[4809]: E1206 05:52:01.388652 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.423627 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.423691 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.423717 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.423741 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.423762 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:01Z","lastTransitionTime":"2025-12-06T05:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.526499 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.526748 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.526810 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.526914 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.527003 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:01Z","lastTransitionTime":"2025-12-06T05:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.630062 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.630405 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.630486 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.630588 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.630667 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:01Z","lastTransitionTime":"2025-12-06T05:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.733088 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.733140 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.733153 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.733174 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.733192 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:01Z","lastTransitionTime":"2025-12-06T05:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.835396 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.835706 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.835814 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.835917 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.836067 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:01Z","lastTransitionTime":"2025-12-06T05:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.939252 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.939295 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.939312 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.939334 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:01 crc kubenswrapper[4809]: I1206 05:52:01.939351 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:01Z","lastTransitionTime":"2025-12-06T05:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.042371 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.042433 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.042443 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.042459 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.042470 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:02Z","lastTransitionTime":"2025-12-06T05:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.145244 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.146042 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.146261 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.146436 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.146579 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:02Z","lastTransitionTime":"2025-12-06T05:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.249415 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.249490 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.249514 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.249543 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.249567 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:02Z","lastTransitionTime":"2025-12-06T05:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.352602 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.352651 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.352664 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.352682 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.352694 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:02Z","lastTransitionTime":"2025-12-06T05:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.387957 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.388064 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:52:02 crc kubenswrapper[4809]: E1206 05:52:02.388200 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:52:02 crc kubenswrapper[4809]: E1206 05:52:02.388274 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.455547 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.455600 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.455611 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.455628 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.455640 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:02Z","lastTransitionTime":"2025-12-06T05:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.557877 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.557912 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.557920 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.557949 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.557958 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:02Z","lastTransitionTime":"2025-12-06T05:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.660990 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.661049 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.661071 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.661100 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.661123 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:02Z","lastTransitionTime":"2025-12-06T05:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.764379 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.764461 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.764473 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.764490 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.764501 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:02Z","lastTransitionTime":"2025-12-06T05:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.866515 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.866593 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.866607 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.866623 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.866634 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:02Z","lastTransitionTime":"2025-12-06T05:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.969235 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.969336 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.969361 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.969392 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:02 crc kubenswrapper[4809]: I1206 05:52:02.969415 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:02Z","lastTransitionTime":"2025-12-06T05:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:03 crc kubenswrapper[4809]: I1206 05:52:03.071442 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:03 crc kubenswrapper[4809]: I1206 05:52:03.071495 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:03 crc kubenswrapper[4809]: I1206 05:52:03.071504 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:03 crc kubenswrapper[4809]: I1206 05:52:03.071516 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:03 crc kubenswrapper[4809]: I1206 05:52:03.071524 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:03Z","lastTransitionTime":"2025-12-06T05:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:03 crc kubenswrapper[4809]: I1206 05:52:03.174801 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:03 crc kubenswrapper[4809]: I1206 05:52:03.175924 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:03 crc kubenswrapper[4809]: I1206 05:52:03.176154 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:03 crc kubenswrapper[4809]: I1206 05:52:03.176323 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:03 crc kubenswrapper[4809]: I1206 05:52:03.176497 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:03Z","lastTransitionTime":"2025-12-06T05:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:03 crc kubenswrapper[4809]: I1206 05:52:03.279542 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:03 crc kubenswrapper[4809]: I1206 05:52:03.279614 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:03 crc kubenswrapper[4809]: I1206 05:52:03.279638 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:03 crc kubenswrapper[4809]: I1206 05:52:03.279669 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:03 crc kubenswrapper[4809]: I1206 05:52:03.279694 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:03Z","lastTransitionTime":"2025-12-06T05:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:03 crc kubenswrapper[4809]: I1206 05:52:03.382555 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:03 crc kubenswrapper[4809]: I1206 05:52:03.382863 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:03 crc kubenswrapper[4809]: I1206 05:52:03.383126 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:03 crc kubenswrapper[4809]: I1206 05:52:03.383377 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:03 crc kubenswrapper[4809]: I1206 05:52:03.383561 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:03Z","lastTransitionTime":"2025-12-06T05:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:03 crc kubenswrapper[4809]: I1206 05:52:03.387897 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:52:03 crc kubenswrapper[4809]: I1206 05:52:03.387978 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:52:03 crc kubenswrapper[4809]: E1206 05:52:03.388698 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:52:03 crc kubenswrapper[4809]: E1206 05:52:03.388300 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:52:03 crc kubenswrapper[4809]: I1206 05:52:03.486409 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:03 crc kubenswrapper[4809]: I1206 05:52:03.486478 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:03 crc kubenswrapper[4809]: I1206 05:52:03.486490 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:03 crc kubenswrapper[4809]: I1206 05:52:03.486527 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:03 crc kubenswrapper[4809]: I1206 05:52:03.486541 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:03Z","lastTransitionTime":"2025-12-06T05:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:03 crc kubenswrapper[4809]: I1206 05:52:03.590191 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:03 crc kubenswrapper[4809]: I1206 05:52:03.590238 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:03 crc kubenswrapper[4809]: I1206 05:52:03.590250 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:03 crc kubenswrapper[4809]: I1206 05:52:03.590270 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:03 crc kubenswrapper[4809]: I1206 05:52:03.590282 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:03Z","lastTransitionTime":"2025-12-06T05:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:03 crc kubenswrapper[4809]: I1206 05:52:03.692322 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:03 crc kubenswrapper[4809]: I1206 05:52:03.692382 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:03 crc kubenswrapper[4809]: I1206 05:52:03.692409 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:03 crc kubenswrapper[4809]: I1206 05:52:03.692466 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:03 crc kubenswrapper[4809]: I1206 05:52:03.692485 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:03Z","lastTransitionTime":"2025-12-06T05:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:03 crc kubenswrapper[4809]: I1206 05:52:03.795088 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:03 crc kubenswrapper[4809]: I1206 05:52:03.795178 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:03 crc kubenswrapper[4809]: I1206 05:52:03.795196 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:03 crc kubenswrapper[4809]: I1206 05:52:03.795217 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:03 crc kubenswrapper[4809]: I1206 05:52:03.795230 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:03Z","lastTransitionTime":"2025-12-06T05:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:03 crc kubenswrapper[4809]: I1206 05:52:03.898455 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:03 crc kubenswrapper[4809]: I1206 05:52:03.898518 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:03 crc kubenswrapper[4809]: I1206 05:52:03.898539 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:03 crc kubenswrapper[4809]: I1206 05:52:03.898561 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:03 crc kubenswrapper[4809]: I1206 05:52:03.898574 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:03Z","lastTransitionTime":"2025-12-06T05:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.001147 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.001182 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.001192 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.001206 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.001217 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:04Z","lastTransitionTime":"2025-12-06T05:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.103311 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.103346 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.103357 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.103371 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.103381 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:04Z","lastTransitionTime":"2025-12-06T05:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.205747 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.205791 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.205803 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.205818 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.205829 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:04Z","lastTransitionTime":"2025-12-06T05:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.307957 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.307999 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.308009 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.308024 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.308034 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:04Z","lastTransitionTime":"2025-12-06T05:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.388339 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.388380 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:52:04 crc kubenswrapper[4809]: E1206 05:52:04.388582 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:52:04 crc kubenswrapper[4809]: E1206 05:52:04.388766 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.410886 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.410971 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.410989 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.411012 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.411028 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:04Z","lastTransitionTime":"2025-12-06T05:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.422343 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.434840 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.444514 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa4dd3f8-c119-46de-ad4e-dd29bb11a32c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa59e4ec9867cf280b308ee3adb5e303eaf94455aca588f5eb74a2b6263d57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5720f7f2977d9af6ed5324a30d08d34a53d55460951b3875b7d7d753d202f349\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dfd977b049a8ff81192d4041adb265366ab88475ee84c5750305d061d21c8db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d5801b61a87c95c9a623c3106e305920199c640efa8a32cf57698245697e54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:04Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.461259 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://010c2681c08dc21bc3a6fb82c04ab3dc2c26dcb76ec12b143c22845cac04ed25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:04Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.484656 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c981561d51870deffeb446e8ca769b7ac0a1a16edbe4e982e0f2c84a45043db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d58d489f532b78aca59f5b92584aad1dc292e6700c112c531cbeb3209fd90c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dcd421b38e5f9ba0c542a09f0047bde6c043942a096d11575afb7423ab2552c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab9b3bd4c8a6d6af7939b06fe01b8a0a50d81bb8c669bc6a27f486953c234e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a3e4453935871bb84d68bc4a8b6c98f299b7c8094ba82291bc90e258e1fec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5be86696b7a6518c5df930accb2c6890d58cfc37104fd34a359724b9e30231b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f973788e6cfc8d6032d0a82dd6dceec13bb5ebdd481636c561d03f469b8746dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f973788e6cfc8d6032d0a82dd6dceec13bb5ebdd481636c561d03f469b8746dd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:51:47Z\\\",\\\"message\\\":\\\"ackage-server-manager-metrics for network=default are: map[]\\\\nI1206 05:51:46.633808 6223 services_controller.go:443] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.110\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1206 05:51:46.633876 6223 services_controller.go:444] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1206 05:51:46.633966 6223 services_controller.go:445] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nI1206 05:51:46.633753 6223 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-operators\\\\\\\"}\\\\nI1206 05:51:46.634043 6223 services_controller.go:360] Finished syncing service redhat-operators on namespace openshift-marketplace for network=default : 1.127407ms\\\\nF1206 05:51:46.633754 6223 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5xvvb_openshift-ovn-kubernetes(ffb3a4f1-1e22-4220-ac80-3b2d69d2db99)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dee8b965be3b2486f1f532f2bdfbbcbd84f9b3b711a11c25e86dae1284314d4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5xvvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:04Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.499182 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5636dd78-3d61-4c6a-b8fd-6d2457e19234\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:04Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.511853 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd45866c51410e910e87d43870bf531b0d05541e9164e700150fd37987e284af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:04Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.514052 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.514096 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.514106 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.514123 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.514133 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:04Z","lastTransitionTime":"2025-12-06T05:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.529914 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:04Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.543237 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:04Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.558922 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:04Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.573410 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947d89be9d04856af8d09facb8d4583f91898bd03a36d3ce8b9b74bc3bb795e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://930788ec9b4d97fdc39353a93111b84707037ce445f1930aa7737a20a44e1d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-npms2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:04Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.586985 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rnzq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0075197-9392-4109-b8a7-39507b15cc17\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53a00b17138b236dc8517cbd4ee71ea3ed7e8885247d06eb2e8840b8b145783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4mgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rnzq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:04Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.599292 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9k8zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4f75fd3-03e8-41e7-8926-f6ff21a5b681\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpdh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpdh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9k8zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:04Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.613038 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f28d0a3aec8a36273fa4480c552d80a5ee23210203861811e37f27bccda1147e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40791d93876595c4093814146d269b8fe70efb64ef2888e03c25aebef0e9295e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:04Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.617026 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.617064 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.617076 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.617090 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.617101 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:04Z","lastTransitionTime":"2025-12-06T05:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.631351 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a48de155-7389-48db-8f86-d175b7db70c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://701274e49f68efbc0a939bc08771288c50662cef9c59da73170f7f114e222337\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9t7pq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:04Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.649253 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nndh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a80ba5e1-90cc-44ce-be19-fdf0c007d5b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3684c4e5e839ce117a1772571ae080042e0187ea1aaca635edca07ac10d94c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2c22k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31d74665ae370eed09b8a3195bd3d122ad64c6e33bf55fc42cd40a862671b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2c22k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nndh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:04Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.670072 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cb4vc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed42d686-41af-470e-910d-a1fabbec66b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f629b20d9f61227553520b30d0ca046750433926a8d31501e5fb77467426cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2nwpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cb4vc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:04Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.682407 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hmlmc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c7558c-89b1-432d-893c-ccb2b3aba183\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c2f8876171399b61abd26285afba0547dac3c24c957bd6d202721b04c912aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm7wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hmlmc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:04Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.719549 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.719591 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.719608 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.719630 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.719647 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:04Z","lastTransitionTime":"2025-12-06T05:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.822804 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.822852 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.822868 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.822890 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.822905 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:04Z","lastTransitionTime":"2025-12-06T05:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.857661 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a4f75fd3-03e8-41e7-8926-f6ff21a5b681-metrics-certs\") pod \"network-metrics-daemon-9k8zx\" (UID: \"a4f75fd3-03e8-41e7-8926-f6ff21a5b681\") " pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:52:04 crc kubenswrapper[4809]: E1206 05:52:04.858039 4809 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 05:52:04 crc kubenswrapper[4809]: E1206 05:52:04.858143 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a4f75fd3-03e8-41e7-8926-f6ff21a5b681-metrics-certs podName:a4f75fd3-03e8-41e7-8926-f6ff21a5b681 nodeName:}" failed. No retries permitted until 2025-12-06 05:52:20.858121503 +0000 UTC m=+65.747104455 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a4f75fd3-03e8-41e7-8926-f6ff21a5b681-metrics-certs") pod "network-metrics-daemon-9k8zx" (UID: "a4f75fd3-03e8-41e7-8926-f6ff21a5b681") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.926457 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.926519 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.926533 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.926564 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:04 crc kubenswrapper[4809]: I1206 05:52:04.926576 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:04Z","lastTransitionTime":"2025-12-06T05:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.029517 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.029559 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.029571 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.029587 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.029600 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:05Z","lastTransitionTime":"2025-12-06T05:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.131495 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.131534 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.131543 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.131556 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.131564 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:05Z","lastTransitionTime":"2025-12-06T05:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.234904 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.235000 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.235019 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.235044 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.235061 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:05Z","lastTransitionTime":"2025-12-06T05:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.338102 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.338141 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.338153 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.338170 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.338182 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:05Z","lastTransitionTime":"2025-12-06T05:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.389336 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.389420 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:52:05 crc kubenswrapper[4809]: E1206 05:52:05.389584 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:52:05 crc kubenswrapper[4809]: E1206 05:52:05.389440 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.390294 4809 scope.go:117] "RemoveContainer" containerID="f973788e6cfc8d6032d0a82dd6dceec13bb5ebdd481636c561d03f469b8746dd" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.409334 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a48de155-7389-48db-8f86-d175b7db70c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://701274e49f68efbc0a939bc08771288c50662cef9c59da73170f7f114e222337\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9t7pq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:05Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.425851 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nndh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a80ba5e1-90cc-44ce-be19-fdf0c007d5b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3684c4e5e839ce117a1772571ae080042e0187ea1aaca635edca07ac10d94c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2c22k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31d74665ae370eed09b8a3195bd3d122ad64c6e33bf55fc42cd40a862671b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2c22k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nndh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:05Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.440399 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.440430 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.440439 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.440451 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.440461 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:05Z","lastTransitionTime":"2025-12-06T05:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.442562 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f28d0a3aec8a36273fa4480c552d80a5ee23210203861811e37f27bccda1147e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40791d93876595c4093814146d269b8fe70efb64ef2888e03c25aebef0e9295e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:05Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.456308 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hmlmc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c7558c-89b1-432d-893c-ccb2b3aba183\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c2f8876171399b61abd26285afba0547dac3c24c957bd6d202721b04c912aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm7wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hmlmc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:05Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.470301 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cb4vc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed42d686-41af-470e-910d-a1fabbec66b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f629b20d9f61227553520b30d0ca046750433926a8d31501e5fb77467426cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2nwpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cb4vc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:05Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.483963 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa4dd3f8-c119-46de-ad4e-dd29bb11a32c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa59e4ec9867cf280b308ee3adb5e303eaf94455aca588f5eb74a2b6263d57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5720f7f2977d9af6ed5324a30d08d34a53d55460951b3875b7d7d753d202f349\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dfd977b049a8ff81192d4041adb265366ab88475ee84c5750305d061d21c8db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d5801b61a87c95c9a623c3106e305920199c640efa8a32cf57698245697e54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:05Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.496377 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://010c2681c08dc21bc3a6fb82c04ab3dc2c26dcb76ec12b143c22845cac04ed25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:05Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.513435 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c981561d51870deffeb446e8ca769b7ac0a1a16edbe4e982e0f2c84a45043db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d58d489f532b78aca59f5b92584aad1dc292e6700c112c531cbeb3209fd90c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dcd421b38e5f9ba0c542a09f0047bde6c043942a096d11575afb7423ab2552c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab9b3bd4c8a6d6af7939b06fe01b8a0a50d81bb8c669bc6a27f486953c234e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a3e4453935871bb84d68bc4a8b6c98f299b7c8094ba82291bc90e258e1fec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5be86696b7a6518c5df930accb2c6890d58cfc37104fd34a359724b9e30231b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f973788e6cfc8d6032d0a82dd6dceec13bb5ebdd481636c561d03f469b8746dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f973788e6cfc8d6032d0a82dd6dceec13bb5ebdd481636c561d03f469b8746dd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:51:47Z\\\",\\\"message\\\":\\\"ackage-server-manager-metrics for network=default are: map[]\\\\nI1206 05:51:46.633808 6223 services_controller.go:443] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.110\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1206 05:51:46.633876 6223 services_controller.go:444] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1206 05:51:46.633966 6223 services_controller.go:445] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nI1206 05:51:46.633753 6223 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-operators\\\\\\\"}\\\\nI1206 05:51:46.634043 6223 services_controller.go:360] Finished syncing service redhat-operators on namespace openshift-marketplace for network=default : 1.127407ms\\\\nF1206 05:51:46.633754 6223 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5xvvb_openshift-ovn-kubernetes(ffb3a4f1-1e22-4220-ac80-3b2d69d2db99)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dee8b965be3b2486f1f532f2bdfbbcbd84f9b3b711a11c25e86dae1284314d4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5xvvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:05Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.526633 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"353c3bc7-39c3-4f30-ab37-9883caae33f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d4b8272a98d4acd09f165939684c0990b9e56b1269f85c20d2bd41c0f75f67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ddfff8cca35bda4796fe58ace60506f5c87e4398193d6711e81aae0d4b7d729\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae336ca136d0a51627e63b1ec86bd4f916a5efaf804e5b91014188331c19802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6965798de9aaf2b8066b7bd8abc23e12545f026bf1f0930dc0085b992553f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc6965798de9aaf2b8066b7bd8abc23e12545f026bf1f0930dc0085b992553f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:05Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.536831 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd45866c51410e910e87d43870bf531b0d05541e9164e700150fd37987e284af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:05Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.542876 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.542905 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.542913 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.542926 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.542952 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:05Z","lastTransitionTime":"2025-12-06T05:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.547671 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:05Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.558194 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:05Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.568289 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:05Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.577727 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947d89be9d04856af8d09facb8d4583f91898bd03a36d3ce8b9b74bc3bb795e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://930788ec9b4d97fdc39353a93111b84707037ce445f1930aa7737a20a44e1d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-npms2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:05Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.598354 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5636dd78-3d61-4c6a-b8fd-6d2457e19234\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:05Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.609045 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9k8zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4f75fd3-03e8-41e7-8926-f6ff21a5b681\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpdh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpdh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9k8zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:05Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.620047 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rnzq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0075197-9392-4109-b8a7-39507b15cc17\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53a00b17138b236dc8517cbd4ee71ea3ed7e8885247d06eb2e8840b8b145783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4mgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rnzq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:05Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.644840 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.644868 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.644876 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.644889 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.644897 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:05Z","lastTransitionTime":"2025-12-06T05:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.746998 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.747035 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.747044 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.747057 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.747067 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:05Z","lastTransitionTime":"2025-12-06T05:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.848593 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.848649 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.848707 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.848729 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.848740 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:05Z","lastTransitionTime":"2025-12-06T05:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.952665 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.952759 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.952780 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.952812 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:05 crc kubenswrapper[4809]: I1206 05:52:05.952833 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:05Z","lastTransitionTime":"2025-12-06T05:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.055074 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.055120 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.055131 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.055149 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.055160 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:06Z","lastTransitionTime":"2025-12-06T05:52:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.156730 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.156757 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.156765 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.156777 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.156785 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:06Z","lastTransitionTime":"2025-12-06T05:52:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.259787 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.259826 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.259837 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.259856 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.259869 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:06Z","lastTransitionTime":"2025-12-06T05:52:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.393250 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.393334 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:52:06 crc kubenswrapper[4809]: E1206 05:52:06.393368 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:52:06 crc kubenswrapper[4809]: E1206 05:52:06.393461 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.394834 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.394857 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.394865 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.394875 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.394883 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:06Z","lastTransitionTime":"2025-12-06T05:52:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.497025 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.497087 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.497103 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.497131 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.497149 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:06Z","lastTransitionTime":"2025-12-06T05:52:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.600062 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.600105 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.600116 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.600130 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.600144 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:06Z","lastTransitionTime":"2025-12-06T05:52:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.703005 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.703070 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.703090 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.703116 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.703134 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:06Z","lastTransitionTime":"2025-12-06T05:52:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.748882 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5xvvb_ffb3a4f1-1e22-4220-ac80-3b2d69d2db99/ovnkube-controller/1.log" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.753173 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" event={"ID":"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99","Type":"ContainerStarted","Data":"80ffc0fe836258c86321ba5a7f8e1c7e289c0f85e0b48bc138f69cf96cf225ab"} Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.753790 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.772028 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"353c3bc7-39c3-4f30-ab37-9883caae33f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d4b8272a98d4acd09f165939684c0990b9e56b1269f85c20d2bd41c0f75f67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ddfff8cca35bda4796fe58ace60506f5c87e4398193d6711e81aae0d4b7d729\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae336ca136d0a51627e63b1ec86bd4f916a5efaf804e5b91014188331c19802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6965798de9aaf2b8066b7bd8abc23e12545f026bf1f0930dc0085b992553f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc6965798de9aaf2b8066b7bd8abc23e12545f026bf1f0930dc0085b992553f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:06Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.792410 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd45866c51410e910e87d43870bf531b0d05541e9164e700150fd37987e284af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:06Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.807796 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.807866 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.807886 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.807913 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.807960 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:06Z","lastTransitionTime":"2025-12-06T05:52:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.809558 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:06Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.826419 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:06Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.842057 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:06Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.854824 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947d89be9d04856af8d09facb8d4583f91898bd03a36d3ce8b9b74bc3bb795e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://930788ec9b4d97fdc39353a93111b84707037ce445f1930aa7737a20a44e1d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-npms2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:06Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.870943 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5636dd78-3d61-4c6a-b8fd-6d2457e19234\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:06Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.905206 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9k8zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4f75fd3-03e8-41e7-8926-f6ff21a5b681\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpdh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpdh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9k8zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:06Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.911255 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.911322 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.911337 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.911354 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.911365 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:06Z","lastTransitionTime":"2025-12-06T05:52:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.929769 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rnzq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0075197-9392-4109-b8a7-39507b15cc17\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53a00b17138b236dc8517cbd4ee71ea3ed7e8885247d06eb2e8840b8b145783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4mgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rnzq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:06Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.960415 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a48de155-7389-48db-8f86-d175b7db70c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://701274e49f68efbc0a939bc08771288c50662cef9c59da73170f7f114e222337\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9t7pq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:06Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.974692 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nndh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a80ba5e1-90cc-44ce-be19-fdf0c007d5b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3684c4e5e839ce117a1772571ae080042e0187ea1aaca635edca07ac10d94c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2c22k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31d74665ae370eed09b8a3195bd3d122ad64c6e33bf55fc42cd40a862671b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2c22k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nndh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:06Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.988371 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f28d0a3aec8a36273fa4480c552d80a5ee23210203861811e37f27bccda1147e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40791d93876595c4093814146d269b8fe70efb64ef2888e03c25aebef0e9295e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:06Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:06 crc kubenswrapper[4809]: I1206 05:52:06.999808 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hmlmc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c7558c-89b1-432d-893c-ccb2b3aba183\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c2f8876171399b61abd26285afba0547dac3c24c957bd6d202721b04c912aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm7wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hmlmc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:06Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.014313 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cb4vc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed42d686-41af-470e-910d-a1fabbec66b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f629b20d9f61227553520b30d0ca046750433926a8d31501e5fb77467426cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2nwpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cb4vc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:07Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.014830 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.014884 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.014900 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.014920 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.014950 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:07Z","lastTransitionTime":"2025-12-06T05:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.034744 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa4dd3f8-c119-46de-ad4e-dd29bb11a32c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa59e4ec9867cf280b308ee3adb5e303eaf94455aca588f5eb74a2b6263d57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5720f7f2977d9af6ed5324a30d08d34a53d55460951b3875b7d7d753d202f349\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dfd977b049a8ff81192d4041adb265366ab88475ee84c5750305d061d21c8db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d5801b61a87c95c9a623c3106e305920199c640efa8a32cf57698245697e54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:07Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.050919 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://010c2681c08dc21bc3a6fb82c04ab3dc2c26dcb76ec12b143c22845cac04ed25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:07Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.078114 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c981561d51870deffeb446e8ca769b7ac0a1a16edbe4e982e0f2c84a45043db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d58d489f532b78aca59f5b92584aad1dc292e6700c112c531cbeb3209fd90c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dcd421b38e5f9ba0c542a09f0047bde6c043942a096d11575afb7423ab2552c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab9b3bd4c8a6d6af7939b06fe01b8a0a50d81bb8c669bc6a27f486953c234e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a3e4453935871bb84d68bc4a8b6c98f299b7c8094ba82291bc90e258e1fec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5be86696b7a6518c5df930accb2c6890d58cfc37104fd34a359724b9e30231b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ffc0fe836258c86321ba5a7f8e1c7e289c0f85e0b48bc138f69cf96cf225ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f973788e6cfc8d6032d0a82dd6dceec13bb5ebdd481636c561d03f469b8746dd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:51:47Z\\\",\\\"message\\\":\\\"ackage-server-manager-metrics for network=default are: map[]\\\\nI1206 05:51:46.633808 6223 services_controller.go:443] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.110\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1206 05:51:46.633876 6223 services_controller.go:444] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1206 05:51:46.633966 6223 services_controller.go:445] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nI1206 05:51:46.633753 6223 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-operators\\\\\\\"}\\\\nI1206 05:51:46.634043 6223 services_controller.go:360] Finished syncing service redhat-operators on namespace openshift-marketplace for network=default : 1.127407ms\\\\nF1206 05:51:46.633754 6223 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dee8b965be3b2486f1f532f2bdfbbcbd84f9b3b711a11c25e86dae1284314d4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5xvvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:07Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.118297 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.118335 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.118348 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.118363 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.118374 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:07Z","lastTransitionTime":"2025-12-06T05:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.203072 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.203217 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.203250 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.203278 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:52:07 crc kubenswrapper[4809]: E1206 05:52:07.203332 4809 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 05:52:07 crc kubenswrapper[4809]: E1206 05:52:07.203351 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 05:52:07 crc kubenswrapper[4809]: E1206 05:52:07.203363 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 05:52:07 crc kubenswrapper[4809]: E1206 05:52:07.203373 4809 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:52:07 crc kubenswrapper[4809]: E1206 05:52:07.203384 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:52:39.203337338 +0000 UTC m=+84.092320320 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:52:07 crc kubenswrapper[4809]: E1206 05:52:07.203429 4809 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 05:52:07 crc kubenswrapper[4809]: E1206 05:52:07.203463 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 05:52:39.203445601 +0000 UTC m=+84.092428553 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 05:52:07 crc kubenswrapper[4809]: E1206 05:52:07.203499 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 05:52:39.203486322 +0000 UTC m=+84.092469274 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:52:07 crc kubenswrapper[4809]: E1206 05:52:07.203520 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 05:52:39.203511222 +0000 UTC m=+84.092494364 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.222142 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.222203 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.222215 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.222237 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.222253 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:07Z","lastTransitionTime":"2025-12-06T05:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.304628 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:52:07 crc kubenswrapper[4809]: E1206 05:52:07.304993 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 05:52:07 crc kubenswrapper[4809]: E1206 05:52:07.305056 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 05:52:07 crc kubenswrapper[4809]: E1206 05:52:07.305072 4809 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:52:07 crc kubenswrapper[4809]: E1206 05:52:07.305167 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 05:52:39.305144568 +0000 UTC m=+84.194127500 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.325160 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.325211 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.325225 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.325245 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.325258 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:07Z","lastTransitionTime":"2025-12-06T05:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.388487 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.388592 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:52:07 crc kubenswrapper[4809]: E1206 05:52:07.388683 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:52:07 crc kubenswrapper[4809]: E1206 05:52:07.388813 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.428379 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.428440 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.428456 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.428477 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.428493 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:07Z","lastTransitionTime":"2025-12-06T05:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.534095 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.534163 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.534179 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.534206 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.534222 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:07Z","lastTransitionTime":"2025-12-06T05:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.637017 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.637059 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.637071 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.637086 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.637098 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:07Z","lastTransitionTime":"2025-12-06T05:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.740050 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.740112 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.740122 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.740139 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.740152 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:07Z","lastTransitionTime":"2025-12-06T05:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.759144 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5xvvb_ffb3a4f1-1e22-4220-ac80-3b2d69d2db99/ovnkube-controller/2.log" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.759736 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5xvvb_ffb3a4f1-1e22-4220-ac80-3b2d69d2db99/ovnkube-controller/1.log" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.762579 4809 generic.go:334] "Generic (PLEG): container finished" podID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerID="80ffc0fe836258c86321ba5a7f8e1c7e289c0f85e0b48bc138f69cf96cf225ab" exitCode=1 Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.762630 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" event={"ID":"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99","Type":"ContainerDied","Data":"80ffc0fe836258c86321ba5a7f8e1c7e289c0f85e0b48bc138f69cf96cf225ab"} Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.762677 4809 scope.go:117] "RemoveContainer" containerID="f973788e6cfc8d6032d0a82dd6dceec13bb5ebdd481636c561d03f469b8746dd" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.763774 4809 scope.go:117] "RemoveContainer" containerID="80ffc0fe836258c86321ba5a7f8e1c7e289c0f85e0b48bc138f69cf96cf225ab" Dec 06 05:52:07 crc kubenswrapper[4809]: E1206 05:52:07.764269 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5xvvb_openshift-ovn-kubernetes(ffb3a4f1-1e22-4220-ac80-3b2d69d2db99)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.778635 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f28d0a3aec8a36273fa4480c552d80a5ee23210203861811e37f27bccda1147e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40791d93876595c4093814146d269b8fe70efb64ef2888e03c25aebef0e9295e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:07Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.796693 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a48de155-7389-48db-8f86-d175b7db70c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://701274e49f68efbc0a939bc08771288c50662cef9c59da73170f7f114e222337\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9t7pq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:07Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.811213 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nndh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a80ba5e1-90cc-44ce-be19-fdf0c007d5b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3684c4e5e839ce117a1772571ae080042e0187ea1aaca635edca07ac10d94c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2c22k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31d74665ae370eed09b8a3195bd3d122ad64c6e33bf55fc42cd40a862671b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2c22k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nndh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:07Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.825892 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cb4vc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed42d686-41af-470e-910d-a1fabbec66b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f629b20d9f61227553520b30d0ca046750433926a8d31501e5fb77467426cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2nwpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cb4vc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:07Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.839903 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hmlmc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c7558c-89b1-432d-893c-ccb2b3aba183\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c2f8876171399b61abd26285afba0547dac3c24c957bd6d202721b04c912aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm7wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hmlmc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:07Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.843019 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.843072 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.843088 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.843107 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.843124 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:07Z","lastTransitionTime":"2025-12-06T05:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.861377 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa4dd3f8-c119-46de-ad4e-dd29bb11a32c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa59e4ec9867cf280b308ee3adb5e303eaf94455aca588f5eb74a2b6263d57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5720f7f2977d9af6ed5324a30d08d34a53d55460951b3875b7d7d753d202f349\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dfd977b049a8ff81192d4041adb265366ab88475ee84c5750305d061d21c8db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d5801b61a87c95c9a623c3106e305920199c640efa8a32cf57698245697e54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:07Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.877543 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://010c2681c08dc21bc3a6fb82c04ab3dc2c26dcb76ec12b143c22845cac04ed25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:07Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.900389 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c981561d51870deffeb446e8ca769b7ac0a1a16edbe4e982e0f2c84a45043db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d58d489f532b78aca59f5b92584aad1dc292e6700c112c531cbeb3209fd90c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dcd421b38e5f9ba0c542a09f0047bde6c043942a096d11575afb7423ab2552c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab9b3bd4c8a6d6af7939b06fe01b8a0a50d81bb8c669bc6a27f486953c234e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a3e4453935871bb84d68bc4a8b6c98f299b7c8094ba82291bc90e258e1fec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5be86696b7a6518c5df930accb2c6890d58cfc37104fd34a359724b9e30231b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ffc0fe836258c86321ba5a7f8e1c7e289c0f85e0b48bc138f69cf96cf225ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f973788e6cfc8d6032d0a82dd6dceec13bb5ebdd481636c561d03f469b8746dd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:51:47Z\\\",\\\"message\\\":\\\"ackage-server-manager-metrics for network=default are: map[]\\\\nI1206 05:51:46.633808 6223 services_controller.go:443] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.110\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1206 05:51:46.633876 6223 services_controller.go:444] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1206 05:51:46.633966 6223 services_controller.go:445] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nI1206 05:51:46.633753 6223 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-operators\\\\\\\"}\\\\nI1206 05:51:46.634043 6223 services_controller.go:360] Finished syncing service redhat-operators on namespace openshift-marketplace for network=default : 1.127407ms\\\\nF1206 05:51:46.633754 6223 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80ffc0fe836258c86321ba5a7f8e1c7e289c0f85e0b48bc138f69cf96cf225ab\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:52:07Z\\\",\\\"message\\\":\\\"52:06.762839 6453 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.59 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {dce28c51-c9f1-478b-97c8-7e209d6e7cbe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 05:52:06.762813 6453 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-dns-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"d937b3b3-82c3-4791-9a66-41b9fed53e9d\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-dns-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, A\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dee8b965be3b2486f1f532f2bdfbbcbd84f9b3b711a11c25e86dae1284314d4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5xvvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:07Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.914798 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947d89be9d04856af8d09facb8d4583f91898bd03a36d3ce8b9b74bc3bb795e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://930788ec9b4d97fdc39353a93111b84707037ce445f1930aa7737a20a44e1d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-npms2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:07Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.932961 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5636dd78-3d61-4c6a-b8fd-6d2457e19234\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:07Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.946203 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.946250 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.946261 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.946279 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.946291 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:07Z","lastTransitionTime":"2025-12-06T05:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.946621 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"353c3bc7-39c3-4f30-ab37-9883caae33f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d4b8272a98d4acd09f165939684c0990b9e56b1269f85c20d2bd41c0f75f67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ddfff8cca35bda4796fe58ace60506f5c87e4398193d6711e81aae0d4b7d729\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae336ca136d0a51627e63b1ec86bd4f916a5efaf804e5b91014188331c19802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6965798de9aaf2b8066b7bd8abc23e12545f026bf1f0930dc0085b992553f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc6965798de9aaf2b8066b7bd8abc23e12545f026bf1f0930dc0085b992553f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:07Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.961451 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd45866c51410e910e87d43870bf531b0d05541e9164e700150fd37987e284af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:07Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.980725 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:07Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:07 crc kubenswrapper[4809]: I1206 05:52:07.997731 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:07Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.013925 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:08Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.028897 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rnzq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0075197-9392-4109-b8a7-39507b15cc17\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53a00b17138b236dc8517cbd4ee71ea3ed7e8885247d06eb2e8840b8b145783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4mgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rnzq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:08Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.041832 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9k8zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4f75fd3-03e8-41e7-8926-f6ff21a5b681\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpdh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpdh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9k8zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:08Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.049153 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.049195 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.049206 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.049225 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.049235 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:08Z","lastTransitionTime":"2025-12-06T05:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.151535 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.151573 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.151583 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.151596 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.151605 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:08Z","lastTransitionTime":"2025-12-06T05:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.253970 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.254022 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.254039 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.254061 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.254078 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:08Z","lastTransitionTime":"2025-12-06T05:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.357470 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.357513 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.357524 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.357541 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.357554 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:08Z","lastTransitionTime":"2025-12-06T05:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.388292 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.388356 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:52:08 crc kubenswrapper[4809]: E1206 05:52:08.388439 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:52:08 crc kubenswrapper[4809]: E1206 05:52:08.388522 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.461431 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.461496 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.461517 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.461543 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.461560 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:08Z","lastTransitionTime":"2025-12-06T05:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.564868 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.565049 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.565075 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.565108 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.565133 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:08Z","lastTransitionTime":"2025-12-06T05:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.668171 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.668211 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.668221 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.668235 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.668245 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:08Z","lastTransitionTime":"2025-12-06T05:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.767301 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5xvvb_ffb3a4f1-1e22-4220-ac80-3b2d69d2db99/ovnkube-controller/2.log" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.769455 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.769478 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.769487 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.769499 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.769508 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:08Z","lastTransitionTime":"2025-12-06T05:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.770887 4809 scope.go:117] "RemoveContainer" containerID="80ffc0fe836258c86321ba5a7f8e1c7e289c0f85e0b48bc138f69cf96cf225ab" Dec 06 05:52:08 crc kubenswrapper[4809]: E1206 05:52:08.771071 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5xvvb_openshift-ovn-kubernetes(ffb3a4f1-1e22-4220-ac80-3b2d69d2db99)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.784268 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa4dd3f8-c119-46de-ad4e-dd29bb11a32c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa59e4ec9867cf280b308ee3adb5e303eaf94455aca588f5eb74a2b6263d57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5720f7f2977d9af6ed5324a30d08d34a53d55460951b3875b7d7d753d202f349\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dfd977b049a8ff81192d4041adb265366ab88475ee84c5750305d061d21c8db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d5801b61a87c95c9a623c3106e305920199c640efa8a32cf57698245697e54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:08Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.798455 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://010c2681c08dc21bc3a6fb82c04ab3dc2c26dcb76ec12b143c22845cac04ed25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:08Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.817314 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c981561d51870deffeb446e8ca769b7ac0a1a16edbe4e982e0f2c84a45043db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d58d489f532b78aca59f5b92584aad1dc292e6700c112c531cbeb3209fd90c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dcd421b38e5f9ba0c542a09f0047bde6c043942a096d11575afb7423ab2552c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab9b3bd4c8a6d6af7939b06fe01b8a0a50d81bb8c669bc6a27f486953c234e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a3e4453935871bb84d68bc4a8b6c98f299b7c8094ba82291bc90e258e1fec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5be86696b7a6518c5df930accb2c6890d58cfc37104fd34a359724b9e30231b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ffc0fe836258c86321ba5a7f8e1c7e289c0f85e0b48bc138f69cf96cf225ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80ffc0fe836258c86321ba5a7f8e1c7e289c0f85e0b48bc138f69cf96cf225ab\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:52:07Z\\\",\\\"message\\\":\\\"52:06.762839 6453 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.59 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {dce28c51-c9f1-478b-97c8-7e209d6e7cbe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 05:52:06.762813 6453 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-dns-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"d937b3b3-82c3-4791-9a66-41b9fed53e9d\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-dns-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, A\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:52:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5xvvb_openshift-ovn-kubernetes(ffb3a4f1-1e22-4220-ac80-3b2d69d2db99)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dee8b965be3b2486f1f532f2bdfbbcbd84f9b3b711a11c25e86dae1284314d4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5xvvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:08Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.830867 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947d89be9d04856af8d09facb8d4583f91898bd03a36d3ce8b9b74bc3bb795e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://930788ec9b4d97fdc39353a93111b84707037ce445f1930aa7737a20a44e1d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-npms2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:08Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.850991 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5636dd78-3d61-4c6a-b8fd-6d2457e19234\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:08Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.865262 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"353c3bc7-39c3-4f30-ab37-9883caae33f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d4b8272a98d4acd09f165939684c0990b9e56b1269f85c20d2bd41c0f75f67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ddfff8cca35bda4796fe58ace60506f5c87e4398193d6711e81aae0d4b7d729\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae336ca136d0a51627e63b1ec86bd4f916a5efaf804e5b91014188331c19802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6965798de9aaf2b8066b7bd8abc23e12545f026bf1f0930dc0085b992553f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc6965798de9aaf2b8066b7bd8abc23e12545f026bf1f0930dc0085b992553f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:08Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.871408 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.871463 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.871475 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.871494 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.871512 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:08Z","lastTransitionTime":"2025-12-06T05:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.883502 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd45866c51410e910e87d43870bf531b0d05541e9164e700150fd37987e284af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:08Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.897886 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:08Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.914565 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:08Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.928926 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:08Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.942159 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rnzq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0075197-9392-4109-b8a7-39507b15cc17\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53a00b17138b236dc8517cbd4ee71ea3ed7e8885247d06eb2e8840b8b145783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4mgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rnzq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:08Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.958044 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9k8zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4f75fd3-03e8-41e7-8926-f6ff21a5b681\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpdh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpdh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9k8zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:08Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.974282 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.974319 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.974330 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.974347 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.974360 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:08Z","lastTransitionTime":"2025-12-06T05:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.978132 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f28d0a3aec8a36273fa4480c552d80a5ee23210203861811e37f27bccda1147e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40791d93876595c4093814146d269b8fe70efb64ef2888e03c25aebef0e9295e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:08Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:08 crc kubenswrapper[4809]: I1206 05:52:08.994262 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a48de155-7389-48db-8f86-d175b7db70c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://701274e49f68efbc0a939bc08771288c50662cef9c59da73170f7f114e222337\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9t7pq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:08Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.010234 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nndh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a80ba5e1-90cc-44ce-be19-fdf0c007d5b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3684c4e5e839ce117a1772571ae080042e0187ea1aaca635edca07ac10d94c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2c22k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31d74665ae370eed09b8a3195bd3d122ad64c6e33bf55fc42cd40a862671b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2c22k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nndh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:09Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.024404 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cb4vc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed42d686-41af-470e-910d-a1fabbec66b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f629b20d9f61227553520b30d0ca046750433926a8d31501e5fb77467426cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2nwpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cb4vc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:09Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.037903 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hmlmc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c7558c-89b1-432d-893c-ccb2b3aba183\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c2f8876171399b61abd26285afba0547dac3c24c957bd6d202721b04c912aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm7wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hmlmc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:09Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.076893 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.077205 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.077270 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.077335 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.077405 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:09Z","lastTransitionTime":"2025-12-06T05:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.180322 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.180351 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.180360 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.180371 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.180380 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:09Z","lastTransitionTime":"2025-12-06T05:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.283147 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.283214 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.283231 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.283255 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.283274 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:09Z","lastTransitionTime":"2025-12-06T05:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.386757 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.386831 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.386851 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.386876 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.386894 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:09Z","lastTransitionTime":"2025-12-06T05:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.387876 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:52:09 crc kubenswrapper[4809]: E1206 05:52:09.388113 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.388166 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:52:09 crc kubenswrapper[4809]: E1206 05:52:09.388336 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.489866 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.489973 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.489993 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.490017 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.490035 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:09Z","lastTransitionTime":"2025-12-06T05:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.592959 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.593004 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.593015 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.593033 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.593091 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:09Z","lastTransitionTime":"2025-12-06T05:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.695449 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.695781 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.695898 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.696017 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.696117 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:09Z","lastTransitionTime":"2025-12-06T05:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.719525 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.719578 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.719594 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.719614 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.719629 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:09Z","lastTransitionTime":"2025-12-06T05:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:09 crc kubenswrapper[4809]: E1206 05:52:09.736416 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"88a28b6b-9c85-4932-b5bc-36346303aee5\\\",\\\"systemUUID\\\":\\\"5931df91-c1df-4f78-9e5f-391a01a76127\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:09Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.741250 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.741300 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.741312 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.741329 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.741342 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:09Z","lastTransitionTime":"2025-12-06T05:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:09 crc kubenswrapper[4809]: E1206 05:52:09.757256 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"88a28b6b-9c85-4932-b5bc-36346303aee5\\\",\\\"systemUUID\\\":\\\"5931df91-c1df-4f78-9e5f-391a01a76127\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:09Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.761792 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.761842 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.761854 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.761873 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.761885 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:09Z","lastTransitionTime":"2025-12-06T05:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:09 crc kubenswrapper[4809]: E1206 05:52:09.776400 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"88a28b6b-9c85-4932-b5bc-36346303aee5\\\",\\\"systemUUID\\\":\\\"5931df91-c1df-4f78-9e5f-391a01a76127\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:09Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.781013 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.781075 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.781095 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.781118 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.781139 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:09Z","lastTransitionTime":"2025-12-06T05:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:09 crc kubenswrapper[4809]: E1206 05:52:09.796081 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"88a28b6b-9c85-4932-b5bc-36346303aee5\\\",\\\"systemUUID\\\":\\\"5931df91-c1df-4f78-9e5f-391a01a76127\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:09Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.800752 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.800793 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.800804 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.800828 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.800843 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:09Z","lastTransitionTime":"2025-12-06T05:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:09 crc kubenswrapper[4809]: E1206 05:52:09.819574 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"88a28b6b-9c85-4932-b5bc-36346303aee5\\\",\\\"systemUUID\\\":\\\"5931df91-c1df-4f78-9e5f-391a01a76127\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:09Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:09 crc kubenswrapper[4809]: E1206 05:52:09.819868 4809 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.822052 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.822120 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.822145 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.822175 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.822197 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:09Z","lastTransitionTime":"2025-12-06T05:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.925281 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.925314 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.925324 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.925340 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:09 crc kubenswrapper[4809]: I1206 05:52:09.925350 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:09Z","lastTransitionTime":"2025-12-06T05:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.028640 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.028729 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.028741 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.028757 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.028766 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:10Z","lastTransitionTime":"2025-12-06T05:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.132255 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.132314 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.132333 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.132357 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.132377 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:10Z","lastTransitionTime":"2025-12-06T05:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.235447 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.235479 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.235487 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.235499 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.235508 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:10Z","lastTransitionTime":"2025-12-06T05:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.338229 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.338270 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.338281 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.338298 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.338310 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:10Z","lastTransitionTime":"2025-12-06T05:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.387772 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:52:10 crc kubenswrapper[4809]: E1206 05:52:10.387955 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.387789 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:52:10 crc kubenswrapper[4809]: E1206 05:52:10.388155 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.445652 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.445790 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.445818 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.445854 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.445892 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:10Z","lastTransitionTime":"2025-12-06T05:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.549416 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.549491 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.549502 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.549523 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.549536 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:10Z","lastTransitionTime":"2025-12-06T05:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.652526 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.652584 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.652594 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.652615 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.652627 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:10Z","lastTransitionTime":"2025-12-06T05:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.755744 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.755785 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.755796 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.755813 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.755823 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:10Z","lastTransitionTime":"2025-12-06T05:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.859058 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.859113 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.859126 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.859145 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.859156 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:10Z","lastTransitionTime":"2025-12-06T05:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.961516 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.961583 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.961598 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.961616 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:10 crc kubenswrapper[4809]: I1206 05:52:10.961650 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:10Z","lastTransitionTime":"2025-12-06T05:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.064658 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.064756 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.064765 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.064778 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.064787 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:11Z","lastTransitionTime":"2025-12-06T05:52:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.167076 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.167118 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.167129 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.167143 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.167154 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:11Z","lastTransitionTime":"2025-12-06T05:52:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.270893 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.270994 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.271015 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.271040 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.271058 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:11Z","lastTransitionTime":"2025-12-06T05:52:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.374113 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.374174 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.374188 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.374207 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.374220 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:11Z","lastTransitionTime":"2025-12-06T05:52:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.387968 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:52:11 crc kubenswrapper[4809]: E1206 05:52:11.388138 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.388227 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:52:11 crc kubenswrapper[4809]: E1206 05:52:11.388454 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.477289 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.477336 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.477348 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.477364 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.477375 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:11Z","lastTransitionTime":"2025-12-06T05:52:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.579674 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.579742 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.579768 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.579790 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.579805 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:11Z","lastTransitionTime":"2025-12-06T05:52:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.682748 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.682821 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.682879 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.682911 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.682988 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:11Z","lastTransitionTime":"2025-12-06T05:52:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.784922 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.784998 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.785010 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.785027 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.785040 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:11Z","lastTransitionTime":"2025-12-06T05:52:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.887653 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.887711 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.887726 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.887747 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.887760 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:11Z","lastTransitionTime":"2025-12-06T05:52:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.991262 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.991328 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.991343 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.991363 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:11 crc kubenswrapper[4809]: I1206 05:52:11.991378 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:11Z","lastTransitionTime":"2025-12-06T05:52:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:12 crc kubenswrapper[4809]: I1206 05:52:12.094527 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:12 crc kubenswrapper[4809]: I1206 05:52:12.094573 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:12 crc kubenswrapper[4809]: I1206 05:52:12.094585 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:12 crc kubenswrapper[4809]: I1206 05:52:12.094607 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:12 crc kubenswrapper[4809]: I1206 05:52:12.094618 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:12Z","lastTransitionTime":"2025-12-06T05:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:12 crc kubenswrapper[4809]: I1206 05:52:12.198216 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:12 crc kubenswrapper[4809]: I1206 05:52:12.198270 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:12 crc kubenswrapper[4809]: I1206 05:52:12.198280 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:12 crc kubenswrapper[4809]: I1206 05:52:12.198302 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:12 crc kubenswrapper[4809]: I1206 05:52:12.198312 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:12Z","lastTransitionTime":"2025-12-06T05:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:12 crc kubenswrapper[4809]: I1206 05:52:12.300692 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:12 crc kubenswrapper[4809]: I1206 05:52:12.300764 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:12 crc kubenswrapper[4809]: I1206 05:52:12.300775 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:12 crc kubenswrapper[4809]: I1206 05:52:12.300792 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:12 crc kubenswrapper[4809]: I1206 05:52:12.300802 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:12Z","lastTransitionTime":"2025-12-06T05:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:12 crc kubenswrapper[4809]: I1206 05:52:12.387500 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:52:12 crc kubenswrapper[4809]: I1206 05:52:12.387569 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:52:12 crc kubenswrapper[4809]: E1206 05:52:12.387661 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:52:12 crc kubenswrapper[4809]: E1206 05:52:12.388020 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:52:12 crc kubenswrapper[4809]: I1206 05:52:12.403021 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:12 crc kubenswrapper[4809]: I1206 05:52:12.403049 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:12 crc kubenswrapper[4809]: I1206 05:52:12.403057 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:12 crc kubenswrapper[4809]: I1206 05:52:12.403069 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:12 crc kubenswrapper[4809]: I1206 05:52:12.403077 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:12Z","lastTransitionTime":"2025-12-06T05:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:12 crc kubenswrapper[4809]: I1206 05:52:12.504833 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:12 crc kubenswrapper[4809]: I1206 05:52:12.504884 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:12 crc kubenswrapper[4809]: I1206 05:52:12.504898 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:12 crc kubenswrapper[4809]: I1206 05:52:12.504916 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:12 crc kubenswrapper[4809]: I1206 05:52:12.504926 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:12Z","lastTransitionTime":"2025-12-06T05:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:12 crc kubenswrapper[4809]: I1206 05:52:12.607476 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:12 crc kubenswrapper[4809]: I1206 05:52:12.607577 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:12 crc kubenswrapper[4809]: I1206 05:52:12.607594 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:12 crc kubenswrapper[4809]: I1206 05:52:12.607610 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:12 crc kubenswrapper[4809]: I1206 05:52:12.607619 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:12Z","lastTransitionTime":"2025-12-06T05:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:12 crc kubenswrapper[4809]: I1206 05:52:12.710251 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:12 crc kubenswrapper[4809]: I1206 05:52:12.710308 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:12 crc kubenswrapper[4809]: I1206 05:52:12.710329 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:12 crc kubenswrapper[4809]: I1206 05:52:12.710354 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:12 crc kubenswrapper[4809]: I1206 05:52:12.710366 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:12Z","lastTransitionTime":"2025-12-06T05:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:12 crc kubenswrapper[4809]: I1206 05:52:12.813302 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:12 crc kubenswrapper[4809]: I1206 05:52:12.813371 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:12 crc kubenswrapper[4809]: I1206 05:52:12.813393 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:12 crc kubenswrapper[4809]: I1206 05:52:12.813422 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:12 crc kubenswrapper[4809]: I1206 05:52:12.813446 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:12Z","lastTransitionTime":"2025-12-06T05:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:12 crc kubenswrapper[4809]: I1206 05:52:12.916507 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:12 crc kubenswrapper[4809]: I1206 05:52:12.916571 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:12 crc kubenswrapper[4809]: I1206 05:52:12.916588 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:12 crc kubenswrapper[4809]: I1206 05:52:12.916612 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:12 crc kubenswrapper[4809]: I1206 05:52:12.916632 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:12Z","lastTransitionTime":"2025-12-06T05:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.019762 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.019828 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.019845 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.019869 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.019886 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:13Z","lastTransitionTime":"2025-12-06T05:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.122766 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.122843 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.122867 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.122896 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.122919 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:13Z","lastTransitionTime":"2025-12-06T05:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.230416 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.230478 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.230495 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.230518 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.230535 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:13Z","lastTransitionTime":"2025-12-06T05:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.333360 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.333398 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.333408 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.333420 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.333430 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:13Z","lastTransitionTime":"2025-12-06T05:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.388376 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:52:13 crc kubenswrapper[4809]: E1206 05:52:13.388541 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.388977 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:52:13 crc kubenswrapper[4809]: E1206 05:52:13.389129 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.435638 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.435684 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.435698 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.435715 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.435726 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:13Z","lastTransitionTime":"2025-12-06T05:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.538564 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.538630 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.538655 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.538683 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.538700 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:13Z","lastTransitionTime":"2025-12-06T05:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.641642 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.641693 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.641703 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.641720 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.641732 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:13Z","lastTransitionTime":"2025-12-06T05:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.745097 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.745157 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.745168 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.745187 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.745199 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:13Z","lastTransitionTime":"2025-12-06T05:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.847873 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.847970 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.847993 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.848015 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.848031 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:13Z","lastTransitionTime":"2025-12-06T05:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.950983 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.951067 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.951092 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.951119 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:13 crc kubenswrapper[4809]: I1206 05:52:13.951141 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:13Z","lastTransitionTime":"2025-12-06T05:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.053983 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.054026 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.054036 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.054050 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.054060 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:14Z","lastTransitionTime":"2025-12-06T05:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.157406 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.157501 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.157532 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.157562 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.157585 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:14Z","lastTransitionTime":"2025-12-06T05:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.260880 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.260990 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.261013 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.261042 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.261063 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:14Z","lastTransitionTime":"2025-12-06T05:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.364027 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.364137 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.364161 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.364190 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.364213 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:14Z","lastTransitionTime":"2025-12-06T05:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.387747 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.387764 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:52:14 crc kubenswrapper[4809]: E1206 05:52:14.387917 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:52:14 crc kubenswrapper[4809]: E1206 05:52:14.388103 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.467753 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.467811 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.467832 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.467859 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.467880 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:14Z","lastTransitionTime":"2025-12-06T05:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.570677 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.570731 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.570742 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.570757 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.570766 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:14Z","lastTransitionTime":"2025-12-06T05:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.673201 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.673245 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.673257 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.673275 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.673289 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:14Z","lastTransitionTime":"2025-12-06T05:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.775972 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.776025 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.776042 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.776062 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.776073 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:14Z","lastTransitionTime":"2025-12-06T05:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.878875 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.878991 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.879018 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.879047 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.879075 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:14Z","lastTransitionTime":"2025-12-06T05:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.981895 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.981946 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.981956 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.981969 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:14 crc kubenswrapper[4809]: I1206 05:52:14.981979 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:14Z","lastTransitionTime":"2025-12-06T05:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.084633 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.084738 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.084992 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.085040 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.085062 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:15Z","lastTransitionTime":"2025-12-06T05:52:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.188224 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.188289 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.188300 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.188322 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.188343 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:15Z","lastTransitionTime":"2025-12-06T05:52:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.291295 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.291383 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.291404 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.291442 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.291473 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:15Z","lastTransitionTime":"2025-12-06T05:52:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.387500 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:52:15 crc kubenswrapper[4809]: E1206 05:52:15.388003 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.387732 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:52:15 crc kubenswrapper[4809]: E1206 05:52:15.388249 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.394234 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.394264 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.394273 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.394286 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.394296 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:15Z","lastTransitionTime":"2025-12-06T05:52:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.411137 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cb4vc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed42d686-41af-470e-910d-a1fabbec66b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f629b20d9f61227553520b30d0ca046750433926a8d31501e5fb77467426cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2nwpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cb4vc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:15Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.423174 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hmlmc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c7558c-89b1-432d-893c-ccb2b3aba183\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c2f8876171399b61abd26285afba0547dac3c24c957bd6d202721b04c912aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm7wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hmlmc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:15Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.444258 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa4dd3f8-c119-46de-ad4e-dd29bb11a32c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa59e4ec9867cf280b308ee3adb5e303eaf94455aca588f5eb74a2b6263d57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5720f7f2977d9af6ed5324a30d08d34a53d55460951b3875b7d7d753d202f349\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dfd977b049a8ff81192d4041adb265366ab88475ee84c5750305d061d21c8db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d5801b61a87c95c9a623c3106e305920199c640efa8a32cf57698245697e54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:15Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.461880 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://010c2681c08dc21bc3a6fb82c04ab3dc2c26dcb76ec12b143c22845cac04ed25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:15Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.493904 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c981561d51870deffeb446e8ca769b7ac0a1a16edbe4e982e0f2c84a45043db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d58d489f532b78aca59f5b92584aad1dc292e6700c112c531cbeb3209fd90c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dcd421b38e5f9ba0c542a09f0047bde6c043942a096d11575afb7423ab2552c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab9b3bd4c8a6d6af7939b06fe01b8a0a50d81bb8c669bc6a27f486953c234e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a3e4453935871bb84d68bc4a8b6c98f299b7c8094ba82291bc90e258e1fec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5be86696b7a6518c5df930accb2c6890d58cfc37104fd34a359724b9e30231b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ffc0fe836258c86321ba5a7f8e1c7e289c0f85e0b48bc138f69cf96cf225ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80ffc0fe836258c86321ba5a7f8e1c7e289c0f85e0b48bc138f69cf96cf225ab\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:52:07Z\\\",\\\"message\\\":\\\"52:06.762839 6453 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.59 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {dce28c51-c9f1-478b-97c8-7e209d6e7cbe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 05:52:06.762813 6453 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-dns-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"d937b3b3-82c3-4791-9a66-41b9fed53e9d\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-dns-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, A\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:52:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5xvvb_openshift-ovn-kubernetes(ffb3a4f1-1e22-4220-ac80-3b2d69d2db99)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dee8b965be3b2486f1f532f2bdfbbcbd84f9b3b711a11c25e86dae1284314d4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5xvvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:15Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.497627 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.497676 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.497694 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.497719 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.497736 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:15Z","lastTransitionTime":"2025-12-06T05:52:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.511820 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947d89be9d04856af8d09facb8d4583f91898bd03a36d3ce8b9b74bc3bb795e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://930788ec9b4d97fdc39353a93111b84707037ce445f1930aa7737a20a44e1d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-npms2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:15Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.530066 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5636dd78-3d61-4c6a-b8fd-6d2457e19234\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:15Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.547625 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"353c3bc7-39c3-4f30-ab37-9883caae33f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d4b8272a98d4acd09f165939684c0990b9e56b1269f85c20d2bd41c0f75f67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ddfff8cca35bda4796fe58ace60506f5c87e4398193d6711e81aae0d4b7d729\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae336ca136d0a51627e63b1ec86bd4f916a5efaf804e5b91014188331c19802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6965798de9aaf2b8066b7bd8abc23e12545f026bf1f0930dc0085b992553f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc6965798de9aaf2b8066b7bd8abc23e12545f026bf1f0930dc0085b992553f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:15Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.567334 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd45866c51410e910e87d43870bf531b0d05541e9164e700150fd37987e284af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:15Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.586366 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:15Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.598205 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:15Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.599944 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.599980 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.599991 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.600007 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.600020 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:15Z","lastTransitionTime":"2025-12-06T05:52:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.611334 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:15Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.623797 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rnzq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0075197-9392-4109-b8a7-39507b15cc17\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53a00b17138b236dc8517cbd4ee71ea3ed7e8885247d06eb2e8840b8b145783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4mgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rnzq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:15Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.635202 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9k8zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4f75fd3-03e8-41e7-8926-f6ff21a5b681\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpdh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpdh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9k8zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:15Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.648422 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f28d0a3aec8a36273fa4480c552d80a5ee23210203861811e37f27bccda1147e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40791d93876595c4093814146d269b8fe70efb64ef2888e03c25aebef0e9295e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:15Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.663456 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a48de155-7389-48db-8f86-d175b7db70c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://701274e49f68efbc0a939bc08771288c50662cef9c59da73170f7f114e222337\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9t7pq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:15Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.678541 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nndh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a80ba5e1-90cc-44ce-be19-fdf0c007d5b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3684c4e5e839ce117a1772571ae080042e0187ea1aaca635edca07ac10d94c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2c22k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31d74665ae370eed09b8a3195bd3d122ad64c6e33bf55fc42cd40a862671b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2c22k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nndh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:15Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.702596 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.702636 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.702647 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.702663 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.702674 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:15Z","lastTransitionTime":"2025-12-06T05:52:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.805070 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.805132 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.805150 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.805176 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.805195 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:15Z","lastTransitionTime":"2025-12-06T05:52:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.908705 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.908752 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.908764 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.908781 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:15 crc kubenswrapper[4809]: I1206 05:52:15.908794 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:15Z","lastTransitionTime":"2025-12-06T05:52:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.011421 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.011487 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.011513 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.011534 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.011560 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:16Z","lastTransitionTime":"2025-12-06T05:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.113777 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.113812 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.113831 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.113847 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.113858 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:16Z","lastTransitionTime":"2025-12-06T05:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.216980 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.217044 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.217067 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.217095 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.217117 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:16Z","lastTransitionTime":"2025-12-06T05:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.320788 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.320888 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.320900 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.320919 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.320950 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:16Z","lastTransitionTime":"2025-12-06T05:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.387821 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:52:16 crc kubenswrapper[4809]: E1206 05:52:16.388002 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.388400 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:52:16 crc kubenswrapper[4809]: E1206 05:52:16.388717 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.424475 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.424540 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.424562 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.424590 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.424619 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:16Z","lastTransitionTime":"2025-12-06T05:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.527677 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.527759 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.527785 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.527821 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.527844 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:16Z","lastTransitionTime":"2025-12-06T05:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.630347 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.630396 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.630407 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.630426 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.630438 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:16Z","lastTransitionTime":"2025-12-06T05:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.733517 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.733568 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.733578 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.733596 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.733610 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:16Z","lastTransitionTime":"2025-12-06T05:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.835884 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.836003 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.836026 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.836111 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.836132 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:16Z","lastTransitionTime":"2025-12-06T05:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.938325 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.938369 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.938377 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.938392 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:16 crc kubenswrapper[4809]: I1206 05:52:16.938405 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:16Z","lastTransitionTime":"2025-12-06T05:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.040731 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.040787 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.040806 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.040865 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.040884 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:17Z","lastTransitionTime":"2025-12-06T05:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.145206 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.145283 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.145301 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.145326 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.145344 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:17Z","lastTransitionTime":"2025-12-06T05:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.248507 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.248603 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.248621 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.248650 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.248669 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:17Z","lastTransitionTime":"2025-12-06T05:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.350921 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.350979 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.350999 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.351015 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.351023 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:17Z","lastTransitionTime":"2025-12-06T05:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.387855 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.387949 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:52:17 crc kubenswrapper[4809]: E1206 05:52:17.388037 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:52:17 crc kubenswrapper[4809]: E1206 05:52:17.388093 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.453644 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.453696 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.453712 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.453733 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.453750 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:17Z","lastTransitionTime":"2025-12-06T05:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.557062 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.557107 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.557124 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.557145 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.557162 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:17Z","lastTransitionTime":"2025-12-06T05:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.659526 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.659572 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.659588 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.659611 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.659627 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:17Z","lastTransitionTime":"2025-12-06T05:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.761942 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.761999 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.762008 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.762023 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.762035 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:17Z","lastTransitionTime":"2025-12-06T05:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.864555 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.864628 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.864645 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.864663 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.864676 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:17Z","lastTransitionTime":"2025-12-06T05:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.968032 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.968089 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.968107 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.968131 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:17 crc kubenswrapper[4809]: I1206 05:52:17.968150 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:17Z","lastTransitionTime":"2025-12-06T05:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.070801 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.070836 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.070846 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.070862 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.070873 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:18Z","lastTransitionTime":"2025-12-06T05:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.173157 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.173487 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.173510 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.173546 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.173559 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:18Z","lastTransitionTime":"2025-12-06T05:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.279705 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.280031 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.280114 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.280184 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.280247 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:18Z","lastTransitionTime":"2025-12-06T05:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.382369 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.382468 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.382487 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.382509 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.382525 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:18Z","lastTransitionTime":"2025-12-06T05:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.388142 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:52:18 crc kubenswrapper[4809]: E1206 05:52:18.388343 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.388156 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:52:18 crc kubenswrapper[4809]: E1206 05:52:18.388614 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.484272 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.484320 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.484332 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.484348 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.484358 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:18Z","lastTransitionTime":"2025-12-06T05:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.587219 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.587266 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.587279 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.587296 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.587307 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:18Z","lastTransitionTime":"2025-12-06T05:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.690461 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.690507 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.690518 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.690533 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.690542 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:18Z","lastTransitionTime":"2025-12-06T05:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.793425 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.793483 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.793496 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.793514 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.793525 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:18Z","lastTransitionTime":"2025-12-06T05:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.896644 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.896982 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.897072 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.897166 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.897250 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:18Z","lastTransitionTime":"2025-12-06T05:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.999345 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.999413 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:18 crc kubenswrapper[4809]: I1206 05:52:18.999434 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:18.999461 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:18.999485 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:18Z","lastTransitionTime":"2025-12-06T05:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:19.101679 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:19.102082 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:19.102270 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:19.102426 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:19.102566 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:19Z","lastTransitionTime":"2025-12-06T05:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:19.204815 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:19.205132 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:19.205284 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:19.205434 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:19.205550 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:19Z","lastTransitionTime":"2025-12-06T05:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:19.308279 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:19.308589 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:19.308682 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:19.308780 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:19.308916 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:19Z","lastTransitionTime":"2025-12-06T05:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:19.387765 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:52:19 crc kubenswrapper[4809]: E1206 05:52:19.388137 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:19.387869 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:52:19 crc kubenswrapper[4809]: E1206 05:52:19.388331 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:19.411837 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:19.411887 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:19.411917 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:19.411975 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:19.411996 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:19Z","lastTransitionTime":"2025-12-06T05:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:19.514550 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:19.514585 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:19.514599 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:19.514616 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:19.514628 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:19Z","lastTransitionTime":"2025-12-06T05:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:19.620725 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:19.620764 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:19.620798 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:19.620812 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:19.620821 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:19Z","lastTransitionTime":"2025-12-06T05:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:19.723643 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:19.723702 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:19.723720 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:19.723744 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:19.723761 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:19Z","lastTransitionTime":"2025-12-06T05:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:19.826461 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:19.826810 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:19.826980 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:19.827122 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:19.827262 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:19Z","lastTransitionTime":"2025-12-06T05:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:19.930559 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:19.930611 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:19.930625 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:19.930645 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:19 crc kubenswrapper[4809]: I1206 05:52:19.930675 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:19Z","lastTransitionTime":"2025-12-06T05:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.033712 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.033752 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.033762 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.033779 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.033789 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:20Z","lastTransitionTime":"2025-12-06T05:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.082491 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.082553 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.082579 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.082698 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.082722 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:20Z","lastTransitionTime":"2025-12-06T05:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:20 crc kubenswrapper[4809]: E1206 05:52:20.101100 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"88a28b6b-9c85-4932-b5bc-36346303aee5\\\",\\\"systemUUID\\\":\\\"5931df91-c1df-4f78-9e5f-391a01a76127\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:20Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.104689 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.104732 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.104750 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.104774 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.104792 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:20Z","lastTransitionTime":"2025-12-06T05:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:20 crc kubenswrapper[4809]: E1206 05:52:20.120203 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"88a28b6b-9c85-4932-b5bc-36346303aee5\\\",\\\"systemUUID\\\":\\\"5931df91-c1df-4f78-9e5f-391a01a76127\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:20Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.124428 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.124477 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.124488 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.124508 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.124524 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:20Z","lastTransitionTime":"2025-12-06T05:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:20 crc kubenswrapper[4809]: E1206 05:52:20.139299 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"88a28b6b-9c85-4932-b5bc-36346303aee5\\\",\\\"systemUUID\\\":\\\"5931df91-c1df-4f78-9e5f-391a01a76127\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:20Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.142561 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.142629 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.142654 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.142684 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.142709 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:20Z","lastTransitionTime":"2025-12-06T05:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:20 crc kubenswrapper[4809]: E1206 05:52:20.161840 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"88a28b6b-9c85-4932-b5bc-36346303aee5\\\",\\\"systemUUID\\\":\\\"5931df91-c1df-4f78-9e5f-391a01a76127\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:20Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.166009 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.166058 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.166077 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.166099 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.166115 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:20Z","lastTransitionTime":"2025-12-06T05:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:20 crc kubenswrapper[4809]: E1206 05:52:20.178595 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"88a28b6b-9c85-4932-b5bc-36346303aee5\\\",\\\"systemUUID\\\":\\\"5931df91-c1df-4f78-9e5f-391a01a76127\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:20Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:20 crc kubenswrapper[4809]: E1206 05:52:20.178910 4809 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.180793 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.180820 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.180829 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.180841 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.180849 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:20Z","lastTransitionTime":"2025-12-06T05:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.283148 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.283201 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.283210 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.283224 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.283233 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:20Z","lastTransitionTime":"2025-12-06T05:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.385171 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.385246 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.385273 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.385301 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.385322 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:20Z","lastTransitionTime":"2025-12-06T05:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.388444 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:52:20 crc kubenswrapper[4809]: E1206 05:52:20.388569 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.388580 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:52:20 crc kubenswrapper[4809]: E1206 05:52:20.389094 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.389297 4809 scope.go:117] "RemoveContainer" containerID="80ffc0fe836258c86321ba5a7f8e1c7e289c0f85e0b48bc138f69cf96cf225ab" Dec 06 05:52:20 crc kubenswrapper[4809]: E1206 05:52:20.389462 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5xvvb_openshift-ovn-kubernetes(ffb3a4f1-1e22-4220-ac80-3b2d69d2db99)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.488026 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.488064 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.488077 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.488094 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.488106 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:20Z","lastTransitionTime":"2025-12-06T05:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.590502 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.590539 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.590547 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.590559 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.590571 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:20Z","lastTransitionTime":"2025-12-06T05:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.692633 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.692678 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.692689 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.692708 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.692719 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:20Z","lastTransitionTime":"2025-12-06T05:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.798681 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.798745 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.798763 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.798790 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.798814 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:20Z","lastTransitionTime":"2025-12-06T05:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.901596 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.902177 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.902337 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.902487 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.902640 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:20Z","lastTransitionTime":"2025-12-06T05:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:20 crc kubenswrapper[4809]: I1206 05:52:20.949142 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a4f75fd3-03e8-41e7-8926-f6ff21a5b681-metrics-certs\") pod \"network-metrics-daemon-9k8zx\" (UID: \"a4f75fd3-03e8-41e7-8926-f6ff21a5b681\") " pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:52:20 crc kubenswrapper[4809]: E1206 05:52:20.949250 4809 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 05:52:20 crc kubenswrapper[4809]: E1206 05:52:20.949299 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a4f75fd3-03e8-41e7-8926-f6ff21a5b681-metrics-certs podName:a4f75fd3-03e8-41e7-8926-f6ff21a5b681 nodeName:}" failed. No retries permitted until 2025-12-06 05:52:52.949285803 +0000 UTC m=+97.838268745 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a4f75fd3-03e8-41e7-8926-f6ff21a5b681-metrics-certs") pod "network-metrics-daemon-9k8zx" (UID: "a4f75fd3-03e8-41e7-8926-f6ff21a5b681") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.005465 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.005768 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.005856 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.005980 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.006080 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:21Z","lastTransitionTime":"2025-12-06T05:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.109067 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.109108 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.109119 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.109134 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.109144 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:21Z","lastTransitionTime":"2025-12-06T05:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.211723 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.211773 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.211791 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.211815 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.211832 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:21Z","lastTransitionTime":"2025-12-06T05:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.314476 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.314797 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.315059 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.315276 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.315473 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:21Z","lastTransitionTime":"2025-12-06T05:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.387750 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:52:21 crc kubenswrapper[4809]: E1206 05:52:21.388373 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.387795 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:52:21 crc kubenswrapper[4809]: E1206 05:52:21.388726 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.417750 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.417807 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.417826 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.417850 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.417867 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:21Z","lastTransitionTime":"2025-12-06T05:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.519763 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.520099 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.520345 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.520548 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.520697 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:21Z","lastTransitionTime":"2025-12-06T05:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.623484 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.623528 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.623545 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.623567 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.623585 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:21Z","lastTransitionTime":"2025-12-06T05:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.726074 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.726112 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.726124 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.726141 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.726153 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:21Z","lastTransitionTime":"2025-12-06T05:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.829183 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.829241 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.829259 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.829285 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.829305 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:21Z","lastTransitionTime":"2025-12-06T05:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.931742 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.931790 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.931803 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.931821 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:21 crc kubenswrapper[4809]: I1206 05:52:21.931835 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:21Z","lastTransitionTime":"2025-12-06T05:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.034513 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.034568 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.034579 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.034592 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.034603 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:22Z","lastTransitionTime":"2025-12-06T05:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.136835 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.136884 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.136894 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.136912 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.136975 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:22Z","lastTransitionTime":"2025-12-06T05:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.239155 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.239207 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.239219 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.239236 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.239248 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:22Z","lastTransitionTime":"2025-12-06T05:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.341868 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.341907 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.341915 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.341945 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.341956 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:22Z","lastTransitionTime":"2025-12-06T05:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.388378 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.388510 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:52:22 crc kubenswrapper[4809]: E1206 05:52:22.388638 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:52:22 crc kubenswrapper[4809]: E1206 05:52:22.388722 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.444176 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.444216 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.444229 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.444243 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.444254 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:22Z","lastTransitionTime":"2025-12-06T05:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.546450 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.546544 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.546558 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.546574 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.546586 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:22Z","lastTransitionTime":"2025-12-06T05:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.648453 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.648490 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.648497 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.648510 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.648518 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:22Z","lastTransitionTime":"2025-12-06T05:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.750381 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.750424 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.750435 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.750448 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.750458 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:22Z","lastTransitionTime":"2025-12-06T05:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.814192 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-cb4vc_ed42d686-41af-470e-910d-a1fabbec66b0/kube-multus/0.log" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.814240 4809 generic.go:334] "Generic (PLEG): container finished" podID="ed42d686-41af-470e-910d-a1fabbec66b0" containerID="73f629b20d9f61227553520b30d0ca046750433926a8d31501e5fb77467426cc" exitCode=1 Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.814272 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-cb4vc" event={"ID":"ed42d686-41af-470e-910d-a1fabbec66b0","Type":"ContainerDied","Data":"73f629b20d9f61227553520b30d0ca046750433926a8d31501e5fb77467426cc"} Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.814613 4809 scope.go:117] "RemoveContainer" containerID="73f629b20d9f61227553520b30d0ca046750433926a8d31501e5fb77467426cc" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.835258 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cb4vc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed42d686-41af-470e-910d-a1fabbec66b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f629b20d9f61227553520b30d0ca046750433926a8d31501e5fb77467426cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73f629b20d9f61227553520b30d0ca046750433926a8d31501e5fb77467426cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:52:22Z\\\",\\\"message\\\":\\\"2025-12-06T05:51:36+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_4426309b-ddbc-4734-9136-9da7465cfbc2\\\\n2025-12-06T05:51:36+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_4426309b-ddbc-4734-9136-9da7465cfbc2 to /host/opt/cni/bin/\\\\n2025-12-06T05:51:36Z [verbose] multus-daemon started\\\\n2025-12-06T05:51:36Z [verbose] Readiness Indicator file check\\\\n2025-12-06T05:52:21Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2nwpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cb4vc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:22Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.846436 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hmlmc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c7558c-89b1-432d-893c-ccb2b3aba183\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c2f8876171399b61abd26285afba0547dac3c24c957bd6d202721b04c912aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm7wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hmlmc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:22Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.851773 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.851810 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.851820 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.851835 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.851846 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:22Z","lastTransitionTime":"2025-12-06T05:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.857543 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://010c2681c08dc21bc3a6fb82c04ab3dc2c26dcb76ec12b143c22845cac04ed25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:22Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.877453 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c981561d51870deffeb446e8ca769b7ac0a1a16edbe4e982e0f2c84a45043db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d58d489f532b78aca59f5b92584aad1dc292e6700c112c531cbeb3209fd90c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dcd421b38e5f9ba0c542a09f0047bde6c043942a096d11575afb7423ab2552c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab9b3bd4c8a6d6af7939b06fe01b8a0a50d81bb8c669bc6a27f486953c234e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a3e4453935871bb84d68bc4a8b6c98f299b7c8094ba82291bc90e258e1fec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5be86696b7a6518c5df930accb2c6890d58cfc37104fd34a359724b9e30231b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ffc0fe836258c86321ba5a7f8e1c7e289c0f85e0b48bc138f69cf96cf225ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80ffc0fe836258c86321ba5a7f8e1c7e289c0f85e0b48bc138f69cf96cf225ab\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:52:07Z\\\",\\\"message\\\":\\\"52:06.762839 6453 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.59 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {dce28c51-c9f1-478b-97c8-7e209d6e7cbe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 05:52:06.762813 6453 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-dns-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"d937b3b3-82c3-4791-9a66-41b9fed53e9d\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-dns-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, A\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:52:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5xvvb_openshift-ovn-kubernetes(ffb3a4f1-1e22-4220-ac80-3b2d69d2db99)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dee8b965be3b2486f1f532f2bdfbbcbd84f9b3b711a11c25e86dae1284314d4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5xvvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:22Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.890999 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa4dd3f8-c119-46de-ad4e-dd29bb11a32c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa59e4ec9867cf280b308ee3adb5e303eaf94455aca588f5eb74a2b6263d57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5720f7f2977d9af6ed5324a30d08d34a53d55460951b3875b7d7d753d202f349\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dfd977b049a8ff81192d4041adb265366ab88475ee84c5750305d061d21c8db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d5801b61a87c95c9a623c3106e305920199c640efa8a32cf57698245697e54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:22Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.902150 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd45866c51410e910e87d43870bf531b0d05541e9164e700150fd37987e284af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:22Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.912696 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:22Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.923621 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:22Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.938024 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:22Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.953961 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.953994 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.954002 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.954014 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.954022 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:22Z","lastTransitionTime":"2025-12-06T05:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.955818 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947d89be9d04856af8d09facb8d4583f91898bd03a36d3ce8b9b74bc3bb795e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://930788ec9b4d97fdc39353a93111b84707037ce445f1930aa7737a20a44e1d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-npms2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:22Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.970821 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5636dd78-3d61-4c6a-b8fd-6d2457e19234\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:22Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.984643 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"353c3bc7-39c3-4f30-ab37-9883caae33f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d4b8272a98d4acd09f165939684c0990b9e56b1269f85c20d2bd41c0f75f67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ddfff8cca35bda4796fe58ace60506f5c87e4398193d6711e81aae0d4b7d729\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae336ca136d0a51627e63b1ec86bd4f916a5efaf804e5b91014188331c19802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6965798de9aaf2b8066b7bd8abc23e12545f026bf1f0930dc0085b992553f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc6965798de9aaf2b8066b7bd8abc23e12545f026bf1f0930dc0085b992553f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:22Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:22 crc kubenswrapper[4809]: I1206 05:52:22.995865 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rnzq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0075197-9392-4109-b8a7-39507b15cc17\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53a00b17138b236dc8517cbd4ee71ea3ed7e8885247d06eb2e8840b8b145783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4mgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rnzq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:22Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.006460 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9k8zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4f75fd3-03e8-41e7-8926-f6ff21a5b681\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpdh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpdh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9k8zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:23Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.024319 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nndh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a80ba5e1-90cc-44ce-be19-fdf0c007d5b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3684c4e5e839ce117a1772571ae080042e0187ea1aaca635edca07ac10d94c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2c22k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31d74665ae370eed09b8a3195bd3d122ad64c6e33bf55fc42cd40a862671b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2c22k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nndh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:23Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.036525 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f28d0a3aec8a36273fa4480c552d80a5ee23210203861811e37f27bccda1147e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40791d93876595c4093814146d269b8fe70efb64ef2888e03c25aebef0e9295e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:23Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.050490 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a48de155-7389-48db-8f86-d175b7db70c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://701274e49f68efbc0a939bc08771288c50662cef9c59da73170f7f114e222337\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9t7pq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:23Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.056352 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.056382 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.056391 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.056407 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.056415 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:23Z","lastTransitionTime":"2025-12-06T05:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.158411 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.158672 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.158731 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.158806 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.158869 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:23Z","lastTransitionTime":"2025-12-06T05:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.262010 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.262066 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.262076 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.262090 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.262099 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:23Z","lastTransitionTime":"2025-12-06T05:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.364241 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.364271 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.364279 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.364292 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.364301 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:23Z","lastTransitionTime":"2025-12-06T05:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.388163 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.388221 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:52:23 crc kubenswrapper[4809]: E1206 05:52:23.388307 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:52:23 crc kubenswrapper[4809]: E1206 05:52:23.388436 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.467347 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.467417 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.467428 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.467442 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.467455 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:23Z","lastTransitionTime":"2025-12-06T05:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.570764 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.570812 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.570822 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.570836 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.570848 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:23Z","lastTransitionTime":"2025-12-06T05:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.673549 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.673594 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.673606 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.673621 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.673633 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:23Z","lastTransitionTime":"2025-12-06T05:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.775800 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.775842 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.775854 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.775872 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.775885 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:23Z","lastTransitionTime":"2025-12-06T05:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.820705 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-cb4vc_ed42d686-41af-470e-910d-a1fabbec66b0/kube-multus/0.log" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.820753 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-cb4vc" event={"ID":"ed42d686-41af-470e-910d-a1fabbec66b0","Type":"ContainerStarted","Data":"2edf7fa65761bc560178a40857780a5668e183a4b01974935cdbc3540915114f"} Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.834900 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cb4vc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed42d686-41af-470e-910d-a1fabbec66b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2edf7fa65761bc560178a40857780a5668e183a4b01974935cdbc3540915114f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73f629b20d9f61227553520b30d0ca046750433926a8d31501e5fb77467426cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:52:22Z\\\",\\\"message\\\":\\\"2025-12-06T05:51:36+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_4426309b-ddbc-4734-9136-9da7465cfbc2\\\\n2025-12-06T05:51:36+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_4426309b-ddbc-4734-9136-9da7465cfbc2 to /host/opt/cni/bin/\\\\n2025-12-06T05:51:36Z [verbose] multus-daemon started\\\\n2025-12-06T05:51:36Z [verbose] Readiness Indicator file check\\\\n2025-12-06T05:52:21Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2nwpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cb4vc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:23Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.844806 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hmlmc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c7558c-89b1-432d-893c-ccb2b3aba183\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c2f8876171399b61abd26285afba0547dac3c24c957bd6d202721b04c912aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm7wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hmlmc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:23Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.862641 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa4dd3f8-c119-46de-ad4e-dd29bb11a32c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa59e4ec9867cf280b308ee3adb5e303eaf94455aca588f5eb74a2b6263d57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5720f7f2977d9af6ed5324a30d08d34a53d55460951b3875b7d7d753d202f349\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dfd977b049a8ff81192d4041adb265366ab88475ee84c5750305d061d21c8db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d5801b61a87c95c9a623c3106e305920199c640efa8a32cf57698245697e54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:23Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.874494 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://010c2681c08dc21bc3a6fb82c04ab3dc2c26dcb76ec12b143c22845cac04ed25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:23Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.878246 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.878357 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.878385 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.878414 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.878437 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:23Z","lastTransitionTime":"2025-12-06T05:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.905382 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c981561d51870deffeb446e8ca769b7ac0a1a16edbe4e982e0f2c84a45043db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d58d489f532b78aca59f5b92584aad1dc292e6700c112c531cbeb3209fd90c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dcd421b38e5f9ba0c542a09f0047bde6c043942a096d11575afb7423ab2552c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab9b3bd4c8a6d6af7939b06fe01b8a0a50d81bb8c669bc6a27f486953c234e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a3e4453935871bb84d68bc4a8b6c98f299b7c8094ba82291bc90e258e1fec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5be86696b7a6518c5df930accb2c6890d58cfc37104fd34a359724b9e30231b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ffc0fe836258c86321ba5a7f8e1c7e289c0f85e0b48bc138f69cf96cf225ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80ffc0fe836258c86321ba5a7f8e1c7e289c0f85e0b48bc138f69cf96cf225ab\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:52:07Z\\\",\\\"message\\\":\\\"52:06.762839 6453 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.59 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {dce28c51-c9f1-478b-97c8-7e209d6e7cbe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 05:52:06.762813 6453 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-dns-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"d937b3b3-82c3-4791-9a66-41b9fed53e9d\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-dns-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, A\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:52:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5xvvb_openshift-ovn-kubernetes(ffb3a4f1-1e22-4220-ac80-3b2d69d2db99)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dee8b965be3b2486f1f532f2bdfbbcbd84f9b3b711a11c25e86dae1284314d4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5xvvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:23Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.920619 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5636dd78-3d61-4c6a-b8fd-6d2457e19234\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:23Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.930718 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"353c3bc7-39c3-4f30-ab37-9883caae33f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d4b8272a98d4acd09f165939684c0990b9e56b1269f85c20d2bd41c0f75f67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ddfff8cca35bda4796fe58ace60506f5c87e4398193d6711e81aae0d4b7d729\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae336ca136d0a51627e63b1ec86bd4f916a5efaf804e5b91014188331c19802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6965798de9aaf2b8066b7bd8abc23e12545f026bf1f0930dc0085b992553f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc6965798de9aaf2b8066b7bd8abc23e12545f026bf1f0930dc0085b992553f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:23Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.947695 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd45866c51410e910e87d43870bf531b0d05541e9164e700150fd37987e284af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:23Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.963054 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:23Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.981051 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.981098 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.981109 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.981128 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.981140 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:23Z","lastTransitionTime":"2025-12-06T05:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:23 crc kubenswrapper[4809]: I1206 05:52:23.984004 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:23Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.003411 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:24Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.017597 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947d89be9d04856af8d09facb8d4583f91898bd03a36d3ce8b9b74bc3bb795e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://930788ec9b4d97fdc39353a93111b84707037ce445f1930aa7737a20a44e1d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-npms2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:24Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.031575 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rnzq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0075197-9392-4109-b8a7-39507b15cc17\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53a00b17138b236dc8517cbd4ee71ea3ed7e8885247d06eb2e8840b8b145783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4mgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rnzq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:24Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.042072 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9k8zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4f75fd3-03e8-41e7-8926-f6ff21a5b681\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpdh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpdh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9k8zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:24Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.057070 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f28d0a3aec8a36273fa4480c552d80a5ee23210203861811e37f27bccda1147e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40791d93876595c4093814146d269b8fe70efb64ef2888e03c25aebef0e9295e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:24Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.071817 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a48de155-7389-48db-8f86-d175b7db70c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://701274e49f68efbc0a939bc08771288c50662cef9c59da73170f7f114e222337\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9t7pq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:24Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.083223 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nndh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a80ba5e1-90cc-44ce-be19-fdf0c007d5b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3684c4e5e839ce117a1772571ae080042e0187ea1aaca635edca07ac10d94c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2c22k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31d74665ae370eed09b8a3195bd3d122ad64c6e33bf55fc42cd40a862671b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2c22k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nndh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:24Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.083962 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.083992 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.084002 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.084018 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.084028 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:24Z","lastTransitionTime":"2025-12-06T05:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.186378 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.186416 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.186427 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.186445 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.186454 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:24Z","lastTransitionTime":"2025-12-06T05:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.288725 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.288761 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.288771 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.288786 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.288810 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:24Z","lastTransitionTime":"2025-12-06T05:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.387988 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.388013 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:52:24 crc kubenswrapper[4809]: E1206 05:52:24.388172 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:52:24 crc kubenswrapper[4809]: E1206 05:52:24.388246 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.391021 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.391068 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.391081 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.391100 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.391112 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:24Z","lastTransitionTime":"2025-12-06T05:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.494107 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.494147 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.494159 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.494174 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.494185 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:24Z","lastTransitionTime":"2025-12-06T05:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.596867 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.596920 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.596961 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.597018 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.597035 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:24Z","lastTransitionTime":"2025-12-06T05:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.699827 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.699865 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.699877 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.699894 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.699906 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:24Z","lastTransitionTime":"2025-12-06T05:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.802119 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.802168 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.802181 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.802216 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.802229 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:24Z","lastTransitionTime":"2025-12-06T05:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.904589 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.904635 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.904646 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.904662 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:24 crc kubenswrapper[4809]: I1206 05:52:24.904673 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:24Z","lastTransitionTime":"2025-12-06T05:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.007829 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.007868 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.007879 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.007894 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.007906 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:25Z","lastTransitionTime":"2025-12-06T05:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.110431 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.110472 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.110483 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.110498 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.110508 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:25Z","lastTransitionTime":"2025-12-06T05:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.213185 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.213216 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.213225 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.213237 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.213262 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:25Z","lastTransitionTime":"2025-12-06T05:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.315254 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.315317 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.315329 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.315345 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.315356 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:25Z","lastTransitionTime":"2025-12-06T05:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.388411 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.388496 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:52:25 crc kubenswrapper[4809]: E1206 05:52:25.388580 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:52:25 crc kubenswrapper[4809]: E1206 05:52:25.388749 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.409892 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c981561d51870deffeb446e8ca769b7ac0a1a16edbe4e982e0f2c84a45043db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d58d489f532b78aca59f5b92584aad1dc292e6700c112c531cbeb3209fd90c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dcd421b38e5f9ba0c542a09f0047bde6c043942a096d11575afb7423ab2552c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab9b3bd4c8a6d6af7939b06fe01b8a0a50d81bb8c669bc6a27f486953c234e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a3e4453935871bb84d68bc4a8b6c98f299b7c8094ba82291bc90e258e1fec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5be86696b7a6518c5df930accb2c6890d58cfc37104fd34a359724b9e30231b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ffc0fe836258c86321ba5a7f8e1c7e289c0f85e0b48bc138f69cf96cf225ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80ffc0fe836258c86321ba5a7f8e1c7e289c0f85e0b48bc138f69cf96cf225ab\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:52:07Z\\\",\\\"message\\\":\\\"52:06.762839 6453 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.59 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {dce28c51-c9f1-478b-97c8-7e209d6e7cbe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 05:52:06.762813 6453 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-dns-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"d937b3b3-82c3-4791-9a66-41b9fed53e9d\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-dns-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, A\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:52:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5xvvb_openshift-ovn-kubernetes(ffb3a4f1-1e22-4220-ac80-3b2d69d2db99)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dee8b965be3b2486f1f532f2bdfbbcbd84f9b3b711a11c25e86dae1284314d4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5xvvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:25Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.417571 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.417621 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.417633 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.417651 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.417662 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:25Z","lastTransitionTime":"2025-12-06T05:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.423154 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa4dd3f8-c119-46de-ad4e-dd29bb11a32c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa59e4ec9867cf280b308ee3adb5e303eaf94455aca588f5eb74a2b6263d57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5720f7f2977d9af6ed5324a30d08d34a53d55460951b3875b7d7d753d202f349\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dfd977b049a8ff81192d4041adb265366ab88475ee84c5750305d061d21c8db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d5801b61a87c95c9a623c3106e305920199c640efa8a32cf57698245697e54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:25Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.438438 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://010c2681c08dc21bc3a6fb82c04ab3dc2c26dcb76ec12b143c22845cac04ed25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:25Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.456426 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd45866c51410e910e87d43870bf531b0d05541e9164e700150fd37987e284af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:25Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.471336 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:25Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.486088 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:25Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.498509 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:25Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.509945 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947d89be9d04856af8d09facb8d4583f91898bd03a36d3ce8b9b74bc3bb795e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://930788ec9b4d97fdc39353a93111b84707037ce445f1930aa7737a20a44e1d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-npms2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:25Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.521506 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.521543 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.521554 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.521569 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.521547 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5636dd78-3d61-4c6a-b8fd-6d2457e19234\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:25Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.521579 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:25Z","lastTransitionTime":"2025-12-06T05:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.532754 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"353c3bc7-39c3-4f30-ab37-9883caae33f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d4b8272a98d4acd09f165939684c0990b9e56b1269f85c20d2bd41c0f75f67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ddfff8cca35bda4796fe58ace60506f5c87e4398193d6711e81aae0d4b7d729\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae336ca136d0a51627e63b1ec86bd4f916a5efaf804e5b91014188331c19802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6965798de9aaf2b8066b7bd8abc23e12545f026bf1f0930dc0085b992553f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc6965798de9aaf2b8066b7bd8abc23e12545f026bf1f0930dc0085b992553f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:25Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.542045 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rnzq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0075197-9392-4109-b8a7-39507b15cc17\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53a00b17138b236dc8517cbd4ee71ea3ed7e8885247d06eb2e8840b8b145783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4mgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rnzq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:25Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.551797 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9k8zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4f75fd3-03e8-41e7-8926-f6ff21a5b681\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpdh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpdh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9k8zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:25Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.562958 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f28d0a3aec8a36273fa4480c552d80a5ee23210203861811e37f27bccda1147e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40791d93876595c4093814146d269b8fe70efb64ef2888e03c25aebef0e9295e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:25Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.577364 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a48de155-7389-48db-8f86-d175b7db70c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://701274e49f68efbc0a939bc08771288c50662cef9c59da73170f7f114e222337\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9t7pq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:25Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.588916 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nndh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a80ba5e1-90cc-44ce-be19-fdf0c007d5b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3684c4e5e839ce117a1772571ae080042e0187ea1aaca635edca07ac10d94c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2c22k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31d74665ae370eed09b8a3195bd3d122ad64c6e33bf55fc42cd40a862671b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2c22k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nndh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:25Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.600523 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cb4vc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed42d686-41af-470e-910d-a1fabbec66b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2edf7fa65761bc560178a40857780a5668e183a4b01974935cdbc3540915114f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73f629b20d9f61227553520b30d0ca046750433926a8d31501e5fb77467426cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:52:22Z\\\",\\\"message\\\":\\\"2025-12-06T05:51:36+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_4426309b-ddbc-4734-9136-9da7465cfbc2\\\\n2025-12-06T05:51:36+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_4426309b-ddbc-4734-9136-9da7465cfbc2 to /host/opt/cni/bin/\\\\n2025-12-06T05:51:36Z [verbose] multus-daemon started\\\\n2025-12-06T05:51:36Z [verbose] Readiness Indicator file check\\\\n2025-12-06T05:52:21Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2nwpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cb4vc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:25Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.609418 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hmlmc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c7558c-89b1-432d-893c-ccb2b3aba183\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c2f8876171399b61abd26285afba0547dac3c24c957bd6d202721b04c912aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm7wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hmlmc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:25Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.624286 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.624310 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.624317 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.624330 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.624339 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:25Z","lastTransitionTime":"2025-12-06T05:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.725988 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.726022 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.726031 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.726045 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.726054 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:25Z","lastTransitionTime":"2025-12-06T05:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.828333 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.828392 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.828401 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.828413 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.828422 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:25Z","lastTransitionTime":"2025-12-06T05:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.931076 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.931143 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.931164 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.931192 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:25 crc kubenswrapper[4809]: I1206 05:52:25.931213 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:25Z","lastTransitionTime":"2025-12-06T05:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.034272 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.034311 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.034322 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.034336 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.034345 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:26Z","lastTransitionTime":"2025-12-06T05:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.136589 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.136630 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.136646 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.136665 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.136680 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:26Z","lastTransitionTime":"2025-12-06T05:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.239005 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.239036 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.239046 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.239059 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.239069 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:26Z","lastTransitionTime":"2025-12-06T05:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.341354 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.341425 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.341440 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.341455 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.341467 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:26Z","lastTransitionTime":"2025-12-06T05:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.388048 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:52:26 crc kubenswrapper[4809]: E1206 05:52:26.388155 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.388208 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:52:26 crc kubenswrapper[4809]: E1206 05:52:26.388266 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.443836 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.444008 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.444035 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.444055 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.444071 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:26Z","lastTransitionTime":"2025-12-06T05:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.546026 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.546093 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.546114 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.546148 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.546182 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:26Z","lastTransitionTime":"2025-12-06T05:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.648259 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.648309 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.648325 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.648346 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.648362 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:26Z","lastTransitionTime":"2025-12-06T05:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.751474 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.751558 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.751576 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.751600 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.751619 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:26Z","lastTransitionTime":"2025-12-06T05:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.859232 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.859273 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.859283 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.859297 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.859307 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:26Z","lastTransitionTime":"2025-12-06T05:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.961978 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.962017 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.962027 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.962042 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:26 crc kubenswrapper[4809]: I1206 05:52:26.962053 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:26Z","lastTransitionTime":"2025-12-06T05:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.066436 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.066654 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.066735 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.067143 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.067559 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:27Z","lastTransitionTime":"2025-12-06T05:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.171192 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.171240 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.171250 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.171264 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.171273 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:27Z","lastTransitionTime":"2025-12-06T05:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.273815 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.273848 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.273856 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.273868 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.273877 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:27Z","lastTransitionTime":"2025-12-06T05:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.375694 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.375727 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.375736 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.375748 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.375756 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:27Z","lastTransitionTime":"2025-12-06T05:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.388100 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:52:27 crc kubenswrapper[4809]: E1206 05:52:27.388200 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.388250 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:52:27 crc kubenswrapper[4809]: E1206 05:52:27.388403 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.478233 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.478273 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.478304 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.478333 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.478342 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:27Z","lastTransitionTime":"2025-12-06T05:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.580419 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.580495 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.580510 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.580525 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.580534 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:27Z","lastTransitionTime":"2025-12-06T05:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.682673 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.682705 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.682714 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.682724 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.682732 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:27Z","lastTransitionTime":"2025-12-06T05:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.784831 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.784867 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.784877 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.784890 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.784898 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:27Z","lastTransitionTime":"2025-12-06T05:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.887786 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.887864 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.887876 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.887891 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.887904 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:27Z","lastTransitionTime":"2025-12-06T05:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.990541 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.990608 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.990621 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.990640 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:27 crc kubenswrapper[4809]: I1206 05:52:27.990652 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:27Z","lastTransitionTime":"2025-12-06T05:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:28 crc kubenswrapper[4809]: I1206 05:52:28.093165 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:28 crc kubenswrapper[4809]: I1206 05:52:28.093215 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:28 crc kubenswrapper[4809]: I1206 05:52:28.093229 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:28 crc kubenswrapper[4809]: I1206 05:52:28.093246 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:28 crc kubenswrapper[4809]: I1206 05:52:28.093257 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:28Z","lastTransitionTime":"2025-12-06T05:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:28 crc kubenswrapper[4809]: I1206 05:52:28.195612 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:28 crc kubenswrapper[4809]: I1206 05:52:28.195653 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:28 crc kubenswrapper[4809]: I1206 05:52:28.195661 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:28 crc kubenswrapper[4809]: I1206 05:52:28.195677 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:28 crc kubenswrapper[4809]: I1206 05:52:28.195686 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:28Z","lastTransitionTime":"2025-12-06T05:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:28 crc kubenswrapper[4809]: I1206 05:52:28.298274 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:28 crc kubenswrapper[4809]: I1206 05:52:28.298326 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:28 crc kubenswrapper[4809]: I1206 05:52:28.298339 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:28 crc kubenswrapper[4809]: I1206 05:52:28.298463 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:28 crc kubenswrapper[4809]: I1206 05:52:28.298481 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:28Z","lastTransitionTime":"2025-12-06T05:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:28 crc kubenswrapper[4809]: I1206 05:52:28.388595 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:52:28 crc kubenswrapper[4809]: E1206 05:52:28.388727 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:52:28 crc kubenswrapper[4809]: I1206 05:52:28.388618 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:52:28 crc kubenswrapper[4809]: E1206 05:52:28.388910 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:52:28 crc kubenswrapper[4809]: I1206 05:52:28.400951 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:28 crc kubenswrapper[4809]: I1206 05:52:28.401040 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:28 crc kubenswrapper[4809]: I1206 05:52:28.401057 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:28 crc kubenswrapper[4809]: I1206 05:52:28.401078 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:28 crc kubenswrapper[4809]: I1206 05:52:28.401095 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:28Z","lastTransitionTime":"2025-12-06T05:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:28 crc kubenswrapper[4809]: I1206 05:52:28.504270 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:28 crc kubenswrapper[4809]: I1206 05:52:28.504544 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:28 crc kubenswrapper[4809]: I1206 05:52:28.504649 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:28 crc kubenswrapper[4809]: I1206 05:52:28.504763 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:28 crc kubenswrapper[4809]: I1206 05:52:28.504908 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:28Z","lastTransitionTime":"2025-12-06T05:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:28 crc kubenswrapper[4809]: I1206 05:52:28.607127 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:28 crc kubenswrapper[4809]: I1206 05:52:28.607180 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:28 crc kubenswrapper[4809]: I1206 05:52:28.607197 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:28 crc kubenswrapper[4809]: I1206 05:52:28.607220 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:28 crc kubenswrapper[4809]: I1206 05:52:28.607235 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:28Z","lastTransitionTime":"2025-12-06T05:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:28 crc kubenswrapper[4809]: I1206 05:52:28.710171 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:28 crc kubenswrapper[4809]: I1206 05:52:28.710200 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:28 crc kubenswrapper[4809]: I1206 05:52:28.710232 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:28 crc kubenswrapper[4809]: I1206 05:52:28.710249 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:28 crc kubenswrapper[4809]: I1206 05:52:28.710261 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:28Z","lastTransitionTime":"2025-12-06T05:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:28 crc kubenswrapper[4809]: I1206 05:52:28.812658 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:28 crc kubenswrapper[4809]: I1206 05:52:28.812706 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:28 crc kubenswrapper[4809]: I1206 05:52:28.812822 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:28 crc kubenswrapper[4809]: I1206 05:52:28.812848 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:28 crc kubenswrapper[4809]: I1206 05:52:28.812882 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:28Z","lastTransitionTime":"2025-12-06T05:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:28 crc kubenswrapper[4809]: I1206 05:52:28.915558 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:28 crc kubenswrapper[4809]: I1206 05:52:28.915600 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:28 crc kubenswrapper[4809]: I1206 05:52:28.915614 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:28 crc kubenswrapper[4809]: I1206 05:52:28.915635 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:28 crc kubenswrapper[4809]: I1206 05:52:28.915652 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:28Z","lastTransitionTime":"2025-12-06T05:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.018501 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.018538 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.018549 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.018566 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.018578 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:29Z","lastTransitionTime":"2025-12-06T05:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.120824 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.120854 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.120867 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.120880 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.120890 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:29Z","lastTransitionTime":"2025-12-06T05:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.223149 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.223182 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.223194 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.223211 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.223222 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:29Z","lastTransitionTime":"2025-12-06T05:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.326568 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.326883 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.327012 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.327123 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.327222 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:29Z","lastTransitionTime":"2025-12-06T05:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.388449 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:52:29 crc kubenswrapper[4809]: E1206 05:52:29.388601 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.388637 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:52:29 crc kubenswrapper[4809]: E1206 05:52:29.388762 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.431313 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.431371 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.431393 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.431420 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.431440 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:29Z","lastTransitionTime":"2025-12-06T05:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.535743 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.535788 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.535800 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.535817 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.535829 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:29Z","lastTransitionTime":"2025-12-06T05:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.639722 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.639764 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.639776 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.639792 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.639804 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:29Z","lastTransitionTime":"2025-12-06T05:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.742700 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.742732 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.742741 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.742753 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.742762 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:29Z","lastTransitionTime":"2025-12-06T05:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.846317 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.846384 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.846403 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.846428 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.846444 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:29Z","lastTransitionTime":"2025-12-06T05:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.950146 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.950192 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.950202 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.950217 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:29 crc kubenswrapper[4809]: I1206 05:52:29.950228 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:29Z","lastTransitionTime":"2025-12-06T05:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.052792 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.052843 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.052855 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.052874 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.052885 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:30Z","lastTransitionTime":"2025-12-06T05:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.156334 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.156420 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.156445 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.156477 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.156501 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:30Z","lastTransitionTime":"2025-12-06T05:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.259553 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.259629 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.259648 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.259674 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.259692 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:30Z","lastTransitionTime":"2025-12-06T05:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.269212 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.269274 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.269296 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.269325 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.269349 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:30Z","lastTransitionTime":"2025-12-06T05:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:30 crc kubenswrapper[4809]: E1206 05:52:30.292771 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"88a28b6b-9c85-4932-b5bc-36346303aee5\\\",\\\"systemUUID\\\":\\\"5931df91-c1df-4f78-9e5f-391a01a76127\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:30Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.297909 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.298000 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.298040 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.298076 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.298102 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:30Z","lastTransitionTime":"2025-12-06T05:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:30 crc kubenswrapper[4809]: E1206 05:52:30.318349 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"88a28b6b-9c85-4932-b5bc-36346303aee5\\\",\\\"systemUUID\\\":\\\"5931df91-c1df-4f78-9e5f-391a01a76127\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:30Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.323586 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.323651 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.323670 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.323693 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.323712 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:30Z","lastTransitionTime":"2025-12-06T05:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:30 crc kubenswrapper[4809]: E1206 05:52:30.343162 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"88a28b6b-9c85-4932-b5bc-36346303aee5\\\",\\\"systemUUID\\\":\\\"5931df91-c1df-4f78-9e5f-391a01a76127\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:30Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.348454 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.348510 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.348528 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.348553 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.348572 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:30Z","lastTransitionTime":"2025-12-06T05:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:30 crc kubenswrapper[4809]: E1206 05:52:30.361497 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"88a28b6b-9c85-4932-b5bc-36346303aee5\\\",\\\"systemUUID\\\":\\\"5931df91-c1df-4f78-9e5f-391a01a76127\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:30Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.365683 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.365729 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.365749 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.365773 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.365790 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:30Z","lastTransitionTime":"2025-12-06T05:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:30 crc kubenswrapper[4809]: E1206 05:52:30.386336 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"88a28b6b-9c85-4932-b5bc-36346303aee5\\\",\\\"systemUUID\\\":\\\"5931df91-c1df-4f78-9e5f-391a01a76127\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:30Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:30 crc kubenswrapper[4809]: E1206 05:52:30.386489 4809 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.387513 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.387713 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:52:30 crc kubenswrapper[4809]: E1206 05:52:30.387848 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:52:30 crc kubenswrapper[4809]: E1206 05:52:30.388035 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.388748 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.388806 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.388830 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.388857 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.388881 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:30Z","lastTransitionTime":"2025-12-06T05:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.401929 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.491967 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.492010 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.492018 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.492031 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.492041 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:30Z","lastTransitionTime":"2025-12-06T05:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.594898 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.594984 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.595010 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.595045 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.595068 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:30Z","lastTransitionTime":"2025-12-06T05:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.698186 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.698250 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.698270 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.698296 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.698316 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:30Z","lastTransitionTime":"2025-12-06T05:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.801347 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.801420 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.801444 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.801473 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.801494 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:30Z","lastTransitionTime":"2025-12-06T05:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.904941 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.905025 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.905042 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.905062 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:30 crc kubenswrapper[4809]: I1206 05:52:30.905075 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:30Z","lastTransitionTime":"2025-12-06T05:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.007706 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.007749 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.007764 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.007784 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.007799 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:31Z","lastTransitionTime":"2025-12-06T05:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.110244 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.110325 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.110351 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.110379 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.110399 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:31Z","lastTransitionTime":"2025-12-06T05:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.212917 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.213100 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.213130 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.213162 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.213183 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:31Z","lastTransitionTime":"2025-12-06T05:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.316480 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.316821 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.317113 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.317142 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.317166 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:31Z","lastTransitionTime":"2025-12-06T05:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.387909 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.388158 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:52:31 crc kubenswrapper[4809]: E1206 05:52:31.388306 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:52:31 crc kubenswrapper[4809]: E1206 05:52:31.388158 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.421663 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.421705 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.421714 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.421728 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.421738 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:31Z","lastTransitionTime":"2025-12-06T05:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.524791 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.524849 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.524868 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.524892 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.524908 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:31Z","lastTransitionTime":"2025-12-06T05:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.627733 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.627812 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.627824 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.627849 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.627863 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:31Z","lastTransitionTime":"2025-12-06T05:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.730228 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.730269 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.730278 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.730292 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.730301 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:31Z","lastTransitionTime":"2025-12-06T05:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.833512 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.833552 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.833560 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.833575 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.833586 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:31Z","lastTransitionTime":"2025-12-06T05:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.936072 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.936113 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.936128 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.936143 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:31 crc kubenswrapper[4809]: I1206 05:52:31.936152 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:31Z","lastTransitionTime":"2025-12-06T05:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.039658 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.039706 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.039721 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.039738 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.039751 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:32Z","lastTransitionTime":"2025-12-06T05:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.141824 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.141875 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.141888 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.141906 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.141916 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:32Z","lastTransitionTime":"2025-12-06T05:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.244802 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.244911 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.244958 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.244985 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.245005 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:32Z","lastTransitionTime":"2025-12-06T05:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.348353 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.348394 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.348407 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.348423 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.348440 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:32Z","lastTransitionTime":"2025-12-06T05:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.388080 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:52:32 crc kubenswrapper[4809]: E1206 05:52:32.388201 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.388085 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:52:32 crc kubenswrapper[4809]: E1206 05:52:32.388291 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.451154 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.451232 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.451254 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.451282 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.451303 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:32Z","lastTransitionTime":"2025-12-06T05:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.554073 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.554159 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.554183 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.554249 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.554272 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:32Z","lastTransitionTime":"2025-12-06T05:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.657300 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.657344 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.657354 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.657370 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.657383 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:32Z","lastTransitionTime":"2025-12-06T05:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.760336 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.760394 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.760410 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.760433 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.760450 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:32Z","lastTransitionTime":"2025-12-06T05:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.862841 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.862867 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.862875 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.862888 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.862897 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:32Z","lastTransitionTime":"2025-12-06T05:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.966305 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.966351 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.966366 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.966382 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:32 crc kubenswrapper[4809]: I1206 05:52:32.966393 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:32Z","lastTransitionTime":"2025-12-06T05:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.069029 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.069079 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.069094 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.069112 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.069124 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:33Z","lastTransitionTime":"2025-12-06T05:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.172193 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.172377 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.172398 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.172426 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.172446 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:33Z","lastTransitionTime":"2025-12-06T05:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.277066 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.277112 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.277123 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.277141 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.277153 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:33Z","lastTransitionTime":"2025-12-06T05:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.379977 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.380030 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.380042 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.380060 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.380072 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:33Z","lastTransitionTime":"2025-12-06T05:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.388579 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.389013 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:52:33 crc kubenswrapper[4809]: E1206 05:52:33.389135 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:52:33 crc kubenswrapper[4809]: E1206 05:52:33.389307 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.389423 4809 scope.go:117] "RemoveContainer" containerID="80ffc0fe836258c86321ba5a7f8e1c7e289c0f85e0b48bc138f69cf96cf225ab" Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.482152 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.482190 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.482203 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.482219 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.482231 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:33Z","lastTransitionTime":"2025-12-06T05:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.583655 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.583703 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.583723 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.583754 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.583770 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:33Z","lastTransitionTime":"2025-12-06T05:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.687008 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.687061 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.687076 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.687097 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.687112 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:33Z","lastTransitionTime":"2025-12-06T05:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.790226 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.790303 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.790327 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.790356 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.790373 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:33Z","lastTransitionTime":"2025-12-06T05:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.894411 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.894477 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.894496 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.894523 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.894546 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:33Z","lastTransitionTime":"2025-12-06T05:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.997728 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.997786 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.997803 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.997825 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:33 crc kubenswrapper[4809]: I1206 05:52:33.997842 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:33Z","lastTransitionTime":"2025-12-06T05:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:34 crc kubenswrapper[4809]: I1206 05:52:34.100375 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:34 crc kubenswrapper[4809]: I1206 05:52:34.100421 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:34 crc kubenswrapper[4809]: I1206 05:52:34.100432 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:34 crc kubenswrapper[4809]: I1206 05:52:34.100448 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:34 crc kubenswrapper[4809]: I1206 05:52:34.100459 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:34Z","lastTransitionTime":"2025-12-06T05:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:34 crc kubenswrapper[4809]: I1206 05:52:34.203523 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:34 crc kubenswrapper[4809]: I1206 05:52:34.203582 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:34 crc kubenswrapper[4809]: I1206 05:52:34.203599 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:34 crc kubenswrapper[4809]: I1206 05:52:34.203622 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:34 crc kubenswrapper[4809]: I1206 05:52:34.203640 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:34Z","lastTransitionTime":"2025-12-06T05:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:34 crc kubenswrapper[4809]: I1206 05:52:34.306019 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:34 crc kubenswrapper[4809]: I1206 05:52:34.306073 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:34 crc kubenswrapper[4809]: I1206 05:52:34.306092 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:34 crc kubenswrapper[4809]: I1206 05:52:34.306121 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:34 crc kubenswrapper[4809]: I1206 05:52:34.306143 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:34Z","lastTransitionTime":"2025-12-06T05:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:34 crc kubenswrapper[4809]: I1206 05:52:34.387575 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:52:34 crc kubenswrapper[4809]: E1206 05:52:34.387724 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:52:34 crc kubenswrapper[4809]: I1206 05:52:34.387917 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:52:34 crc kubenswrapper[4809]: E1206 05:52:34.388017 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:52:34 crc kubenswrapper[4809]: I1206 05:52:34.408526 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:34 crc kubenswrapper[4809]: I1206 05:52:34.408562 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:34 crc kubenswrapper[4809]: I1206 05:52:34.408573 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:34 crc kubenswrapper[4809]: I1206 05:52:34.408587 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:34 crc kubenswrapper[4809]: I1206 05:52:34.408599 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:34Z","lastTransitionTime":"2025-12-06T05:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:34 crc kubenswrapper[4809]: I1206 05:52:34.510642 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:34 crc kubenswrapper[4809]: I1206 05:52:34.510703 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:34 crc kubenswrapper[4809]: I1206 05:52:34.510721 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:34 crc kubenswrapper[4809]: I1206 05:52:34.510746 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:34 crc kubenswrapper[4809]: I1206 05:52:34.510763 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:34Z","lastTransitionTime":"2025-12-06T05:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:34 crc kubenswrapper[4809]: I1206 05:52:34.613795 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:34 crc kubenswrapper[4809]: I1206 05:52:34.613855 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:34 crc kubenswrapper[4809]: I1206 05:52:34.613872 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:34 crc kubenswrapper[4809]: I1206 05:52:34.613897 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:34 crc kubenswrapper[4809]: I1206 05:52:34.613924 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:34Z","lastTransitionTime":"2025-12-06T05:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:34 crc kubenswrapper[4809]: I1206 05:52:34.716852 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:34 crc kubenswrapper[4809]: I1206 05:52:34.716899 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:34 crc kubenswrapper[4809]: I1206 05:52:34.716914 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:34 crc kubenswrapper[4809]: I1206 05:52:34.716934 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:34 crc kubenswrapper[4809]: I1206 05:52:34.716946 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:34Z","lastTransitionTime":"2025-12-06T05:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:34 crc kubenswrapper[4809]: I1206 05:52:34.819737 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:34 crc kubenswrapper[4809]: I1206 05:52:34.819780 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:34 crc kubenswrapper[4809]: I1206 05:52:34.819790 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:34 crc kubenswrapper[4809]: I1206 05:52:34.819822 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:34 crc kubenswrapper[4809]: I1206 05:52:34.819832 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:34Z","lastTransitionTime":"2025-12-06T05:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:34 crc kubenswrapper[4809]: I1206 05:52:34.922633 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:34 crc kubenswrapper[4809]: I1206 05:52:34.922694 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:34 crc kubenswrapper[4809]: I1206 05:52:34.922708 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:34 crc kubenswrapper[4809]: I1206 05:52:34.922727 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:34 crc kubenswrapper[4809]: I1206 05:52:34.922740 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:34Z","lastTransitionTime":"2025-12-06T05:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.025488 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.025522 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.025533 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.025550 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.025562 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:35Z","lastTransitionTime":"2025-12-06T05:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.128254 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.128295 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.128305 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.128320 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.128330 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:35Z","lastTransitionTime":"2025-12-06T05:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.231756 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.231836 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.231849 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.231868 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.231883 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:35Z","lastTransitionTime":"2025-12-06T05:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.335095 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.335160 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.335177 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.335198 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.335221 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:35Z","lastTransitionTime":"2025-12-06T05:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.390714 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:52:35 crc kubenswrapper[4809]: E1206 05:52:35.390902 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.391113 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:52:35 crc kubenswrapper[4809]: E1206 05:52:35.391237 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.414261 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd45866c51410e910e87d43870bf531b0d05541e9164e700150fd37987e284af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:35Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.430074 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:35Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.440166 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.440205 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.440216 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.440231 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.440242 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:35Z","lastTransitionTime":"2025-12-06T05:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.442401 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:35Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.456581 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:35Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.468820 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947d89be9d04856af8d09facb8d4583f91898bd03a36d3ce8b9b74bc3bb795e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://930788ec9b4d97fdc39353a93111b84707037ce445f1930aa7737a20a44e1d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-npms2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:35Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.481550 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5636dd78-3d61-4c6a-b8fd-6d2457e19234\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:35Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.491179 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"353c3bc7-39c3-4f30-ab37-9883caae33f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d4b8272a98d4acd09f165939684c0990b9e56b1269f85c20d2bd41c0f75f67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ddfff8cca35bda4796fe58ace60506f5c87e4398193d6711e81aae0d4b7d729\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae336ca136d0a51627e63b1ec86bd4f916a5efaf804e5b91014188331c19802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6965798de9aaf2b8066b7bd8abc23e12545f026bf1f0930dc0085b992553f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc6965798de9aaf2b8066b7bd8abc23e12545f026bf1f0930dc0085b992553f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:35Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.500190 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7b8a286-4824-49e9-bcd3-77ed87fd7638\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95eafea368aebd35fb8e9a20bb01938c0ee9b1109d1a4ac5c1062e0d4ce7465c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a37233831095a0ba4b488c136d3ede54ee833ab0d0141448f20ddb34fd4cc8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a37233831095a0ba4b488c136d3ede54ee833ab0d0141448f20ddb34fd4cc8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:35Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.509321 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rnzq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0075197-9392-4109-b8a7-39507b15cc17\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53a00b17138b236dc8517cbd4ee71ea3ed7e8885247d06eb2e8840b8b145783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4mgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rnzq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:35Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.518743 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9k8zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4f75fd3-03e8-41e7-8926-f6ff21a5b681\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpdh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpdh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9k8zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:35Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.529353 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f28d0a3aec8a36273fa4480c552d80a5ee23210203861811e37f27bccda1147e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40791d93876595c4093814146d269b8fe70efb64ef2888e03c25aebef0e9295e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:35Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.541020 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a48de155-7389-48db-8f86-d175b7db70c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://701274e49f68efbc0a939bc08771288c50662cef9c59da73170f7f114e222337\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9t7pq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:35Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.542744 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.542769 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.542780 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.542842 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.542855 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:35Z","lastTransitionTime":"2025-12-06T05:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.553026 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nndh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a80ba5e1-90cc-44ce-be19-fdf0c007d5b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3684c4e5e839ce117a1772571ae080042e0187ea1aaca635edca07ac10d94c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2c22k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31d74665ae370eed09b8a3195bd3d122ad64c6e33bf55fc42cd40a862671b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2c22k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nndh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:35Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.566414 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cb4vc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed42d686-41af-470e-910d-a1fabbec66b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2edf7fa65761bc560178a40857780a5668e183a4b01974935cdbc3540915114f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73f629b20d9f61227553520b30d0ca046750433926a8d31501e5fb77467426cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:52:22Z\\\",\\\"message\\\":\\\"2025-12-06T05:51:36+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_4426309b-ddbc-4734-9136-9da7465cfbc2\\\\n2025-12-06T05:51:36+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_4426309b-ddbc-4734-9136-9da7465cfbc2 to /host/opt/cni/bin/\\\\n2025-12-06T05:51:36Z [verbose] multus-daemon started\\\\n2025-12-06T05:51:36Z [verbose] Readiness Indicator file check\\\\n2025-12-06T05:52:21Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2nwpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cb4vc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:35Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.578151 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hmlmc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c7558c-89b1-432d-893c-ccb2b3aba183\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c2f8876171399b61abd26285afba0547dac3c24c957bd6d202721b04c912aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm7wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hmlmc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:35Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.602437 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c981561d51870deffeb446e8ca769b7ac0a1a16edbe4e982e0f2c84a45043db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d58d489f532b78aca59f5b92584aad1dc292e6700c112c531cbeb3209fd90c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dcd421b38e5f9ba0c542a09f0047bde6c043942a096d11575afb7423ab2552c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab9b3bd4c8a6d6af7939b06fe01b8a0a50d81bb8c669bc6a27f486953c234e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a3e4453935871bb84d68bc4a8b6c98f299b7c8094ba82291bc90e258e1fec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5be86696b7a6518c5df930accb2c6890d58cfc37104fd34a359724b9e30231b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ffc0fe836258c86321ba5a7f8e1c7e289c0f85e0b48bc138f69cf96cf225ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80ffc0fe836258c86321ba5a7f8e1c7e289c0f85e0b48bc138f69cf96cf225ab\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:52:07Z\\\",\\\"message\\\":\\\"52:06.762839 6453 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.59 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {dce28c51-c9f1-478b-97c8-7e209d6e7cbe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 05:52:06.762813 6453 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-dns-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"d937b3b3-82c3-4791-9a66-41b9fed53e9d\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-dns-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, A\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:52:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5xvvb_openshift-ovn-kubernetes(ffb3a4f1-1e22-4220-ac80-3b2d69d2db99)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dee8b965be3b2486f1f532f2bdfbbcbd84f9b3b711a11c25e86dae1284314d4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5xvvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:35Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.616481 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa4dd3f8-c119-46de-ad4e-dd29bb11a32c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa59e4ec9867cf280b308ee3adb5e303eaf94455aca588f5eb74a2b6263d57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5720f7f2977d9af6ed5324a30d08d34a53d55460951b3875b7d7d753d202f349\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dfd977b049a8ff81192d4041adb265366ab88475ee84c5750305d061d21c8db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d5801b61a87c95c9a623c3106e305920199c640efa8a32cf57698245697e54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:35Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.628350 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://010c2681c08dc21bc3a6fb82c04ab3dc2c26dcb76ec12b143c22845cac04ed25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:35Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.645425 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.645468 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.645478 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.645492 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.645502 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:35Z","lastTransitionTime":"2025-12-06T05:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.748449 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.748492 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.748504 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.748521 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.748533 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:35Z","lastTransitionTime":"2025-12-06T05:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.851176 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.851514 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.851527 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.851543 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.851555 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:35Z","lastTransitionTime":"2025-12-06T05:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.863759 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5xvvb_ffb3a4f1-1e22-4220-ac80-3b2d69d2db99/ovnkube-controller/2.log" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.866774 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" event={"ID":"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99","Type":"ContainerStarted","Data":"4a4a7efb6435b711f6fc3d07f0f82258ae97be843a5917fe8017c2afd690c200"} Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.867460 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.882066 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f28d0a3aec8a36273fa4480c552d80a5ee23210203861811e37f27bccda1147e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40791d93876595c4093814146d269b8fe70efb64ef2888e03c25aebef0e9295e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:35Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.897526 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a48de155-7389-48db-8f86-d175b7db70c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://701274e49f68efbc0a939bc08771288c50662cef9c59da73170f7f114e222337\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9t7pq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:35Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.908936 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nndh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a80ba5e1-90cc-44ce-be19-fdf0c007d5b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3684c4e5e839ce117a1772571ae080042e0187ea1aaca635edca07ac10d94c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2c22k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31d74665ae370eed09b8a3195bd3d122ad64c6e33bf55fc42cd40a862671b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2c22k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nndh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:35Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.923532 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cb4vc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed42d686-41af-470e-910d-a1fabbec66b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2edf7fa65761bc560178a40857780a5668e183a4b01974935cdbc3540915114f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73f629b20d9f61227553520b30d0ca046750433926a8d31501e5fb77467426cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:52:22Z\\\",\\\"message\\\":\\\"2025-12-06T05:51:36+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_4426309b-ddbc-4734-9136-9da7465cfbc2\\\\n2025-12-06T05:51:36+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_4426309b-ddbc-4734-9136-9da7465cfbc2 to /host/opt/cni/bin/\\\\n2025-12-06T05:51:36Z [verbose] multus-daemon started\\\\n2025-12-06T05:51:36Z [verbose] Readiness Indicator file check\\\\n2025-12-06T05:52:21Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2nwpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cb4vc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:35Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.947097 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hmlmc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c7558c-89b1-432d-893c-ccb2b3aba183\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c2f8876171399b61abd26285afba0547dac3c24c957bd6d202721b04c912aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm7wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hmlmc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:35Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.954392 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.954437 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.954448 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.954463 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.954475 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:35Z","lastTransitionTime":"2025-12-06T05:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.975200 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa4dd3f8-c119-46de-ad4e-dd29bb11a32c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa59e4ec9867cf280b308ee3adb5e303eaf94455aca588f5eb74a2b6263d57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5720f7f2977d9af6ed5324a30d08d34a53d55460951b3875b7d7d753d202f349\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dfd977b049a8ff81192d4041adb265366ab88475ee84c5750305d061d21c8db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d5801b61a87c95c9a623c3106e305920199c640efa8a32cf57698245697e54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:35Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:35 crc kubenswrapper[4809]: I1206 05:52:35.991878 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://010c2681c08dc21bc3a6fb82c04ab3dc2c26dcb76ec12b143c22845cac04ed25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:35Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.024212 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c981561d51870deffeb446e8ca769b7ac0a1a16edbe4e982e0f2c84a45043db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d58d489f532b78aca59f5b92584aad1dc292e6700c112c531cbeb3209fd90c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dcd421b38e5f9ba0c542a09f0047bde6c043942a096d11575afb7423ab2552c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab9b3bd4c8a6d6af7939b06fe01b8a0a50d81bb8c669bc6a27f486953c234e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a3e4453935871bb84d68bc4a8b6c98f299b7c8094ba82291bc90e258e1fec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5be86696b7a6518c5df930accb2c6890d58cfc37104fd34a359724b9e30231b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a4a7efb6435b711f6fc3d07f0f82258ae97be843a5917fe8017c2afd690c200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80ffc0fe836258c86321ba5a7f8e1c7e289c0f85e0b48bc138f69cf96cf225ab\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:52:07Z\\\",\\\"message\\\":\\\"52:06.762839 6453 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.59 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {dce28c51-c9f1-478b-97c8-7e209d6e7cbe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 05:52:06.762813 6453 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-dns-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"d937b3b3-82c3-4791-9a66-41b9fed53e9d\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-dns-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, A\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:52:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dee8b965be3b2486f1f532f2bdfbbcbd84f9b3b711a11c25e86dae1284314d4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5xvvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:36Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.039136 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:36Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.050615 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947d89be9d04856af8d09facb8d4583f91898bd03a36d3ce8b9b74bc3bb795e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://930788ec9b4d97fdc39353a93111b84707037ce445f1930aa7737a20a44e1d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-npms2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:36Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.056924 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.056968 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.056976 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.057016 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.057026 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:36Z","lastTransitionTime":"2025-12-06T05:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.068467 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5636dd78-3d61-4c6a-b8fd-6d2457e19234\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:36Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.079451 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"353c3bc7-39c3-4f30-ab37-9883caae33f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d4b8272a98d4acd09f165939684c0990b9e56b1269f85c20d2bd41c0f75f67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ddfff8cca35bda4796fe58ace60506f5c87e4398193d6711e81aae0d4b7d729\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae336ca136d0a51627e63b1ec86bd4f916a5efaf804e5b91014188331c19802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6965798de9aaf2b8066b7bd8abc23e12545f026bf1f0930dc0085b992553f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc6965798de9aaf2b8066b7bd8abc23e12545f026bf1f0930dc0085b992553f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:36Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.091378 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7b8a286-4824-49e9-bcd3-77ed87fd7638\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95eafea368aebd35fb8e9a20bb01938c0ee9b1109d1a4ac5c1062e0d4ce7465c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a37233831095a0ba4b488c136d3ede54ee833ab0d0141448f20ddb34fd4cc8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a37233831095a0ba4b488c136d3ede54ee833ab0d0141448f20ddb34fd4cc8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:36Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.105990 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd45866c51410e910e87d43870bf531b0d05541e9164e700150fd37987e284af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:36Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.120739 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:36Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.134306 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:36Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.148592 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rnzq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0075197-9392-4109-b8a7-39507b15cc17\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53a00b17138b236dc8517cbd4ee71ea3ed7e8885247d06eb2e8840b8b145783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4mgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rnzq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:36Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.159639 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.159679 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.159692 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.159707 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.159718 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:36Z","lastTransitionTime":"2025-12-06T05:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.164067 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9k8zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4f75fd3-03e8-41e7-8926-f6ff21a5b681\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpdh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpdh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9k8zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:36Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.261989 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.262033 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.262043 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.262062 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.262074 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:36Z","lastTransitionTime":"2025-12-06T05:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.364131 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.364162 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.364171 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.364186 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.364195 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:36Z","lastTransitionTime":"2025-12-06T05:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.387830 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:52:36 crc kubenswrapper[4809]: E1206 05:52:36.387965 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.387834 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:52:36 crc kubenswrapper[4809]: E1206 05:52:36.388193 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.466347 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.466388 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.466398 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.466415 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.466427 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:36Z","lastTransitionTime":"2025-12-06T05:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.569441 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.569512 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.569535 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.569567 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.569590 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:36Z","lastTransitionTime":"2025-12-06T05:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.673040 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.673105 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.673120 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.673138 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.673149 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:36Z","lastTransitionTime":"2025-12-06T05:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.775838 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.775885 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.775896 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.775912 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.775925 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:36Z","lastTransitionTime":"2025-12-06T05:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.878668 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.878709 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.878721 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.878736 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.878747 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:36Z","lastTransitionTime":"2025-12-06T05:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.982227 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.982273 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.982285 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.982302 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:36 crc kubenswrapper[4809]: I1206 05:52:36.982316 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:36Z","lastTransitionTime":"2025-12-06T05:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.085110 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.085155 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.085166 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.085183 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.085195 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:37Z","lastTransitionTime":"2025-12-06T05:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.188108 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.188183 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.188202 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.188257 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.188282 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:37Z","lastTransitionTime":"2025-12-06T05:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.291075 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.291140 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.291161 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.291189 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.291210 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:37Z","lastTransitionTime":"2025-12-06T05:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.388517 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.388531 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:52:37 crc kubenswrapper[4809]: E1206 05:52:37.388874 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:52:37 crc kubenswrapper[4809]: E1206 05:52:37.389066 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.392749 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.392790 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.392803 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.392816 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.392828 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:37Z","lastTransitionTime":"2025-12-06T05:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.495769 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.495847 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.495870 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.495903 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.495969 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:37Z","lastTransitionTime":"2025-12-06T05:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.599209 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.599277 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.599295 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.599322 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.599347 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:37Z","lastTransitionTime":"2025-12-06T05:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.703378 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.703434 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.703451 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.703475 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.703492 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:37Z","lastTransitionTime":"2025-12-06T05:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.806077 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.806132 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.806148 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.806170 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.806189 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:37Z","lastTransitionTime":"2025-12-06T05:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.875713 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5xvvb_ffb3a4f1-1e22-4220-ac80-3b2d69d2db99/ovnkube-controller/3.log" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.876906 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5xvvb_ffb3a4f1-1e22-4220-ac80-3b2d69d2db99/ovnkube-controller/2.log" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.881666 4809 generic.go:334] "Generic (PLEG): container finished" podID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerID="4a4a7efb6435b711f6fc3d07f0f82258ae97be843a5917fe8017c2afd690c200" exitCode=1 Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.881726 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" event={"ID":"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99","Type":"ContainerDied","Data":"4a4a7efb6435b711f6fc3d07f0f82258ae97be843a5917fe8017c2afd690c200"} Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.881773 4809 scope.go:117] "RemoveContainer" containerID="80ffc0fe836258c86321ba5a7f8e1c7e289c0f85e0b48bc138f69cf96cf225ab" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.883296 4809 scope.go:117] "RemoveContainer" containerID="4a4a7efb6435b711f6fc3d07f0f82258ae97be843a5917fe8017c2afd690c200" Dec 06 05:52:37 crc kubenswrapper[4809]: E1206 05:52:37.883634 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5xvvb_openshift-ovn-kubernetes(ffb3a4f1-1e22-4220-ac80-3b2d69d2db99)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.903703 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa4dd3f8-c119-46de-ad4e-dd29bb11a32c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa59e4ec9867cf280b308ee3adb5e303eaf94455aca588f5eb74a2b6263d57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5720f7f2977d9af6ed5324a30d08d34a53d55460951b3875b7d7d753d202f349\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dfd977b049a8ff81192d4041adb265366ab88475ee84c5750305d061d21c8db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d5801b61a87c95c9a623c3106e305920199c640efa8a32cf57698245697e54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:37Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.908471 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.908514 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.908525 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.908540 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.908552 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:37Z","lastTransitionTime":"2025-12-06T05:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.918405 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://010c2681c08dc21bc3a6fb82c04ab3dc2c26dcb76ec12b143c22845cac04ed25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:37Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.938942 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c981561d51870deffeb446e8ca769b7ac0a1a16edbe4e982e0f2c84a45043db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d58d489f532b78aca59f5b92584aad1dc292e6700c112c531cbeb3209fd90c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dcd421b38e5f9ba0c542a09f0047bde6c043942a096d11575afb7423ab2552c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab9b3bd4c8a6d6af7939b06fe01b8a0a50d81bb8c669bc6a27f486953c234e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a3e4453935871bb84d68bc4a8b6c98f299b7c8094ba82291bc90e258e1fec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5be86696b7a6518c5df930accb2c6890d58cfc37104fd34a359724b9e30231b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a4a7efb6435b711f6fc3d07f0f82258ae97be843a5917fe8017c2afd690c200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80ffc0fe836258c86321ba5a7f8e1c7e289c0f85e0b48bc138f69cf96cf225ab\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:52:07Z\\\",\\\"message\\\":\\\"52:06.762839 6453 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.59 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {dce28c51-c9f1-478b-97c8-7e209d6e7cbe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 05:52:06.762813 6453 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-dns-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"d937b3b3-82c3-4791-9a66-41b9fed53e9d\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-dns-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, A\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:52:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a4a7efb6435b711f6fc3d07f0f82258ae97be843a5917fe8017c2afd690c200\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:52:37Z\\\",\\\"message\\\":\\\"d already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:36Z is after 2025-08-24T17:21:41Z]\\\\nI1206 05:52:36.258911 6813 services_controller.go:451] Built service openshift-network-console/networking-console-plugin cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-network-console/networking-console-plugin_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.246\\\\\\\", Port:9443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1206 05:52:36.258967 6813 services_controller.go:452] Built service openshift-network-\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dee8b965be3b2486f1f532f2bdfbbcbd84f9b3b711a11c25e86dae1284314d4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5xvvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:37Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.953484 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5636dd78-3d61-4c6a-b8fd-6d2457e19234\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:37Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.965617 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"353c3bc7-39c3-4f30-ab37-9883caae33f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d4b8272a98d4acd09f165939684c0990b9e56b1269f85c20d2bd41c0f75f67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ddfff8cca35bda4796fe58ace60506f5c87e4398193d6711e81aae0d4b7d729\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae336ca136d0a51627e63b1ec86bd4f916a5efaf804e5b91014188331c19802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6965798de9aaf2b8066b7bd8abc23e12545f026bf1f0930dc0085b992553f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc6965798de9aaf2b8066b7bd8abc23e12545f026bf1f0930dc0085b992553f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:37Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.974965 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7b8a286-4824-49e9-bcd3-77ed87fd7638\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95eafea368aebd35fb8e9a20bb01938c0ee9b1109d1a4ac5c1062e0d4ce7465c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a37233831095a0ba4b488c136d3ede54ee833ab0d0141448f20ddb34fd4cc8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a37233831095a0ba4b488c136d3ede54ee833ab0d0141448f20ddb34fd4cc8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:37Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.987636 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd45866c51410e910e87d43870bf531b0d05541e9164e700150fd37987e284af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:37Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:37 crc kubenswrapper[4809]: I1206 05:52:37.999284 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:37Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.011143 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.011195 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.011207 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.011225 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.011236 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:38Z","lastTransitionTime":"2025-12-06T05:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.013238 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:38Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.023192 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:38Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.033134 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947d89be9d04856af8d09facb8d4583f91898bd03a36d3ce8b9b74bc3bb795e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://930788ec9b4d97fdc39353a93111b84707037ce445f1930aa7737a20a44e1d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-npms2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:38Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.042316 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rnzq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0075197-9392-4109-b8a7-39507b15cc17\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53a00b17138b236dc8517cbd4ee71ea3ed7e8885247d06eb2e8840b8b145783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4mgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rnzq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:38Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.051135 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9k8zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4f75fd3-03e8-41e7-8926-f6ff21a5b681\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpdh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpdh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9k8zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:38Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.062983 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f28d0a3aec8a36273fa4480c552d80a5ee23210203861811e37f27bccda1147e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40791d93876595c4093814146d269b8fe70efb64ef2888e03c25aebef0e9295e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:38Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.075790 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a48de155-7389-48db-8f86-d175b7db70c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://701274e49f68efbc0a939bc08771288c50662cef9c59da73170f7f114e222337\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9t7pq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:38Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.108365 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nndh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a80ba5e1-90cc-44ce-be19-fdf0c007d5b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3684c4e5e839ce117a1772571ae080042e0187ea1aaca635edca07ac10d94c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2c22k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31d74665ae370eed09b8a3195bd3d122ad64c6e33bf55fc42cd40a862671b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2c22k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nndh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:38Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.112781 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.112809 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.112821 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.112835 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.112845 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:38Z","lastTransitionTime":"2025-12-06T05:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.120735 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cb4vc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed42d686-41af-470e-910d-a1fabbec66b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2edf7fa65761bc560178a40857780a5668e183a4b01974935cdbc3540915114f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73f629b20d9f61227553520b30d0ca046750433926a8d31501e5fb77467426cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:52:22Z\\\",\\\"message\\\":\\\"2025-12-06T05:51:36+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_4426309b-ddbc-4734-9136-9da7465cfbc2\\\\n2025-12-06T05:51:36+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_4426309b-ddbc-4734-9136-9da7465cfbc2 to /host/opt/cni/bin/\\\\n2025-12-06T05:51:36Z [verbose] multus-daemon started\\\\n2025-12-06T05:51:36Z [verbose] Readiness Indicator file check\\\\n2025-12-06T05:52:21Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2nwpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cb4vc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:38Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.133420 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hmlmc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c7558c-89b1-432d-893c-ccb2b3aba183\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c2f8876171399b61abd26285afba0547dac3c24c957bd6d202721b04c912aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm7wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hmlmc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:38Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.215275 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.215307 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.215318 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.215334 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.215344 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:38Z","lastTransitionTime":"2025-12-06T05:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.318090 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.318116 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.318125 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.318136 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.318145 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:38Z","lastTransitionTime":"2025-12-06T05:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.388332 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.388423 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:52:38 crc kubenswrapper[4809]: E1206 05:52:38.388551 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:52:38 crc kubenswrapper[4809]: E1206 05:52:38.389018 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.420763 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.420811 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.420825 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.420840 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.420852 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:38Z","lastTransitionTime":"2025-12-06T05:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.523778 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.523815 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.523825 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.523842 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.523853 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:38Z","lastTransitionTime":"2025-12-06T05:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.626477 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.626859 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.627052 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.627261 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.627429 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:38Z","lastTransitionTime":"2025-12-06T05:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.731129 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.731344 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.731415 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.731553 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.731634 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:38Z","lastTransitionTime":"2025-12-06T05:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.834496 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.834852 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.835295 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.835515 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.835707 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:38Z","lastTransitionTime":"2025-12-06T05:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.888411 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5xvvb_ffb3a4f1-1e22-4220-ac80-3b2d69d2db99/ovnkube-controller/3.log" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.939246 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.939307 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.939324 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.939347 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:38 crc kubenswrapper[4809]: I1206 05:52:38.939364 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:38Z","lastTransitionTime":"2025-12-06T05:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.041585 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.041685 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.041703 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.041728 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.041745 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:39Z","lastTransitionTime":"2025-12-06T05:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.145109 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.145161 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.145177 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.145198 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.145215 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:39Z","lastTransitionTime":"2025-12-06T05:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.247531 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.247571 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.247579 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.247594 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.247607 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:39Z","lastTransitionTime":"2025-12-06T05:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.270517 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.270719 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.270757 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:52:39 crc kubenswrapper[4809]: E1206 05:52:39.270808 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:43.270788338 +0000 UTC m=+148.159771280 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.270844 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:52:39 crc kubenswrapper[4809]: E1206 05:52:39.270855 4809 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 05:52:39 crc kubenswrapper[4809]: E1206 05:52:39.270975 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 05:53:43.270920292 +0000 UTC m=+148.159903284 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 05:52:39 crc kubenswrapper[4809]: E1206 05:52:39.270979 4809 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 05:52:39 crc kubenswrapper[4809]: E1206 05:52:39.270998 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 05:52:39 crc kubenswrapper[4809]: E1206 05:52:39.271050 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 05:52:39 crc kubenswrapper[4809]: E1206 05:52:39.271021 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 05:53:43.271012624 +0000 UTC m=+148.159995686 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 05:52:39 crc kubenswrapper[4809]: E1206 05:52:39.271064 4809 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:52:39 crc kubenswrapper[4809]: E1206 05:52:39.271157 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 05:53:43.271136927 +0000 UTC m=+148.160119889 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.349882 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.349918 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.349932 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.349978 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.349990 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:39Z","lastTransitionTime":"2025-12-06T05:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.371561 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:52:39 crc kubenswrapper[4809]: E1206 05:52:39.371720 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 05:52:39 crc kubenswrapper[4809]: E1206 05:52:39.371799 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 05:52:39 crc kubenswrapper[4809]: E1206 05:52:39.371819 4809 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:52:39 crc kubenswrapper[4809]: E1206 05:52:39.371888 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 05:53:43.371867394 +0000 UTC m=+148.260850346 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.388118 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:52:39 crc kubenswrapper[4809]: E1206 05:52:39.388221 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.388123 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:52:39 crc kubenswrapper[4809]: E1206 05:52:39.388476 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.452525 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.452585 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.452603 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.452626 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.452642 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:39Z","lastTransitionTime":"2025-12-06T05:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.555456 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.555508 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.555520 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.555538 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.555548 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:39Z","lastTransitionTime":"2025-12-06T05:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.658137 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.658198 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.658212 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.658233 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.658248 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:39Z","lastTransitionTime":"2025-12-06T05:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.761125 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.761214 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.761234 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.761255 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.761270 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:39Z","lastTransitionTime":"2025-12-06T05:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.864480 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.864528 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.864540 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.864556 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.864568 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:39Z","lastTransitionTime":"2025-12-06T05:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.966644 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.966692 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.966703 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.966722 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:39 crc kubenswrapper[4809]: I1206 05:52:39.966736 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:39Z","lastTransitionTime":"2025-12-06T05:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.069165 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.069223 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.069240 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.069266 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.069283 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:40Z","lastTransitionTime":"2025-12-06T05:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.171536 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.171603 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.171626 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.171658 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.171682 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:40Z","lastTransitionTime":"2025-12-06T05:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.275251 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.275311 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.275321 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.275337 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.275346 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:40Z","lastTransitionTime":"2025-12-06T05:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.378172 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.379877 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.380100 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.380282 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.380411 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:40Z","lastTransitionTime":"2025-12-06T05:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.387649 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:52:40 crc kubenswrapper[4809]: E1206 05:52:40.387778 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.388386 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:52:40 crc kubenswrapper[4809]: E1206 05:52:40.388566 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.483331 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.483382 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.483393 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.483409 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.483420 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:40Z","lastTransitionTime":"2025-12-06T05:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.488772 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.489044 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.489254 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.489408 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.489525 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:40Z","lastTransitionTime":"2025-12-06T05:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:40 crc kubenswrapper[4809]: E1206 05:52:40.505136 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"88a28b6b-9c85-4932-b5bc-36346303aee5\\\",\\\"systemUUID\\\":\\\"5931df91-c1df-4f78-9e5f-391a01a76127\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.508744 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.508781 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.508794 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.508811 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.508849 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:40Z","lastTransitionTime":"2025-12-06T05:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:40 crc kubenswrapper[4809]: E1206 05:52:40.521645 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"88a28b6b-9c85-4932-b5bc-36346303aee5\\\",\\\"systemUUID\\\":\\\"5931df91-c1df-4f78-9e5f-391a01a76127\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.525654 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.526231 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.526442 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.526616 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.526787 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:40Z","lastTransitionTime":"2025-12-06T05:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:40 crc kubenswrapper[4809]: E1206 05:52:40.540678 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"88a28b6b-9c85-4932-b5bc-36346303aee5\\\",\\\"systemUUID\\\":\\\"5931df91-c1df-4f78-9e5f-391a01a76127\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.543822 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.543857 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.543870 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.543888 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.543899 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:40Z","lastTransitionTime":"2025-12-06T05:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:40 crc kubenswrapper[4809]: E1206 05:52:40.560038 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"88a28b6b-9c85-4932-b5bc-36346303aee5\\\",\\\"systemUUID\\\":\\\"5931df91-c1df-4f78-9e5f-391a01a76127\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.563034 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.563096 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.563107 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.563121 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.563130 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:40Z","lastTransitionTime":"2025-12-06T05:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:40 crc kubenswrapper[4809]: E1206 05:52:40.576660 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"88a28b6b-9c85-4932-b5bc-36346303aee5\\\",\\\"systemUUID\\\":\\\"5931df91-c1df-4f78-9e5f-391a01a76127\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:40 crc kubenswrapper[4809]: E1206 05:52:40.576800 4809 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.586153 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.586197 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.586209 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.586225 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.586240 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:40Z","lastTransitionTime":"2025-12-06T05:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.689088 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.689152 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.689170 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.689195 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.689212 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:40Z","lastTransitionTime":"2025-12-06T05:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.793047 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.793079 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.793099 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.793115 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.793126 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:40Z","lastTransitionTime":"2025-12-06T05:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.895360 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.895409 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.895423 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.895443 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.895458 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:40Z","lastTransitionTime":"2025-12-06T05:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.998276 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.998352 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.998376 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.998405 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:40 crc kubenswrapper[4809]: I1206 05:52:40.998427 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:40Z","lastTransitionTime":"2025-12-06T05:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:41 crc kubenswrapper[4809]: I1206 05:52:41.101858 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:41 crc kubenswrapper[4809]: I1206 05:52:41.101896 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:41 crc kubenswrapper[4809]: I1206 05:52:41.101906 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:41 crc kubenswrapper[4809]: I1206 05:52:41.101922 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:41 crc kubenswrapper[4809]: I1206 05:52:41.101952 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:41Z","lastTransitionTime":"2025-12-06T05:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:41 crc kubenswrapper[4809]: I1206 05:52:41.204759 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:41 crc kubenswrapper[4809]: I1206 05:52:41.204813 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:41 crc kubenswrapper[4809]: I1206 05:52:41.204824 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:41 crc kubenswrapper[4809]: I1206 05:52:41.204840 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:41 crc kubenswrapper[4809]: I1206 05:52:41.204852 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:41Z","lastTransitionTime":"2025-12-06T05:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:41 crc kubenswrapper[4809]: I1206 05:52:41.307077 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:41 crc kubenswrapper[4809]: I1206 05:52:41.307113 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:41 crc kubenswrapper[4809]: I1206 05:52:41.307121 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:41 crc kubenswrapper[4809]: I1206 05:52:41.307134 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:41 crc kubenswrapper[4809]: I1206 05:52:41.307143 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:41Z","lastTransitionTime":"2025-12-06T05:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:41 crc kubenswrapper[4809]: I1206 05:52:41.388385 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:52:41 crc kubenswrapper[4809]: E1206 05:52:41.388678 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:52:41 crc kubenswrapper[4809]: I1206 05:52:41.388770 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:52:41 crc kubenswrapper[4809]: E1206 05:52:41.388973 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:52:41 crc kubenswrapper[4809]: I1206 05:52:41.410094 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:41 crc kubenswrapper[4809]: I1206 05:52:41.410143 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:41 crc kubenswrapper[4809]: I1206 05:52:41.410161 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:41 crc kubenswrapper[4809]: I1206 05:52:41.410183 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:41 crc kubenswrapper[4809]: I1206 05:52:41.410204 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:41Z","lastTransitionTime":"2025-12-06T05:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:41 crc kubenswrapper[4809]: I1206 05:52:41.517874 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:41 crc kubenswrapper[4809]: I1206 05:52:41.517918 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:41 crc kubenswrapper[4809]: I1206 05:52:41.517976 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:41 crc kubenswrapper[4809]: I1206 05:52:41.517995 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:41 crc kubenswrapper[4809]: I1206 05:52:41.518041 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:41Z","lastTransitionTime":"2025-12-06T05:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:41 crc kubenswrapper[4809]: I1206 05:52:41.620409 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:41 crc kubenswrapper[4809]: I1206 05:52:41.620457 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:41 crc kubenswrapper[4809]: I1206 05:52:41.620471 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:41 crc kubenswrapper[4809]: I1206 05:52:41.620491 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:41 crc kubenswrapper[4809]: I1206 05:52:41.620503 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:41Z","lastTransitionTime":"2025-12-06T05:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:41 crc kubenswrapper[4809]: I1206 05:52:41.723541 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:41 crc kubenswrapper[4809]: I1206 05:52:41.723602 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:41 crc kubenswrapper[4809]: I1206 05:52:41.723620 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:41 crc kubenswrapper[4809]: I1206 05:52:41.723643 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:41 crc kubenswrapper[4809]: I1206 05:52:41.723660 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:41Z","lastTransitionTime":"2025-12-06T05:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:41 crc kubenswrapper[4809]: I1206 05:52:41.826439 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:41 crc kubenswrapper[4809]: I1206 05:52:41.826497 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:41 crc kubenswrapper[4809]: I1206 05:52:41.826515 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:41 crc kubenswrapper[4809]: I1206 05:52:41.826540 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:41 crc kubenswrapper[4809]: I1206 05:52:41.826558 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:41Z","lastTransitionTime":"2025-12-06T05:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:41 crc kubenswrapper[4809]: I1206 05:52:41.929095 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:41 crc kubenswrapper[4809]: I1206 05:52:41.929186 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:41 crc kubenswrapper[4809]: I1206 05:52:41.929203 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:41 crc kubenswrapper[4809]: I1206 05:52:41.929230 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:41 crc kubenswrapper[4809]: I1206 05:52:41.929247 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:41Z","lastTransitionTime":"2025-12-06T05:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.032603 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.033058 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.033152 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.033255 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.033343 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:42Z","lastTransitionTime":"2025-12-06T05:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.136848 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.137193 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.137280 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.137320 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.137344 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:42Z","lastTransitionTime":"2025-12-06T05:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.239453 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.239521 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.239538 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.239562 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.239579 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:42Z","lastTransitionTime":"2025-12-06T05:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.341495 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.341526 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.341534 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.341547 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.341557 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:42Z","lastTransitionTime":"2025-12-06T05:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.388454 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.388690 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:52:42 crc kubenswrapper[4809]: E1206 05:52:42.388891 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:52:42 crc kubenswrapper[4809]: E1206 05:52:42.389119 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.445206 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.445274 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.445294 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.445328 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.445347 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:42Z","lastTransitionTime":"2025-12-06T05:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.548216 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.548297 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.548320 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.548349 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.548375 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:42Z","lastTransitionTime":"2025-12-06T05:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.651220 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.651255 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.651263 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.651281 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.651291 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:42Z","lastTransitionTime":"2025-12-06T05:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.753652 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.753731 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.753756 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.753803 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.753858 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:42Z","lastTransitionTime":"2025-12-06T05:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.856606 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.856664 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.856679 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.856700 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.856715 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:42Z","lastTransitionTime":"2025-12-06T05:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.959568 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.959628 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.959645 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.959665 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:42 crc kubenswrapper[4809]: I1206 05:52:42.959677 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:42Z","lastTransitionTime":"2025-12-06T05:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.062294 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.062341 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.062355 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.062371 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.062383 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:43Z","lastTransitionTime":"2025-12-06T05:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.164075 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.164119 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.164128 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.164140 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.164151 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:43Z","lastTransitionTime":"2025-12-06T05:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.268623 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.268664 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.268679 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.268697 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.268709 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:43Z","lastTransitionTime":"2025-12-06T05:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.371002 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.371042 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.371053 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.371068 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.371078 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:43Z","lastTransitionTime":"2025-12-06T05:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.387595 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.387690 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:52:43 crc kubenswrapper[4809]: E1206 05:52:43.388120 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:52:43 crc kubenswrapper[4809]: E1206 05:52:43.387979 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.473909 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.473970 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.473979 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.473993 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.474002 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:43Z","lastTransitionTime":"2025-12-06T05:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.575975 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.576017 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.576051 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.576066 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.576074 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:43Z","lastTransitionTime":"2025-12-06T05:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.679022 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.679084 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.679101 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.679132 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.679148 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:43Z","lastTransitionTime":"2025-12-06T05:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.781716 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.781763 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.781773 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.781788 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.781799 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:43Z","lastTransitionTime":"2025-12-06T05:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.883479 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.883518 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.883529 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.883544 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.883554 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:43Z","lastTransitionTime":"2025-12-06T05:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.986011 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.986079 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.986091 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.986110 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:43 crc kubenswrapper[4809]: I1206 05:52:43.986126 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:43Z","lastTransitionTime":"2025-12-06T05:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:44 crc kubenswrapper[4809]: I1206 05:52:44.089251 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:44 crc kubenswrapper[4809]: I1206 05:52:44.089671 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:44 crc kubenswrapper[4809]: I1206 05:52:44.089749 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:44 crc kubenswrapper[4809]: I1206 05:52:44.089821 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:44 crc kubenswrapper[4809]: I1206 05:52:44.089890 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:44Z","lastTransitionTime":"2025-12-06T05:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:44 crc kubenswrapper[4809]: I1206 05:52:44.192490 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:44 crc kubenswrapper[4809]: I1206 05:52:44.192537 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:44 crc kubenswrapper[4809]: I1206 05:52:44.192548 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:44 crc kubenswrapper[4809]: I1206 05:52:44.192565 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:44 crc kubenswrapper[4809]: I1206 05:52:44.192576 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:44Z","lastTransitionTime":"2025-12-06T05:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:44 crc kubenswrapper[4809]: I1206 05:52:44.294905 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:44 crc kubenswrapper[4809]: I1206 05:52:44.294973 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:44 crc kubenswrapper[4809]: I1206 05:52:44.294986 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:44 crc kubenswrapper[4809]: I1206 05:52:44.295002 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:44 crc kubenswrapper[4809]: I1206 05:52:44.295013 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:44Z","lastTransitionTime":"2025-12-06T05:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:44 crc kubenswrapper[4809]: I1206 05:52:44.388406 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:52:44 crc kubenswrapper[4809]: E1206 05:52:44.388546 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:52:44 crc kubenswrapper[4809]: I1206 05:52:44.388793 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:52:44 crc kubenswrapper[4809]: E1206 05:52:44.388884 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:52:44 crc kubenswrapper[4809]: I1206 05:52:44.397149 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:44 crc kubenswrapper[4809]: I1206 05:52:44.397182 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:44 crc kubenswrapper[4809]: I1206 05:52:44.397193 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:44 crc kubenswrapper[4809]: I1206 05:52:44.397209 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:44 crc kubenswrapper[4809]: I1206 05:52:44.397221 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:44Z","lastTransitionTime":"2025-12-06T05:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:44 crc kubenswrapper[4809]: I1206 05:52:44.499712 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:44 crc kubenswrapper[4809]: I1206 05:52:44.499755 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:44 crc kubenswrapper[4809]: I1206 05:52:44.499792 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:44 crc kubenswrapper[4809]: I1206 05:52:44.499813 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:44 crc kubenswrapper[4809]: I1206 05:52:44.499826 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:44Z","lastTransitionTime":"2025-12-06T05:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:44 crc kubenswrapper[4809]: I1206 05:52:44.602343 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:44 crc kubenswrapper[4809]: I1206 05:52:44.602417 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:44 crc kubenswrapper[4809]: I1206 05:52:44.602434 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:44 crc kubenswrapper[4809]: I1206 05:52:44.602489 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:44 crc kubenswrapper[4809]: I1206 05:52:44.602504 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:44Z","lastTransitionTime":"2025-12-06T05:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:44 crc kubenswrapper[4809]: I1206 05:52:44.705383 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:44 crc kubenswrapper[4809]: I1206 05:52:44.705423 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:44 crc kubenswrapper[4809]: I1206 05:52:44.705433 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:44 crc kubenswrapper[4809]: I1206 05:52:44.705449 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:44 crc kubenswrapper[4809]: I1206 05:52:44.705462 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:44Z","lastTransitionTime":"2025-12-06T05:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:44 crc kubenswrapper[4809]: I1206 05:52:44.808771 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:44 crc kubenswrapper[4809]: I1206 05:52:44.808884 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:44 crc kubenswrapper[4809]: I1206 05:52:44.809038 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:44 crc kubenswrapper[4809]: I1206 05:52:44.809079 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:44 crc kubenswrapper[4809]: I1206 05:52:44.809280 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:44Z","lastTransitionTime":"2025-12-06T05:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:44 crc kubenswrapper[4809]: I1206 05:52:44.911442 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:44 crc kubenswrapper[4809]: I1206 05:52:44.911514 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:44 crc kubenswrapper[4809]: I1206 05:52:44.911539 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:44 crc kubenswrapper[4809]: I1206 05:52:44.911570 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:44 crc kubenswrapper[4809]: I1206 05:52:44.911593 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:44Z","lastTransitionTime":"2025-12-06T05:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.014095 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.014137 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.014148 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.014165 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.014177 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:45Z","lastTransitionTime":"2025-12-06T05:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.116884 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.117002 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.117027 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.117056 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.117075 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:45Z","lastTransitionTime":"2025-12-06T05:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.219602 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.219661 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.219676 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.219696 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.219709 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:45Z","lastTransitionTime":"2025-12-06T05:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.321349 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.321403 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.321414 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.321427 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.321436 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:45Z","lastTransitionTime":"2025-12-06T05:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.388506 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.388566 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:52:45 crc kubenswrapper[4809]: E1206 05:52:45.388674 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:52:45 crc kubenswrapper[4809]: E1206 05:52:45.388773 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.404673 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://010c2681c08dc21bc3a6fb82c04ab3dc2c26dcb76ec12b143c22845cac04ed25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.422618 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c981561d51870deffeb446e8ca769b7ac0a1a16edbe4e982e0f2c84a45043db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d58d489f532b78aca59f5b92584aad1dc292e6700c112c531cbeb3209fd90c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dcd421b38e5f9ba0c542a09f0047bde6c043942a096d11575afb7423ab2552c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab9b3bd4c8a6d6af7939b06fe01b8a0a50d81bb8c669bc6a27f486953c234e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a3e4453935871bb84d68bc4a8b6c98f299b7c8094ba82291bc90e258e1fec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5be86696b7a6518c5df930accb2c6890d58cfc37104fd34a359724b9e30231b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a4a7efb6435b711f6fc3d07f0f82258ae97be843a5917fe8017c2afd690c200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80ffc0fe836258c86321ba5a7f8e1c7e289c0f85e0b48bc138f69cf96cf225ab\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:52:07Z\\\",\\\"message\\\":\\\"52:06.762839 6453 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.59 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {dce28c51-c9f1-478b-97c8-7e209d6e7cbe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 05:52:06.762813 6453 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-dns-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"d937b3b3-82c3-4791-9a66-41b9fed53e9d\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-dns-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, A\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:52:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a4a7efb6435b711f6fc3d07f0f82258ae97be843a5917fe8017c2afd690c200\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:52:37Z\\\",\\\"message\\\":\\\"d already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:36Z is after 2025-08-24T17:21:41Z]\\\\nI1206 05:52:36.258911 6813 services_controller.go:451] Built service openshift-network-console/networking-console-plugin cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-network-console/networking-console-plugin_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.246\\\\\\\", Port:9443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1206 05:52:36.258967 6813 services_controller.go:452] Built service openshift-network-\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dee8b965be3b2486f1f532f2bdfbbcbd84f9b3b711a11c25e86dae1284314d4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5xvvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.424568 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.424623 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.424639 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.424658 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.424678 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:45Z","lastTransitionTime":"2025-12-06T05:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.442779 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa4dd3f8-c119-46de-ad4e-dd29bb11a32c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa59e4ec9867cf280b308ee3adb5e303eaf94455aca588f5eb74a2b6263d57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5720f7f2977d9af6ed5324a30d08d34a53d55460951b3875b7d7d753d202f349\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dfd977b049a8ff81192d4041adb265366ab88475ee84c5750305d061d21c8db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d5801b61a87c95c9a623c3106e305920199c640efa8a32cf57698245697e54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.455078 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7b8a286-4824-49e9-bcd3-77ed87fd7638\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95eafea368aebd35fb8e9a20bb01938c0ee9b1109d1a4ac5c1062e0d4ce7465c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a37233831095a0ba4b488c136d3ede54ee833ab0d0141448f20ddb34fd4cc8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a37233831095a0ba4b488c136d3ede54ee833ab0d0141448f20ddb34fd4cc8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.469022 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd45866c51410e910e87d43870bf531b0d05541e9164e700150fd37987e284af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.481582 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.492053 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.504115 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.515699 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947d89be9d04856af8d09facb8d4583f91898bd03a36d3ce8b9b74bc3bb795e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://930788ec9b4d97fdc39353a93111b84707037ce445f1930aa7737a20a44e1d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-npms2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.528461 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.528533 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.528547 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.528574 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.528589 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:45Z","lastTransitionTime":"2025-12-06T05:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.529540 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5636dd78-3d61-4c6a-b8fd-6d2457e19234\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.540878 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"353c3bc7-39c3-4f30-ab37-9883caae33f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d4b8272a98d4acd09f165939684c0990b9e56b1269f85c20d2bd41c0f75f67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ddfff8cca35bda4796fe58ace60506f5c87e4398193d6711e81aae0d4b7d729\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae336ca136d0a51627e63b1ec86bd4f916a5efaf804e5b91014188331c19802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6965798de9aaf2b8066b7bd8abc23e12545f026bf1f0930dc0085b992553f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc6965798de9aaf2b8066b7bd8abc23e12545f026bf1f0930dc0085b992553f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.550435 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rnzq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0075197-9392-4109-b8a7-39507b15cc17\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53a00b17138b236dc8517cbd4ee71ea3ed7e8885247d06eb2e8840b8b145783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4mgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rnzq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.561308 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9k8zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4f75fd3-03e8-41e7-8926-f6ff21a5b681\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpdh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpdh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9k8zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.574749 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nndh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a80ba5e1-90cc-44ce-be19-fdf0c007d5b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3684c4e5e839ce117a1772571ae080042e0187ea1aaca635edca07ac10d94c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2c22k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31d74665ae370eed09b8a3195bd3d122ad64c6e33bf55fc42cd40a862671b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2c22k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nndh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.587147 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f28d0a3aec8a36273fa4480c552d80a5ee23210203861811e37f27bccda1147e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40791d93876595c4093814146d269b8fe70efb64ef2888e03c25aebef0e9295e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.600027 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a48de155-7389-48db-8f86-d175b7db70c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://701274e49f68efbc0a939bc08771288c50662cef9c59da73170f7f114e222337\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9t7pq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.617611 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cb4vc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed42d686-41af-470e-910d-a1fabbec66b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2edf7fa65761bc560178a40857780a5668e183a4b01974935cdbc3540915114f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73f629b20d9f61227553520b30d0ca046750433926a8d31501e5fb77467426cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:52:22Z\\\",\\\"message\\\":\\\"2025-12-06T05:51:36+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_4426309b-ddbc-4734-9136-9da7465cfbc2\\\\n2025-12-06T05:51:36+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_4426309b-ddbc-4734-9136-9da7465cfbc2 to /host/opt/cni/bin/\\\\n2025-12-06T05:51:36Z [verbose] multus-daemon started\\\\n2025-12-06T05:51:36Z [verbose] Readiness Indicator file check\\\\n2025-12-06T05:52:21Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2nwpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cb4vc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.630438 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hmlmc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c7558c-89b1-432d-893c-ccb2b3aba183\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c2f8876171399b61abd26285afba0547dac3c24c957bd6d202721b04c912aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm7wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hmlmc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.630919 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.630983 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.630996 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.631013 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.631050 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:45Z","lastTransitionTime":"2025-12-06T05:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.734035 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.734108 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.734119 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.734133 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.734143 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:45Z","lastTransitionTime":"2025-12-06T05:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.837283 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.837377 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.837395 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.837418 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.837466 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:45Z","lastTransitionTime":"2025-12-06T05:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.939879 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.939961 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.939979 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.940001 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:45 crc kubenswrapper[4809]: I1206 05:52:45.940018 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:45Z","lastTransitionTime":"2025-12-06T05:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.042826 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.042922 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.043002 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.043026 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.043043 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:46Z","lastTransitionTime":"2025-12-06T05:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.146251 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.146305 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.146322 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.146344 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.146361 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:46Z","lastTransitionTime":"2025-12-06T05:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.247999 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.248042 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.248053 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.248070 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.248082 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:46Z","lastTransitionTime":"2025-12-06T05:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.350274 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.350356 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.350378 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.350410 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.350430 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:46Z","lastTransitionTime":"2025-12-06T05:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.388133 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.388277 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:52:46 crc kubenswrapper[4809]: E1206 05:52:46.388329 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:52:46 crc kubenswrapper[4809]: E1206 05:52:46.388482 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.453142 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.453198 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.453210 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.453229 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.453244 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:46Z","lastTransitionTime":"2025-12-06T05:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.555904 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.556019 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.556039 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.556061 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.556079 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:46Z","lastTransitionTime":"2025-12-06T05:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.658674 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.658753 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.658775 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.658801 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.658819 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:46Z","lastTransitionTime":"2025-12-06T05:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.761697 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.761762 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.761781 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.761807 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.761826 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:46Z","lastTransitionTime":"2025-12-06T05:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.865537 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.865605 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.865629 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.865662 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.865682 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:46Z","lastTransitionTime":"2025-12-06T05:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.968424 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.968482 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.968500 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.968523 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:46 crc kubenswrapper[4809]: I1206 05:52:46.968570 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:46Z","lastTransitionTime":"2025-12-06T05:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.072121 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.072178 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.072214 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.072251 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.072275 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:47Z","lastTransitionTime":"2025-12-06T05:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.175980 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.176111 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.176122 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.176137 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.176146 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:47Z","lastTransitionTime":"2025-12-06T05:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.279336 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.279374 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.279382 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.279396 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.279405 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:47Z","lastTransitionTime":"2025-12-06T05:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.381837 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.381889 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.381901 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.381920 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.381950 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:47Z","lastTransitionTime":"2025-12-06T05:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.388448 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.388822 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:52:47 crc kubenswrapper[4809]: E1206 05:52:47.389046 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:52:47 crc kubenswrapper[4809]: E1206 05:52:47.389198 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.484287 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.484346 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.484372 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.484393 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.484410 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:47Z","lastTransitionTime":"2025-12-06T05:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.587071 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.587134 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.587143 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.587158 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.587168 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:47Z","lastTransitionTime":"2025-12-06T05:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.689913 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.689956 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.689966 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.689979 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.689987 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:47Z","lastTransitionTime":"2025-12-06T05:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.792879 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.792926 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.792962 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.792979 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.792989 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:47Z","lastTransitionTime":"2025-12-06T05:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.895458 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.895500 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.895508 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.895522 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.895534 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:47Z","lastTransitionTime":"2025-12-06T05:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.998630 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.998692 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.998706 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.998727 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:47 crc kubenswrapper[4809]: I1206 05:52:47.998740 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:47Z","lastTransitionTime":"2025-12-06T05:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:48 crc kubenswrapper[4809]: I1206 05:52:48.101474 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:48 crc kubenswrapper[4809]: I1206 05:52:48.101547 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:48 crc kubenswrapper[4809]: I1206 05:52:48.101561 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:48 crc kubenswrapper[4809]: I1206 05:52:48.101578 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:48 crc kubenswrapper[4809]: I1206 05:52:48.101592 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:48Z","lastTransitionTime":"2025-12-06T05:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:48 crc kubenswrapper[4809]: I1206 05:52:48.204371 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:48 crc kubenswrapper[4809]: I1206 05:52:48.204420 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:48 crc kubenswrapper[4809]: I1206 05:52:48.204431 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:48 crc kubenswrapper[4809]: I1206 05:52:48.204447 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:48 crc kubenswrapper[4809]: I1206 05:52:48.204456 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:48Z","lastTransitionTime":"2025-12-06T05:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:48 crc kubenswrapper[4809]: I1206 05:52:48.307354 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:48 crc kubenswrapper[4809]: I1206 05:52:48.307389 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:48 crc kubenswrapper[4809]: I1206 05:52:48.307399 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:48 crc kubenswrapper[4809]: I1206 05:52:48.307414 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:48 crc kubenswrapper[4809]: I1206 05:52:48.307425 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:48Z","lastTransitionTime":"2025-12-06T05:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:48 crc kubenswrapper[4809]: I1206 05:52:48.388483 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:52:48 crc kubenswrapper[4809]: I1206 05:52:48.388502 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:52:48 crc kubenswrapper[4809]: E1206 05:52:48.388625 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:52:48 crc kubenswrapper[4809]: E1206 05:52:48.388712 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:52:48 crc kubenswrapper[4809]: I1206 05:52:48.410084 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:48 crc kubenswrapper[4809]: I1206 05:52:48.410158 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:48 crc kubenswrapper[4809]: I1206 05:52:48.410176 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:48 crc kubenswrapper[4809]: I1206 05:52:48.410196 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:48 crc kubenswrapper[4809]: I1206 05:52:48.410212 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:48Z","lastTransitionTime":"2025-12-06T05:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:48 crc kubenswrapper[4809]: I1206 05:52:48.513328 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:48 crc kubenswrapper[4809]: I1206 05:52:48.513374 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:48 crc kubenswrapper[4809]: I1206 05:52:48.513390 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:48 crc kubenswrapper[4809]: I1206 05:52:48.513411 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:48 crc kubenswrapper[4809]: I1206 05:52:48.513428 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:48Z","lastTransitionTime":"2025-12-06T05:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:48 crc kubenswrapper[4809]: I1206 05:52:48.616391 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:48 crc kubenswrapper[4809]: I1206 05:52:48.616430 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:48 crc kubenswrapper[4809]: I1206 05:52:48.616440 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:48 crc kubenswrapper[4809]: I1206 05:52:48.616456 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:48 crc kubenswrapper[4809]: I1206 05:52:48.616469 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:48Z","lastTransitionTime":"2025-12-06T05:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:48 crc kubenswrapper[4809]: I1206 05:52:48.719226 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:48 crc kubenswrapper[4809]: I1206 05:52:48.719300 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:48 crc kubenswrapper[4809]: I1206 05:52:48.719323 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:48 crc kubenswrapper[4809]: I1206 05:52:48.719356 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:48 crc kubenswrapper[4809]: I1206 05:52:48.719380 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:48Z","lastTransitionTime":"2025-12-06T05:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:48 crc kubenswrapper[4809]: I1206 05:52:48.822639 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:48 crc kubenswrapper[4809]: I1206 05:52:48.822803 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:48 crc kubenswrapper[4809]: I1206 05:52:48.822843 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:48 crc kubenswrapper[4809]: I1206 05:52:48.822886 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:48 crc kubenswrapper[4809]: I1206 05:52:48.822918 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:48Z","lastTransitionTime":"2025-12-06T05:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:48 crc kubenswrapper[4809]: I1206 05:52:48.925074 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:48 crc kubenswrapper[4809]: I1206 05:52:48.925135 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:48 crc kubenswrapper[4809]: I1206 05:52:48.925152 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:48 crc kubenswrapper[4809]: I1206 05:52:48.925177 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:48 crc kubenswrapper[4809]: I1206 05:52:48.925193 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:48Z","lastTransitionTime":"2025-12-06T05:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.028255 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.028291 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.028299 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.028314 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.028324 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:49Z","lastTransitionTime":"2025-12-06T05:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.130471 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.130544 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.130568 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.130595 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.130616 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:49Z","lastTransitionTime":"2025-12-06T05:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.233565 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.233662 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.234115 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.234191 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.234211 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:49Z","lastTransitionTime":"2025-12-06T05:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.337621 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.337694 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.337718 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.337745 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.337767 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:49Z","lastTransitionTime":"2025-12-06T05:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.387990 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.388020 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:52:49 crc kubenswrapper[4809]: E1206 05:52:49.388227 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:52:49 crc kubenswrapper[4809]: E1206 05:52:49.388647 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.440379 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.440425 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.440436 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.440460 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.440474 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:49Z","lastTransitionTime":"2025-12-06T05:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.543428 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.543488 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.543503 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.543522 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.543536 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:49Z","lastTransitionTime":"2025-12-06T05:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.646274 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.646324 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.646334 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.646350 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.646361 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:49Z","lastTransitionTime":"2025-12-06T05:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.752265 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.752316 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.752329 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.752349 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.752360 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:49Z","lastTransitionTime":"2025-12-06T05:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.854796 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.854852 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.854863 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.854880 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.854892 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:49Z","lastTransitionTime":"2025-12-06T05:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.957424 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.957512 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.957547 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.957577 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:49 crc kubenswrapper[4809]: I1206 05:52:49.957601 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:49Z","lastTransitionTime":"2025-12-06T05:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.060619 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.060678 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.060694 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.060715 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.060731 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:50Z","lastTransitionTime":"2025-12-06T05:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.163171 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.163225 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.163239 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.163256 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.163267 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:50Z","lastTransitionTime":"2025-12-06T05:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.266225 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.266307 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.266319 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.266337 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.266349 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:50Z","lastTransitionTime":"2025-12-06T05:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.369015 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.369064 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.369076 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.369095 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.369107 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:50Z","lastTransitionTime":"2025-12-06T05:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.387597 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.387679 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:52:50 crc kubenswrapper[4809]: E1206 05:52:50.387745 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:52:50 crc kubenswrapper[4809]: E1206 05:52:50.388230 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.388820 4809 scope.go:117] "RemoveContainer" containerID="4a4a7efb6435b711f6fc3d07f0f82258ae97be843a5917fe8017c2afd690c200" Dec 06 05:52:50 crc kubenswrapper[4809]: E1206 05:52:50.389150 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5xvvb_openshift-ovn-kubernetes(ffb3a4f1-1e22-4220-ac80-3b2d69d2db99)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.400857 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cb4vc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed42d686-41af-470e-910d-a1fabbec66b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2edf7fa65761bc560178a40857780a5668e183a4b01974935cdbc3540915114f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73f629b20d9f61227553520b30d0ca046750433926a8d31501e5fb77467426cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:52:22Z\\\",\\\"message\\\":\\\"2025-12-06T05:51:36+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_4426309b-ddbc-4734-9136-9da7465cfbc2\\\\n2025-12-06T05:51:36+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_4426309b-ddbc-4734-9136-9da7465cfbc2 to /host/opt/cni/bin/\\\\n2025-12-06T05:51:36Z [verbose] multus-daemon started\\\\n2025-12-06T05:51:36Z [verbose] Readiness Indicator file check\\\\n2025-12-06T05:52:21Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2nwpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cb4vc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.410965 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hmlmc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c7558c-89b1-432d-893c-ccb2b3aba183\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c2f8876171399b61abd26285afba0547dac3c24c957bd6d202721b04c912aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm7wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hmlmc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.425736 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa4dd3f8-c119-46de-ad4e-dd29bb11a32c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa59e4ec9867cf280b308ee3adb5e303eaf94455aca588f5eb74a2b6263d57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5720f7f2977d9af6ed5324a30d08d34a53d55460951b3875b7d7d753d202f349\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dfd977b049a8ff81192d4041adb265366ab88475ee84c5750305d061d21c8db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d5801b61a87c95c9a623c3106e305920199c640efa8a32cf57698245697e54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.438635 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://010c2681c08dc21bc3a6fb82c04ab3dc2c26dcb76ec12b143c22845cac04ed25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.460157 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c981561d51870deffeb446e8ca769b7ac0a1a16edbe4e982e0f2c84a45043db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d58d489f532b78aca59f5b92584aad1dc292e6700c112c531cbeb3209fd90c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dcd421b38e5f9ba0c542a09f0047bde6c043942a096d11575afb7423ab2552c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab9b3bd4c8a6d6af7939b06fe01b8a0a50d81bb8c669bc6a27f486953c234e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a3e4453935871bb84d68bc4a8b6c98f299b7c8094ba82291bc90e258e1fec55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5be86696b7a6518c5df930accb2c6890d58cfc37104fd34a359724b9e30231b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a4a7efb6435b711f6fc3d07f0f82258ae97be843a5917fe8017c2afd690c200\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a4a7efb6435b711f6fc3d07f0f82258ae97be843a5917fe8017c2afd690c200\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:52:37Z\\\",\\\"message\\\":\\\"d already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:36Z is after 2025-08-24T17:21:41Z]\\\\nI1206 05:52:36.258911 6813 services_controller.go:451] Built service openshift-network-console/networking-console-plugin cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-network-console/networking-console-plugin_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.246\\\\\\\", Port:9443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1206 05:52:36.258967 6813 services_controller.go:452] Built service openshift-network-\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:52:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5xvvb_openshift-ovn-kubernetes(ffb3a4f1-1e22-4220-ac80-3b2d69d2db99)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dee8b965be3b2486f1f532f2bdfbbcbd84f9b3b711a11c25e86dae1284314d4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-227vc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5xvvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.472571 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.472725 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.472738 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.472754 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.472765 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:50Z","lastTransitionTime":"2025-12-06T05:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.472701 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.487119 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.501576 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.514050 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbf4cf62-024e-4703-a8b8-9aecda9cd26a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947d89be9d04856af8d09facb8d4583f91898bd03a36d3ce8b9b74bc3bb795e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://930788ec9b4d97fdc39353a93111b84707037ce445f1930aa7737a20a44e1d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mqln\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-npms2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.529181 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5636dd78-3d61-4c6a-b8fd-6d2457e19234\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.543788 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"353c3bc7-39c3-4f30-ab37-9883caae33f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54d4b8272a98d4acd09f165939684c0990b9e56b1269f85c20d2bd41c0f75f67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ddfff8cca35bda4796fe58ace60506f5c87e4398193d6711e81aae0d4b7d729\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae336ca136d0a51627e63b1ec86bd4f916a5efaf804e5b91014188331c19802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6965798de9aaf2b8066b7bd8abc23e12545f026bf1f0930dc0085b992553f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc6965798de9aaf2b8066b7bd8abc23e12545f026bf1f0930dc0085b992553f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.562581 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7b8a286-4824-49e9-bcd3-77ed87fd7638\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95eafea368aebd35fb8e9a20bb01938c0ee9b1109d1a4ac5c1062e0d4ce7465c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a37233831095a0ba4b488c136d3ede54ee833ab0d0141448f20ddb34fd4cc8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a37233831095a0ba4b488c136d3ede54ee833ab0d0141448f20ddb34fd4cc8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.575723 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.575770 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.575783 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.575800 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.575811 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:50Z","lastTransitionTime":"2025-12-06T05:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.578259 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd45866c51410e910e87d43870bf531b0d05541e9164e700150fd37987e284af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.592435 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rnzq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0075197-9392-4109-b8a7-39507b15cc17\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53a00b17138b236dc8517cbd4ee71ea3ed7e8885247d06eb2e8840b8b145783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4mgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rnzq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.602180 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9k8zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4f75fd3-03e8-41e7-8926-f6ff21a5b681\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpdh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpdh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9k8zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.614542 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f28d0a3aec8a36273fa4480c552d80a5ee23210203861811e37f27bccda1147e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40791d93876595c4093814146d269b8fe70efb64ef2888e03c25aebef0e9295e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.629514 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a48de155-7389-48db-8f86-d175b7db70c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://701274e49f68efbc0a939bc08771288c50662cef9c59da73170f7f114e222337\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7604fc02594ff0d9d6a2d6548072259adac9725e40d48f89cb1a683c53c3c66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7d90ad7a54df4c8e4163128091b4597adb49a7abd41018e2b1e50c9ea0c233d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bedaeb88e38a0a02a08321013cbc66533af36f96e6451c655080e689af145b1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f782901aff6789653f2fd56228d662ff0dfe1950198d76ce9bb4a2a0280127f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d81e3435f869d42681f8512ee7f43ed44046d3375bbead9abd3d9fad33fb61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fc22a2872573288183b85290dcd66a1bacdf0161fe0a4c6694aa76662709b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4z265\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9t7pq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.641021 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nndh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a80ba5e1-90cc-44ce-be19-fdf0c007d5b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:51:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3684c4e5e839ce117a1772571ae080042e0187ea1aaca635edca07ac10d94c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2c22k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31d74665ae370eed09b8a3195bd3d122ad64c6e33bf55fc42cd40a862671b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:51:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2c22k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:51:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nndh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.678368 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.678456 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.678474 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.678498 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.678541 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:50Z","lastTransitionTime":"2025-12-06T05:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.781081 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.781146 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.781160 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.781180 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.781195 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:50Z","lastTransitionTime":"2025-12-06T05:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.883967 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.884023 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.884037 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.884086 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.884102 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:50Z","lastTransitionTime":"2025-12-06T05:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.906798 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.906842 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.906856 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.906893 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.906908 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:50Z","lastTransitionTime":"2025-12-06T05:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:50 crc kubenswrapper[4809]: E1206 05:52:50.919337 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"88a28b6b-9c85-4932-b5bc-36346303aee5\\\",\\\"systemUUID\\\":\\\"5931df91-c1df-4f78-9e5f-391a01a76127\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.922448 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.922482 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.922494 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.922509 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.922521 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:50Z","lastTransitionTime":"2025-12-06T05:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:50 crc kubenswrapper[4809]: E1206 05:52:50.936727 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"88a28b6b-9c85-4932-b5bc-36346303aee5\\\",\\\"systemUUID\\\":\\\"5931df91-c1df-4f78-9e5f-391a01a76127\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.941249 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.941279 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.941290 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.941303 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.941314 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:50Z","lastTransitionTime":"2025-12-06T05:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:50 crc kubenswrapper[4809]: E1206 05:52:50.953091 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"88a28b6b-9c85-4932-b5bc-36346303aee5\\\",\\\"systemUUID\\\":\\\"5931df91-c1df-4f78-9e5f-391a01a76127\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.957712 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.957738 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.957747 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.957760 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.957769 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:50Z","lastTransitionTime":"2025-12-06T05:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:50 crc kubenswrapper[4809]: E1206 05:52:50.971506 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"88a28b6b-9c85-4932-b5bc-36346303aee5\\\",\\\"systemUUID\\\":\\\"5931df91-c1df-4f78-9e5f-391a01a76127\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.974709 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.974775 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.974793 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.975230 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.975331 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:50Z","lastTransitionTime":"2025-12-06T05:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:50 crc kubenswrapper[4809]: E1206 05:52:50.988664 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:52:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:52:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"88a28b6b-9c85-4932-b5bc-36346303aee5\\\",\\\"systemUUID\\\":\\\"5931df91-c1df-4f78-9e5f-391a01a76127\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:52:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:52:50 crc kubenswrapper[4809]: E1206 05:52:50.988893 4809 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.990472 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.990510 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.990525 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.990547 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:50 crc kubenswrapper[4809]: I1206 05:52:50.990562 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:50Z","lastTransitionTime":"2025-12-06T05:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:51 crc kubenswrapper[4809]: I1206 05:52:51.093308 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:51 crc kubenswrapper[4809]: I1206 05:52:51.093360 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:51 crc kubenswrapper[4809]: I1206 05:52:51.093372 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:51 crc kubenswrapper[4809]: I1206 05:52:51.093391 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:51 crc kubenswrapper[4809]: I1206 05:52:51.093404 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:51Z","lastTransitionTime":"2025-12-06T05:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:51 crc kubenswrapper[4809]: I1206 05:52:51.195960 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:51 crc kubenswrapper[4809]: I1206 05:52:51.196001 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:51 crc kubenswrapper[4809]: I1206 05:52:51.196018 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:51 crc kubenswrapper[4809]: I1206 05:52:51.196040 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:51 crc kubenswrapper[4809]: I1206 05:52:51.196056 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:51Z","lastTransitionTime":"2025-12-06T05:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:51 crc kubenswrapper[4809]: I1206 05:52:51.298439 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:51 crc kubenswrapper[4809]: I1206 05:52:51.298482 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:51 crc kubenswrapper[4809]: I1206 05:52:51.298492 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:51 crc kubenswrapper[4809]: I1206 05:52:51.298508 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:51 crc kubenswrapper[4809]: I1206 05:52:51.298521 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:51Z","lastTransitionTime":"2025-12-06T05:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:51 crc kubenswrapper[4809]: I1206 05:52:51.388110 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:52:51 crc kubenswrapper[4809]: I1206 05:52:51.388166 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:52:51 crc kubenswrapper[4809]: E1206 05:52:51.388270 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:52:51 crc kubenswrapper[4809]: E1206 05:52:51.388412 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:52:51 crc kubenswrapper[4809]: I1206 05:52:51.403764 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:51 crc kubenswrapper[4809]: I1206 05:52:51.403815 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:51 crc kubenswrapper[4809]: I1206 05:52:51.403825 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:51 crc kubenswrapper[4809]: I1206 05:52:51.403839 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:51 crc kubenswrapper[4809]: I1206 05:52:51.403847 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:51Z","lastTransitionTime":"2025-12-06T05:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:51 crc kubenswrapper[4809]: I1206 05:52:51.506731 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:51 crc kubenswrapper[4809]: I1206 05:52:51.506780 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:51 crc kubenswrapper[4809]: I1206 05:52:51.506791 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:51 crc kubenswrapper[4809]: I1206 05:52:51.506807 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:51 crc kubenswrapper[4809]: I1206 05:52:51.506818 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:51Z","lastTransitionTime":"2025-12-06T05:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:51 crc kubenswrapper[4809]: I1206 05:52:51.609281 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:51 crc kubenswrapper[4809]: I1206 05:52:51.609320 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:51 crc kubenswrapper[4809]: I1206 05:52:51.609330 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:51 crc kubenswrapper[4809]: I1206 05:52:51.609343 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:51 crc kubenswrapper[4809]: I1206 05:52:51.609352 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:51Z","lastTransitionTime":"2025-12-06T05:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:51 crc kubenswrapper[4809]: I1206 05:52:51.711591 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:51 crc kubenswrapper[4809]: I1206 05:52:51.711647 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:51 crc kubenswrapper[4809]: I1206 05:52:51.711658 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:51 crc kubenswrapper[4809]: I1206 05:52:51.711673 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:51 crc kubenswrapper[4809]: I1206 05:52:51.711683 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:51Z","lastTransitionTime":"2025-12-06T05:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:51 crc kubenswrapper[4809]: I1206 05:52:51.813817 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:51 crc kubenswrapper[4809]: I1206 05:52:51.813862 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:51 crc kubenswrapper[4809]: I1206 05:52:51.813881 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:51 crc kubenswrapper[4809]: I1206 05:52:51.813900 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:51 crc kubenswrapper[4809]: I1206 05:52:51.813912 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:51Z","lastTransitionTime":"2025-12-06T05:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:51 crc kubenswrapper[4809]: I1206 05:52:51.916872 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:51 crc kubenswrapper[4809]: I1206 05:52:51.916915 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:51 crc kubenswrapper[4809]: I1206 05:52:51.916927 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:51 crc kubenswrapper[4809]: I1206 05:52:51.916959 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:51 crc kubenswrapper[4809]: I1206 05:52:51.916970 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:51Z","lastTransitionTime":"2025-12-06T05:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.019769 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.019816 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.019830 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.019846 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.019858 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:52Z","lastTransitionTime":"2025-12-06T05:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.122814 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.122888 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.122917 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.122952 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.122961 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:52Z","lastTransitionTime":"2025-12-06T05:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.225440 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.225481 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.225492 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.225510 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.225521 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:52Z","lastTransitionTime":"2025-12-06T05:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.327852 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.327898 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.327909 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.327925 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.327958 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:52Z","lastTransitionTime":"2025-12-06T05:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.387865 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.387908 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:52:52 crc kubenswrapper[4809]: E1206 05:52:52.388040 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:52:52 crc kubenswrapper[4809]: E1206 05:52:52.388403 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.430273 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.430329 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.430345 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.430361 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.430373 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:52Z","lastTransitionTime":"2025-12-06T05:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.532584 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.532672 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.532684 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.532701 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.532713 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:52Z","lastTransitionTime":"2025-12-06T05:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.635358 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.635691 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.635788 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.636017 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.636104 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:52Z","lastTransitionTime":"2025-12-06T05:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.738764 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.738803 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.738812 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.738826 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.738837 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:52Z","lastTransitionTime":"2025-12-06T05:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.840501 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.840554 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.840564 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.840579 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.840589 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:52Z","lastTransitionTime":"2025-12-06T05:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.942460 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.942504 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.942516 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.942530 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:52 crc kubenswrapper[4809]: I1206 05:52:52.942541 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:52Z","lastTransitionTime":"2025-12-06T05:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.016559 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a4f75fd3-03e8-41e7-8926-f6ff21a5b681-metrics-certs\") pod \"network-metrics-daemon-9k8zx\" (UID: \"a4f75fd3-03e8-41e7-8926-f6ff21a5b681\") " pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:52:53 crc kubenswrapper[4809]: E1206 05:52:53.016714 4809 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 05:52:53 crc kubenswrapper[4809]: E1206 05:52:53.016767 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a4f75fd3-03e8-41e7-8926-f6ff21a5b681-metrics-certs podName:a4f75fd3-03e8-41e7-8926-f6ff21a5b681 nodeName:}" failed. No retries permitted until 2025-12-06 05:53:57.01675099 +0000 UTC m=+161.905733932 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a4f75fd3-03e8-41e7-8926-f6ff21a5b681-metrics-certs") pod "network-metrics-daemon-9k8zx" (UID: "a4f75fd3-03e8-41e7-8926-f6ff21a5b681") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.045234 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.045282 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.045294 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.045309 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.045323 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:53Z","lastTransitionTime":"2025-12-06T05:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.148161 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.148197 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.148208 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.148222 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.148231 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:53Z","lastTransitionTime":"2025-12-06T05:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.251151 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.251208 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.251221 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.251243 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.251255 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:53Z","lastTransitionTime":"2025-12-06T05:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.354183 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.354254 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.354278 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.354308 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.354338 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:53Z","lastTransitionTime":"2025-12-06T05:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.388050 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.388094 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:52:53 crc kubenswrapper[4809]: E1206 05:52:53.388394 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:52:53 crc kubenswrapper[4809]: E1206 05:52:53.388659 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.456573 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.456603 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.456614 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.456628 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.456647 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:53Z","lastTransitionTime":"2025-12-06T05:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.559309 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.559357 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.559369 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.559386 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.559398 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:53Z","lastTransitionTime":"2025-12-06T05:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.661537 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.661594 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.661605 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.661621 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.661631 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:53Z","lastTransitionTime":"2025-12-06T05:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.763742 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.763824 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.763833 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.763847 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.763857 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:53Z","lastTransitionTime":"2025-12-06T05:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.866498 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.866540 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.866551 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.866570 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.866580 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:53Z","lastTransitionTime":"2025-12-06T05:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.970267 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.970310 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.970319 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.970337 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:53 crc kubenswrapper[4809]: I1206 05:52:53.970347 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:53Z","lastTransitionTime":"2025-12-06T05:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.073490 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.073531 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.073542 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.073558 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.073569 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:54Z","lastTransitionTime":"2025-12-06T05:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.176199 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.176243 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.176254 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.176269 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.176280 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:54Z","lastTransitionTime":"2025-12-06T05:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.278347 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.278372 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.278384 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.278399 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.278413 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:54Z","lastTransitionTime":"2025-12-06T05:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.381987 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.382039 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.382050 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.382066 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.382077 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:54Z","lastTransitionTime":"2025-12-06T05:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.387487 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.387600 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:52:54 crc kubenswrapper[4809]: E1206 05:52:54.387760 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:52:54 crc kubenswrapper[4809]: E1206 05:52:54.387883 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.484701 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.484751 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.484761 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.484781 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.484792 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:54Z","lastTransitionTime":"2025-12-06T05:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.587325 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.587390 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.587417 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.587440 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.587456 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:54Z","lastTransitionTime":"2025-12-06T05:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.689651 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.689697 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.689713 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.689731 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.689744 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:54Z","lastTransitionTime":"2025-12-06T05:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.792608 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.792644 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.792654 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.792671 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.792683 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:54Z","lastTransitionTime":"2025-12-06T05:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.895396 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.895429 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.895437 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.895448 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.895457 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:54Z","lastTransitionTime":"2025-12-06T05:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.998548 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.998629 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.998656 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.998686 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:54 crc kubenswrapper[4809]: I1206 05:52:54.998708 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:54Z","lastTransitionTime":"2025-12-06T05:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.102192 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.102274 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.102300 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.102328 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.102392 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:55Z","lastTransitionTime":"2025-12-06T05:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.205770 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.205845 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.205861 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.205883 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.205898 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:55Z","lastTransitionTime":"2025-12-06T05:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.308556 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.308601 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.308611 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.308630 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.308643 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:55Z","lastTransitionTime":"2025-12-06T05:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.388993 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.389092 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:52:55 crc kubenswrapper[4809]: E1206 05:52:55.389390 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:52:55 crc kubenswrapper[4809]: E1206 05:52:55.389560 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.413057 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.413124 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.413135 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.413148 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.413161 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:55Z","lastTransitionTime":"2025-12-06T05:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.462530 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-9t7pq" podStartSLOduration=81.462510647 podStartE2EDuration="1m21.462510647s" podCreationTimestamp="2025-12-06 05:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:52:55.437031333 +0000 UTC m=+100.326014285" watchObservedRunningTime="2025-12-06 05:52:55.462510647 +0000 UTC m=+100.351493589" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.482037 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-cb4vc" podStartSLOduration=81.482014665 podStartE2EDuration="1m21.482014665s" podCreationTimestamp="2025-12-06 05:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:52:55.481589185 +0000 UTC m=+100.370572127" watchObservedRunningTime="2025-12-06 05:52:55.482014665 +0000 UTC m=+100.370997607" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.482189 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nndh4" podStartSLOduration=80.48218177 podStartE2EDuration="1m20.48218177s" podCreationTimestamp="2025-12-06 05:51:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:52:55.462430345 +0000 UTC m=+100.351413287" watchObservedRunningTime="2025-12-06 05:52:55.48218177 +0000 UTC m=+100.371164712" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.492189 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-hmlmc" podStartSLOduration=81.49216254 podStartE2EDuration="1m21.49216254s" podCreationTimestamp="2025-12-06 05:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:52:55.491668087 +0000 UTC m=+100.380651039" watchObservedRunningTime="2025-12-06 05:52:55.49216254 +0000 UTC m=+100.381145492" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.515118 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.515164 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.515175 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.515188 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.515198 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:55Z","lastTransitionTime":"2025-12-06T05:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.533059 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=81.533040736 podStartE2EDuration="1m21.533040736s" podCreationTimestamp="2025-12-06 05:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:52:55.532583334 +0000 UTC m=+100.421566276" watchObservedRunningTime="2025-12-06 05:52:55.533040736 +0000 UTC m=+100.422023678" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.616818 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podStartSLOduration=81.616796659 podStartE2EDuration="1m21.616796659s" podCreationTimestamp="2025-12-06 05:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:52:55.616462261 +0000 UTC m=+100.505445213" watchObservedRunningTime="2025-12-06 05:52:55.616796659 +0000 UTC m=+100.505779601" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.617125 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.617149 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.617157 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.617170 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.617178 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:55Z","lastTransitionTime":"2025-12-06T05:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.633328 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=81.63331112 podStartE2EDuration="1m21.63331112s" podCreationTimestamp="2025-12-06 05:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:52:55.633012092 +0000 UTC m=+100.521995064" watchObservedRunningTime="2025-12-06 05:52:55.63331112 +0000 UTC m=+100.522294062" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.649646 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=51.649629076 podStartE2EDuration="51.649629076s" podCreationTimestamp="2025-12-06 05:52:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:52:55.648457065 +0000 UTC m=+100.537440007" watchObservedRunningTime="2025-12-06 05:52:55.649629076 +0000 UTC m=+100.538612018" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.661610 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=25.661593807 podStartE2EDuration="25.661593807s" podCreationTimestamp="2025-12-06 05:52:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:52:55.661107495 +0000 UTC m=+100.550090437" watchObservedRunningTime="2025-12-06 05:52:55.661593807 +0000 UTC m=+100.550576749" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.685647 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-rnzq4" podStartSLOduration=81.685630065 podStartE2EDuration="1m21.685630065s" podCreationTimestamp="2025-12-06 05:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:52:55.67355292 +0000 UTC m=+100.562535862" watchObservedRunningTime="2025-12-06 05:52:55.685630065 +0000 UTC m=+100.574613007" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.753595 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.753643 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.753654 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.753671 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.753683 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:55Z","lastTransitionTime":"2025-12-06T05:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.854878 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.854921 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.854947 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.854964 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.854976 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:55Z","lastTransitionTime":"2025-12-06T05:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.957686 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.957966 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.958064 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.958156 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:55 crc kubenswrapper[4809]: I1206 05:52:55.958233 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:55Z","lastTransitionTime":"2025-12-06T05:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.060824 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.060867 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.060877 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.060890 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.060900 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:56Z","lastTransitionTime":"2025-12-06T05:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.162981 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.163007 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.163015 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.163027 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.163037 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:56Z","lastTransitionTime":"2025-12-06T05:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.265835 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.265875 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.265885 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.265900 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.265909 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:56Z","lastTransitionTime":"2025-12-06T05:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.367811 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.367850 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.367882 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.367901 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.367912 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:56Z","lastTransitionTime":"2025-12-06T05:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.388532 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:52:56 crc kubenswrapper[4809]: E1206 05:52:56.388656 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.389384 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:52:56 crc kubenswrapper[4809]: E1206 05:52:56.389507 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.470528 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.470578 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.470591 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.470609 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.470621 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:56Z","lastTransitionTime":"2025-12-06T05:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.573699 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.573761 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.573778 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.573798 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.573813 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:56Z","lastTransitionTime":"2025-12-06T05:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.677048 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.677102 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.677120 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.677145 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.677164 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:56Z","lastTransitionTime":"2025-12-06T05:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.780078 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.780144 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.780157 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.780180 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.780204 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:56Z","lastTransitionTime":"2025-12-06T05:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.882827 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.882864 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.882872 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.882885 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.882895 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:56Z","lastTransitionTime":"2025-12-06T05:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.985044 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.985074 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.985085 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.985102 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:56 crc kubenswrapper[4809]: I1206 05:52:56.985117 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:56Z","lastTransitionTime":"2025-12-06T05:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:57 crc kubenswrapper[4809]: I1206 05:52:57.088298 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:57 crc kubenswrapper[4809]: I1206 05:52:57.088562 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:57 crc kubenswrapper[4809]: I1206 05:52:57.088640 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:57 crc kubenswrapper[4809]: I1206 05:52:57.088709 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:57 crc kubenswrapper[4809]: I1206 05:52:57.088768 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:57Z","lastTransitionTime":"2025-12-06T05:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:57 crc kubenswrapper[4809]: I1206 05:52:57.191193 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:57 crc kubenswrapper[4809]: I1206 05:52:57.191224 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:57 crc kubenswrapper[4809]: I1206 05:52:57.191232 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:57 crc kubenswrapper[4809]: I1206 05:52:57.191245 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:57 crc kubenswrapper[4809]: I1206 05:52:57.191254 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:57Z","lastTransitionTime":"2025-12-06T05:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:57 crc kubenswrapper[4809]: I1206 05:52:57.293784 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:57 crc kubenswrapper[4809]: I1206 05:52:57.293822 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:57 crc kubenswrapper[4809]: I1206 05:52:57.293830 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:57 crc kubenswrapper[4809]: I1206 05:52:57.293841 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:57 crc kubenswrapper[4809]: I1206 05:52:57.293850 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:57Z","lastTransitionTime":"2025-12-06T05:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:57 crc kubenswrapper[4809]: I1206 05:52:57.388212 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:52:57 crc kubenswrapper[4809]: E1206 05:52:57.388324 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:52:57 crc kubenswrapper[4809]: I1206 05:52:57.388386 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:52:57 crc kubenswrapper[4809]: E1206 05:52:57.388548 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:52:57 crc kubenswrapper[4809]: I1206 05:52:57.396665 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:57 crc kubenswrapper[4809]: I1206 05:52:57.397084 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:57 crc kubenswrapper[4809]: I1206 05:52:57.397194 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:57 crc kubenswrapper[4809]: I1206 05:52:57.397294 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:57 crc kubenswrapper[4809]: I1206 05:52:57.397385 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:57Z","lastTransitionTime":"2025-12-06T05:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:57 crc kubenswrapper[4809]: I1206 05:52:57.500410 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:57 crc kubenswrapper[4809]: I1206 05:52:57.500456 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:57 crc kubenswrapper[4809]: I1206 05:52:57.500470 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:57 crc kubenswrapper[4809]: I1206 05:52:57.500487 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:57 crc kubenswrapper[4809]: I1206 05:52:57.500499 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:57Z","lastTransitionTime":"2025-12-06T05:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:57 crc kubenswrapper[4809]: I1206 05:52:57.602806 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:57 crc kubenswrapper[4809]: I1206 05:52:57.603018 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:57 crc kubenswrapper[4809]: I1206 05:52:57.603042 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:57 crc kubenswrapper[4809]: I1206 05:52:57.603066 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:57 crc kubenswrapper[4809]: I1206 05:52:57.603082 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:57Z","lastTransitionTime":"2025-12-06T05:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:57 crc kubenswrapper[4809]: I1206 05:52:57.705355 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:57 crc kubenswrapper[4809]: I1206 05:52:57.705413 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:57 crc kubenswrapper[4809]: I1206 05:52:57.705425 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:57 crc kubenswrapper[4809]: I1206 05:52:57.705440 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:57 crc kubenswrapper[4809]: I1206 05:52:57.705450 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:57Z","lastTransitionTime":"2025-12-06T05:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:57 crc kubenswrapper[4809]: I1206 05:52:57.807803 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:57 crc kubenswrapper[4809]: I1206 05:52:57.807877 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:57 crc kubenswrapper[4809]: I1206 05:52:57.807890 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:57 crc kubenswrapper[4809]: I1206 05:52:57.807908 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:57 crc kubenswrapper[4809]: I1206 05:52:57.807919 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:57Z","lastTransitionTime":"2025-12-06T05:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:57 crc kubenswrapper[4809]: I1206 05:52:57.910177 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:57 crc kubenswrapper[4809]: I1206 05:52:57.910240 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:57 crc kubenswrapper[4809]: I1206 05:52:57.910253 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:57 crc kubenswrapper[4809]: I1206 05:52:57.910271 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:57 crc kubenswrapper[4809]: I1206 05:52:57.910281 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:57Z","lastTransitionTime":"2025-12-06T05:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.012981 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.013053 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.013077 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.013108 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.013130 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:58Z","lastTransitionTime":"2025-12-06T05:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.115114 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.115174 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.115188 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.115204 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.115217 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:58Z","lastTransitionTime":"2025-12-06T05:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.217442 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.217505 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.217515 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.217529 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.217540 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:58Z","lastTransitionTime":"2025-12-06T05:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.322227 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.322304 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.322317 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.322340 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.322353 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:58Z","lastTransitionTime":"2025-12-06T05:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.387535 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.387677 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:52:58 crc kubenswrapper[4809]: E1206 05:52:58.387694 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:52:58 crc kubenswrapper[4809]: E1206 05:52:58.388153 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.425155 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.425199 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.425209 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.425224 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.425235 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:58Z","lastTransitionTime":"2025-12-06T05:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.527692 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.527766 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.527788 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.527818 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.527838 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:58Z","lastTransitionTime":"2025-12-06T05:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.630785 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.630844 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.630862 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.630889 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.630904 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:58Z","lastTransitionTime":"2025-12-06T05:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.734403 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.734466 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.734499 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.734518 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.734543 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:58Z","lastTransitionTime":"2025-12-06T05:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.837891 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.837922 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.837945 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.837958 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.837967 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:58Z","lastTransitionTime":"2025-12-06T05:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.940874 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.941815 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.942044 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.942205 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:58 crc kubenswrapper[4809]: I1206 05:52:58.942358 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:58Z","lastTransitionTime":"2025-12-06T05:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.044846 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.044887 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.044897 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.044956 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.044967 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:59Z","lastTransitionTime":"2025-12-06T05:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.148119 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.148178 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.148195 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.148218 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.148234 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:59Z","lastTransitionTime":"2025-12-06T05:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.250472 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.250517 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.250531 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.250552 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.250566 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:59Z","lastTransitionTime":"2025-12-06T05:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.353728 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.353780 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.353791 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.353811 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.353823 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:59Z","lastTransitionTime":"2025-12-06T05:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.388609 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.388716 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:52:59 crc kubenswrapper[4809]: E1206 05:52:59.388909 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:52:59 crc kubenswrapper[4809]: E1206 05:52:59.389007 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.456767 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.456825 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.456837 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.456856 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.456868 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:59Z","lastTransitionTime":"2025-12-06T05:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.560848 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.560905 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.560917 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.560955 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.560969 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:59Z","lastTransitionTime":"2025-12-06T05:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.663095 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.663136 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.663146 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.663161 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.663172 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:59Z","lastTransitionTime":"2025-12-06T05:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.765164 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.765217 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.765233 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.765252 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.765265 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:59Z","lastTransitionTime":"2025-12-06T05:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.868103 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.868400 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.868540 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.868758 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.869012 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:59Z","lastTransitionTime":"2025-12-06T05:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.971248 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.971901 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.972051 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.972188 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:52:59 crc kubenswrapper[4809]: I1206 05:52:59.972302 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:52:59Z","lastTransitionTime":"2025-12-06T05:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.078670 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.078742 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.078759 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.078778 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.078790 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:53:00Z","lastTransitionTime":"2025-12-06T05:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.181368 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.181429 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.181440 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.181463 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.181477 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:53:00Z","lastTransitionTime":"2025-12-06T05:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.284011 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.284056 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.284065 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.284082 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.284093 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:53:00Z","lastTransitionTime":"2025-12-06T05:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.386300 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.386359 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.386374 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.386394 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.386407 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:53:00Z","lastTransitionTime":"2025-12-06T05:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.387496 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:53:00 crc kubenswrapper[4809]: E1206 05:53:00.387640 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.387669 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:53:00 crc kubenswrapper[4809]: E1206 05:53:00.387799 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.402278 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.489476 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.489525 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.489539 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.489555 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.489566 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:53:00Z","lastTransitionTime":"2025-12-06T05:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.592036 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.592094 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.592104 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.592123 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.592134 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:53:00Z","lastTransitionTime":"2025-12-06T05:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.694773 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.694831 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.694843 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.694861 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.694874 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:53:00Z","lastTransitionTime":"2025-12-06T05:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.798253 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.798309 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.798321 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.798340 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.798351 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:53:00Z","lastTransitionTime":"2025-12-06T05:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.900896 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.901002 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.901024 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.901053 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:53:00 crc kubenswrapper[4809]: I1206 05:53:00.901073 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:53:00Z","lastTransitionTime":"2025-12-06T05:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:53:01 crc kubenswrapper[4809]: I1206 05:53:01.004376 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:53:01 crc kubenswrapper[4809]: I1206 05:53:01.004448 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:53:01 crc kubenswrapper[4809]: I1206 05:53:01.004463 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:53:01 crc kubenswrapper[4809]: I1206 05:53:01.004510 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:53:01 crc kubenswrapper[4809]: I1206 05:53:01.004523 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:53:01Z","lastTransitionTime":"2025-12-06T05:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:53:01 crc kubenswrapper[4809]: I1206 05:53:01.086900 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:53:01 crc kubenswrapper[4809]: I1206 05:53:01.086998 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:53:01 crc kubenswrapper[4809]: I1206 05:53:01.087014 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:53:01 crc kubenswrapper[4809]: I1206 05:53:01.087036 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:53:01 crc kubenswrapper[4809]: I1206 05:53:01.087049 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:53:01Z","lastTransitionTime":"2025-12-06T05:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:53:01 crc kubenswrapper[4809]: I1206 05:53:01.131165 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-xl6b5"] Dec 06 05:53:01 crc kubenswrapper[4809]: I1206 05:53:01.131624 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xl6b5" Dec 06 05:53:01 crc kubenswrapper[4809]: I1206 05:53:01.135905 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 06 05:53:01 crc kubenswrapper[4809]: I1206 05:53:01.136445 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 06 05:53:01 crc kubenswrapper[4809]: I1206 05:53:01.136643 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 06 05:53:01 crc kubenswrapper[4809]: I1206 05:53:01.137403 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 06 05:53:01 crc kubenswrapper[4809]: I1206 05:53:01.163801 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=1.163779721 podStartE2EDuration="1.163779721s" podCreationTimestamp="2025-12-06 05:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:53:01.161987814 +0000 UTC m=+106.050970776" watchObservedRunningTime="2025-12-06 05:53:01.163779721 +0000 UTC m=+106.052762663" Dec 06 05:53:01 crc kubenswrapper[4809]: I1206 05:53:01.215207 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/829123d4-cddd-4d0a-b772-3d98dc7fa522-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-xl6b5\" (UID: \"829123d4-cddd-4d0a-b772-3d98dc7fa522\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xl6b5" Dec 06 05:53:01 crc kubenswrapper[4809]: I1206 05:53:01.215246 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/829123d4-cddd-4d0a-b772-3d98dc7fa522-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-xl6b5\" (UID: \"829123d4-cddd-4d0a-b772-3d98dc7fa522\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xl6b5" Dec 06 05:53:01 crc kubenswrapper[4809]: I1206 05:53:01.215280 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/829123d4-cddd-4d0a-b772-3d98dc7fa522-service-ca\") pod \"cluster-version-operator-5c965bbfc6-xl6b5\" (UID: \"829123d4-cddd-4d0a-b772-3d98dc7fa522\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xl6b5" Dec 06 05:53:01 crc kubenswrapper[4809]: I1206 05:53:01.215311 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/829123d4-cddd-4d0a-b772-3d98dc7fa522-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-xl6b5\" (UID: \"829123d4-cddd-4d0a-b772-3d98dc7fa522\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xl6b5" Dec 06 05:53:01 crc kubenswrapper[4809]: I1206 05:53:01.215352 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/829123d4-cddd-4d0a-b772-3d98dc7fa522-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-xl6b5\" (UID: \"829123d4-cddd-4d0a-b772-3d98dc7fa522\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xl6b5" Dec 06 05:53:01 crc kubenswrapper[4809]: I1206 05:53:01.316704 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/829123d4-cddd-4d0a-b772-3d98dc7fa522-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-xl6b5\" (UID: \"829123d4-cddd-4d0a-b772-3d98dc7fa522\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xl6b5" Dec 06 05:53:01 crc kubenswrapper[4809]: I1206 05:53:01.316783 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/829123d4-cddd-4d0a-b772-3d98dc7fa522-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-xl6b5\" (UID: \"829123d4-cddd-4d0a-b772-3d98dc7fa522\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xl6b5" Dec 06 05:53:01 crc kubenswrapper[4809]: I1206 05:53:01.316858 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/829123d4-cddd-4d0a-b772-3d98dc7fa522-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-xl6b5\" (UID: \"829123d4-cddd-4d0a-b772-3d98dc7fa522\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xl6b5" Dec 06 05:53:01 crc kubenswrapper[4809]: I1206 05:53:01.316924 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/829123d4-cddd-4d0a-b772-3d98dc7fa522-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-xl6b5\" (UID: \"829123d4-cddd-4d0a-b772-3d98dc7fa522\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xl6b5" Dec 06 05:53:01 crc kubenswrapper[4809]: I1206 05:53:01.316979 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/829123d4-cddd-4d0a-b772-3d98dc7fa522-service-ca\") pod \"cluster-version-operator-5c965bbfc6-xl6b5\" (UID: \"829123d4-cddd-4d0a-b772-3d98dc7fa522\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xl6b5" Dec 06 05:53:01 crc kubenswrapper[4809]: I1206 05:53:01.317502 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/829123d4-cddd-4d0a-b772-3d98dc7fa522-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-xl6b5\" (UID: \"829123d4-cddd-4d0a-b772-3d98dc7fa522\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xl6b5" Dec 06 05:53:01 crc kubenswrapper[4809]: I1206 05:53:01.317640 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/829123d4-cddd-4d0a-b772-3d98dc7fa522-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-xl6b5\" (UID: \"829123d4-cddd-4d0a-b772-3d98dc7fa522\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xl6b5" Dec 06 05:53:01 crc kubenswrapper[4809]: I1206 05:53:01.318041 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/829123d4-cddd-4d0a-b772-3d98dc7fa522-service-ca\") pod \"cluster-version-operator-5c965bbfc6-xl6b5\" (UID: \"829123d4-cddd-4d0a-b772-3d98dc7fa522\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xl6b5" Dec 06 05:53:01 crc kubenswrapper[4809]: I1206 05:53:01.326101 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/829123d4-cddd-4d0a-b772-3d98dc7fa522-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-xl6b5\" (UID: \"829123d4-cddd-4d0a-b772-3d98dc7fa522\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xl6b5" Dec 06 05:53:01 crc kubenswrapper[4809]: I1206 05:53:01.337031 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/829123d4-cddd-4d0a-b772-3d98dc7fa522-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-xl6b5\" (UID: \"829123d4-cddd-4d0a-b772-3d98dc7fa522\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xl6b5" Dec 06 05:53:01 crc kubenswrapper[4809]: I1206 05:53:01.387893 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:53:01 crc kubenswrapper[4809]: I1206 05:53:01.388489 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:53:01 crc kubenswrapper[4809]: I1206 05:53:01.388671 4809 scope.go:117] "RemoveContainer" containerID="4a4a7efb6435b711f6fc3d07f0f82258ae97be843a5917fe8017c2afd690c200" Dec 06 05:53:01 crc kubenswrapper[4809]: E1206 05:53:01.388869 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5xvvb_openshift-ovn-kubernetes(ffb3a4f1-1e22-4220-ac80-3b2d69d2db99)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" Dec 06 05:53:01 crc kubenswrapper[4809]: E1206 05:53:01.389041 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:53:01 crc kubenswrapper[4809]: E1206 05:53:01.389106 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:53:01 crc kubenswrapper[4809]: I1206 05:53:01.445729 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xl6b5" Dec 06 05:53:01 crc kubenswrapper[4809]: I1206 05:53:01.960643 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xl6b5" event={"ID":"829123d4-cddd-4d0a-b772-3d98dc7fa522","Type":"ContainerStarted","Data":"0a11f37967872c6d4ad66b341f1dabeed07fe99f838ce0c9eaf0a94c2cf5a58b"} Dec 06 05:53:01 crc kubenswrapper[4809]: I1206 05:53:01.962056 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xl6b5" event={"ID":"829123d4-cddd-4d0a-b772-3d98dc7fa522","Type":"ContainerStarted","Data":"e7465f61409b770a738d5e8e2c7d84f67d585d1748e41ec332f784d098b6d4dc"} Dec 06 05:53:01 crc kubenswrapper[4809]: I1206 05:53:01.974993 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xl6b5" podStartSLOduration=87.97497708 podStartE2EDuration="1m27.97497708s" podCreationTimestamp="2025-12-06 05:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:53:01.974494228 +0000 UTC m=+106.863477170" watchObservedRunningTime="2025-12-06 05:53:01.97497708 +0000 UTC m=+106.863960012" Dec 06 05:53:02 crc kubenswrapper[4809]: I1206 05:53:02.388266 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:53:02 crc kubenswrapper[4809]: E1206 05:53:02.388842 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:53:02 crc kubenswrapper[4809]: I1206 05:53:02.388550 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:53:02 crc kubenswrapper[4809]: E1206 05:53:02.389107 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:53:03 crc kubenswrapper[4809]: I1206 05:53:03.387585 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:53:03 crc kubenswrapper[4809]: E1206 05:53:03.387764 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:53:03 crc kubenswrapper[4809]: I1206 05:53:03.387585 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:53:03 crc kubenswrapper[4809]: E1206 05:53:03.387867 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:53:04 crc kubenswrapper[4809]: I1206 05:53:04.388120 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:53:04 crc kubenswrapper[4809]: I1206 05:53:04.388153 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:53:04 crc kubenswrapper[4809]: E1206 05:53:04.389039 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:53:04 crc kubenswrapper[4809]: E1206 05:53:04.388899 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:53:05 crc kubenswrapper[4809]: I1206 05:53:05.388096 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:53:05 crc kubenswrapper[4809]: I1206 05:53:05.388184 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:53:05 crc kubenswrapper[4809]: E1206 05:53:05.389734 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:53:05 crc kubenswrapper[4809]: E1206 05:53:05.389875 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:53:06 crc kubenswrapper[4809]: I1206 05:53:06.388168 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:53:06 crc kubenswrapper[4809]: I1206 05:53:06.388206 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:53:06 crc kubenswrapper[4809]: E1206 05:53:06.388281 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:53:06 crc kubenswrapper[4809]: E1206 05:53:06.388402 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:53:07 crc kubenswrapper[4809]: I1206 05:53:07.388144 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:53:07 crc kubenswrapper[4809]: I1206 05:53:07.388145 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:53:07 crc kubenswrapper[4809]: E1206 05:53:07.388375 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:53:07 crc kubenswrapper[4809]: E1206 05:53:07.388554 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:53:08 crc kubenswrapper[4809]: I1206 05:53:08.388064 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:53:08 crc kubenswrapper[4809]: I1206 05:53:08.388069 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:53:08 crc kubenswrapper[4809]: E1206 05:53:08.388215 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:53:08 crc kubenswrapper[4809]: E1206 05:53:08.388305 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:53:08 crc kubenswrapper[4809]: I1206 05:53:08.982028 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-cb4vc_ed42d686-41af-470e-910d-a1fabbec66b0/kube-multus/1.log" Dec 06 05:53:08 crc kubenswrapper[4809]: I1206 05:53:08.982992 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-cb4vc_ed42d686-41af-470e-910d-a1fabbec66b0/kube-multus/0.log" Dec 06 05:53:08 crc kubenswrapper[4809]: I1206 05:53:08.983057 4809 generic.go:334] "Generic (PLEG): container finished" podID="ed42d686-41af-470e-910d-a1fabbec66b0" containerID="2edf7fa65761bc560178a40857780a5668e183a4b01974935cdbc3540915114f" exitCode=1 Dec 06 05:53:08 crc kubenswrapper[4809]: I1206 05:53:08.983092 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-cb4vc" event={"ID":"ed42d686-41af-470e-910d-a1fabbec66b0","Type":"ContainerDied","Data":"2edf7fa65761bc560178a40857780a5668e183a4b01974935cdbc3540915114f"} Dec 06 05:53:08 crc kubenswrapper[4809]: I1206 05:53:08.983127 4809 scope.go:117] "RemoveContainer" containerID="73f629b20d9f61227553520b30d0ca046750433926a8d31501e5fb77467426cc" Dec 06 05:53:08 crc kubenswrapper[4809]: I1206 05:53:08.984327 4809 scope.go:117] "RemoveContainer" containerID="2edf7fa65761bc560178a40857780a5668e183a4b01974935cdbc3540915114f" Dec 06 05:53:08 crc kubenswrapper[4809]: E1206 05:53:08.985117 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-cb4vc_openshift-multus(ed42d686-41af-470e-910d-a1fabbec66b0)\"" pod="openshift-multus/multus-cb4vc" podUID="ed42d686-41af-470e-910d-a1fabbec66b0" Dec 06 05:53:09 crc kubenswrapper[4809]: I1206 05:53:09.387647 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:53:09 crc kubenswrapper[4809]: I1206 05:53:09.387793 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:53:09 crc kubenswrapper[4809]: E1206 05:53:09.387844 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:53:09 crc kubenswrapper[4809]: E1206 05:53:09.387964 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:53:09 crc kubenswrapper[4809]: I1206 05:53:09.988766 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-cb4vc_ed42d686-41af-470e-910d-a1fabbec66b0/kube-multus/1.log" Dec 06 05:53:10 crc kubenswrapper[4809]: I1206 05:53:10.388199 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:53:10 crc kubenswrapper[4809]: E1206 05:53:10.388337 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:53:10 crc kubenswrapper[4809]: I1206 05:53:10.388199 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:53:10 crc kubenswrapper[4809]: E1206 05:53:10.388443 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:53:11 crc kubenswrapper[4809]: I1206 05:53:11.388007 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:53:11 crc kubenswrapper[4809]: E1206 05:53:11.388418 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:53:11 crc kubenswrapper[4809]: I1206 05:53:11.388049 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:53:11 crc kubenswrapper[4809]: E1206 05:53:11.388630 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:53:12 crc kubenswrapper[4809]: I1206 05:53:12.387830 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:53:12 crc kubenswrapper[4809]: I1206 05:53:12.387861 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:53:12 crc kubenswrapper[4809]: E1206 05:53:12.388123 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:53:12 crc kubenswrapper[4809]: E1206 05:53:12.388271 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:53:13 crc kubenswrapper[4809]: I1206 05:53:13.387632 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:53:13 crc kubenswrapper[4809]: I1206 05:53:13.387686 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:53:13 crc kubenswrapper[4809]: E1206 05:53:13.387795 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:53:13 crc kubenswrapper[4809]: E1206 05:53:13.387958 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:53:14 crc kubenswrapper[4809]: I1206 05:53:14.388258 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:53:14 crc kubenswrapper[4809]: I1206 05:53:14.388296 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:53:14 crc kubenswrapper[4809]: E1206 05:53:14.389047 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:53:14 crc kubenswrapper[4809]: E1206 05:53:14.389161 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:53:15 crc kubenswrapper[4809]: I1206 05:53:15.388278 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:53:15 crc kubenswrapper[4809]: I1206 05:53:15.388376 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:53:15 crc kubenswrapper[4809]: E1206 05:53:15.389963 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:53:15 crc kubenswrapper[4809]: E1206 05:53:15.390582 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:53:15 crc kubenswrapper[4809]: E1206 05:53:15.410448 4809 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 06 05:53:15 crc kubenswrapper[4809]: E1206 05:53:15.848883 4809 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 06 05:53:16 crc kubenswrapper[4809]: I1206 05:53:16.388515 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:53:16 crc kubenswrapper[4809]: I1206 05:53:16.388530 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:53:16 crc kubenswrapper[4809]: E1206 05:53:16.388635 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:53:16 crc kubenswrapper[4809]: E1206 05:53:16.388852 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:53:16 crc kubenswrapper[4809]: I1206 05:53:16.389618 4809 scope.go:117] "RemoveContainer" containerID="4a4a7efb6435b711f6fc3d07f0f82258ae97be843a5917fe8017c2afd690c200" Dec 06 05:53:16 crc kubenswrapper[4809]: E1206 05:53:16.389797 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5xvvb_openshift-ovn-kubernetes(ffb3a4f1-1e22-4220-ac80-3b2d69d2db99)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" Dec 06 05:53:17 crc kubenswrapper[4809]: I1206 05:53:17.388227 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:53:17 crc kubenswrapper[4809]: I1206 05:53:17.388271 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:53:17 crc kubenswrapper[4809]: E1206 05:53:17.388439 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:53:17 crc kubenswrapper[4809]: E1206 05:53:17.388678 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:53:18 crc kubenswrapper[4809]: I1206 05:53:18.387591 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:53:18 crc kubenswrapper[4809]: I1206 05:53:18.387614 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:53:18 crc kubenswrapper[4809]: E1206 05:53:18.387769 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:53:18 crc kubenswrapper[4809]: E1206 05:53:18.387901 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:53:19 crc kubenswrapper[4809]: I1206 05:53:19.387807 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:53:19 crc kubenswrapper[4809]: I1206 05:53:19.387806 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:53:19 crc kubenswrapper[4809]: E1206 05:53:19.388001 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:53:19 crc kubenswrapper[4809]: E1206 05:53:19.388037 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:53:20 crc kubenswrapper[4809]: I1206 05:53:20.388104 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:53:20 crc kubenswrapper[4809]: I1206 05:53:20.388151 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:53:20 crc kubenswrapper[4809]: E1206 05:53:20.388295 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:53:20 crc kubenswrapper[4809]: E1206 05:53:20.388460 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:53:20 crc kubenswrapper[4809]: E1206 05:53:20.850199 4809 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 06 05:53:21 crc kubenswrapper[4809]: I1206 05:53:21.388220 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:53:21 crc kubenswrapper[4809]: E1206 05:53:21.388341 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:53:21 crc kubenswrapper[4809]: I1206 05:53:21.388220 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:53:21 crc kubenswrapper[4809]: E1206 05:53:21.388471 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:53:22 crc kubenswrapper[4809]: I1206 05:53:22.388149 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:53:22 crc kubenswrapper[4809]: I1206 05:53:22.388305 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:53:22 crc kubenswrapper[4809]: E1206 05:53:22.388465 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:53:22 crc kubenswrapper[4809]: E1206 05:53:22.388703 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:53:22 crc kubenswrapper[4809]: I1206 05:53:22.389072 4809 scope.go:117] "RemoveContainer" containerID="2edf7fa65761bc560178a40857780a5668e183a4b01974935cdbc3540915114f" Dec 06 05:53:23 crc kubenswrapper[4809]: I1206 05:53:23.031649 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-cb4vc_ed42d686-41af-470e-910d-a1fabbec66b0/kube-multus/1.log" Dec 06 05:53:23 crc kubenswrapper[4809]: I1206 05:53:23.032149 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-cb4vc" event={"ID":"ed42d686-41af-470e-910d-a1fabbec66b0","Type":"ContainerStarted","Data":"b950762e7c636b40a9c1701711f6bee3048580e99b743938cbcb59c938281b24"} Dec 06 05:53:23 crc kubenswrapper[4809]: I1206 05:53:23.387878 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:53:23 crc kubenswrapper[4809]: E1206 05:53:23.388043 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:53:23 crc kubenswrapper[4809]: I1206 05:53:23.388138 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:53:23 crc kubenswrapper[4809]: E1206 05:53:23.388249 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:53:24 crc kubenswrapper[4809]: I1206 05:53:24.387710 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:53:24 crc kubenswrapper[4809]: I1206 05:53:24.387710 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:53:24 crc kubenswrapper[4809]: E1206 05:53:24.387895 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:53:24 crc kubenswrapper[4809]: E1206 05:53:24.388021 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:53:25 crc kubenswrapper[4809]: I1206 05:53:25.388450 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:53:25 crc kubenswrapper[4809]: I1206 05:53:25.388484 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:53:25 crc kubenswrapper[4809]: E1206 05:53:25.389765 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:53:25 crc kubenswrapper[4809]: E1206 05:53:25.389961 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:53:25 crc kubenswrapper[4809]: E1206 05:53:25.851130 4809 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 06 05:53:26 crc kubenswrapper[4809]: I1206 05:53:26.387440 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:53:26 crc kubenswrapper[4809]: I1206 05:53:26.387520 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:53:26 crc kubenswrapper[4809]: E1206 05:53:26.387576 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:53:26 crc kubenswrapper[4809]: E1206 05:53:26.387692 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:53:27 crc kubenswrapper[4809]: I1206 05:53:27.388636 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:53:27 crc kubenswrapper[4809]: I1206 05:53:27.388750 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:53:27 crc kubenswrapper[4809]: E1206 05:53:27.388821 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:53:27 crc kubenswrapper[4809]: E1206 05:53:27.388979 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:53:28 crc kubenswrapper[4809]: I1206 05:53:28.388465 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:53:28 crc kubenswrapper[4809]: E1206 05:53:28.388691 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:53:28 crc kubenswrapper[4809]: I1206 05:53:28.389083 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:53:28 crc kubenswrapper[4809]: E1206 05:53:28.389233 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:53:29 crc kubenswrapper[4809]: I1206 05:53:29.388651 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:53:29 crc kubenswrapper[4809]: I1206 05:53:29.388731 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:53:29 crc kubenswrapper[4809]: E1206 05:53:29.388822 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:53:29 crc kubenswrapper[4809]: E1206 05:53:29.389219 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:53:30 crc kubenswrapper[4809]: I1206 05:53:30.388119 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:53:30 crc kubenswrapper[4809]: E1206 05:53:30.388276 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:53:30 crc kubenswrapper[4809]: I1206 05:53:30.388151 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:53:30 crc kubenswrapper[4809]: E1206 05:53:30.388455 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:53:30 crc kubenswrapper[4809]: E1206 05:53:30.851890 4809 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 06 05:53:31 crc kubenswrapper[4809]: I1206 05:53:31.388039 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:53:31 crc kubenswrapper[4809]: I1206 05:53:31.387900 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:53:31 crc kubenswrapper[4809]: E1206 05:53:31.388580 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:53:31 crc kubenswrapper[4809]: I1206 05:53:31.388750 4809 scope.go:117] "RemoveContainer" containerID="4a4a7efb6435b711f6fc3d07f0f82258ae97be843a5917fe8017c2afd690c200" Dec 06 05:53:31 crc kubenswrapper[4809]: E1206 05:53:31.388804 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:53:32 crc kubenswrapper[4809]: I1206 05:53:32.388133 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:53:32 crc kubenswrapper[4809]: I1206 05:53:32.388172 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:53:32 crc kubenswrapper[4809]: E1206 05:53:32.388400 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:53:32 crc kubenswrapper[4809]: E1206 05:53:32.388565 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:53:33 crc kubenswrapper[4809]: I1206 05:53:33.066091 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5xvvb_ffb3a4f1-1e22-4220-ac80-3b2d69d2db99/ovnkube-controller/3.log" Dec 06 05:53:33 crc kubenswrapper[4809]: I1206 05:53:33.069198 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" event={"ID":"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99","Type":"ContainerStarted","Data":"1cf68aa264056dd94a3e6687d35def2e9eb5580a4ee72a3acd46a061f4d06437"} Dec 06 05:53:33 crc kubenswrapper[4809]: I1206 05:53:33.070350 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:53:33 crc kubenswrapper[4809]: I1206 05:53:33.387821 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:53:33 crc kubenswrapper[4809]: E1206 05:53:33.387961 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:53:33 crc kubenswrapper[4809]: I1206 05:53:33.387819 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:53:33 crc kubenswrapper[4809]: E1206 05:53:33.388100 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:53:33 crc kubenswrapper[4809]: I1206 05:53:33.923639 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" podStartSLOduration=119.923616211 podStartE2EDuration="1m59.923616211s" podCreationTimestamp="2025-12-06 05:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:53:33.09498384 +0000 UTC m=+137.983966792" watchObservedRunningTime="2025-12-06 05:53:33.923616211 +0000 UTC m=+138.812599183" Dec 06 05:53:33 crc kubenswrapper[4809]: I1206 05:53:33.924430 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-9k8zx"] Dec 06 05:53:33 crc kubenswrapper[4809]: I1206 05:53:33.924567 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:53:33 crc kubenswrapper[4809]: E1206 05:53:33.924746 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:53:34 crc kubenswrapper[4809]: I1206 05:53:34.387801 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:53:34 crc kubenswrapper[4809]: E1206 05:53:34.388074 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:53:35 crc kubenswrapper[4809]: I1206 05:53:35.387556 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:53:35 crc kubenswrapper[4809]: E1206 05:53:35.387694 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:53:35 crc kubenswrapper[4809]: I1206 05:53:35.388786 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:53:35 crc kubenswrapper[4809]: E1206 05:53:35.388834 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:53:35 crc kubenswrapper[4809]: I1206 05:53:35.389049 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:53:35 crc kubenswrapper[4809]: E1206 05:53:35.389094 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:53:35 crc kubenswrapper[4809]: E1206 05:53:35.852486 4809 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 06 05:53:36 crc kubenswrapper[4809]: I1206 05:53:36.388134 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:53:36 crc kubenswrapper[4809]: E1206 05:53:36.388336 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:53:37 crc kubenswrapper[4809]: I1206 05:53:37.388142 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:53:37 crc kubenswrapper[4809]: I1206 05:53:37.388157 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:53:37 crc kubenswrapper[4809]: I1206 05:53:37.388164 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:53:37 crc kubenswrapper[4809]: E1206 05:53:37.388444 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:53:37 crc kubenswrapper[4809]: E1206 05:53:37.388466 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:53:37 crc kubenswrapper[4809]: E1206 05:53:37.388303 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:53:38 crc kubenswrapper[4809]: I1206 05:53:38.388029 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:53:38 crc kubenswrapper[4809]: E1206 05:53:38.388181 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:53:39 crc kubenswrapper[4809]: I1206 05:53:39.388587 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:53:39 crc kubenswrapper[4809]: E1206 05:53:39.388711 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:53:39 crc kubenswrapper[4809]: I1206 05:53:39.388747 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:53:39 crc kubenswrapper[4809]: I1206 05:53:39.388806 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:53:39 crc kubenswrapper[4809]: E1206 05:53:39.388897 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:53:39 crc kubenswrapper[4809]: E1206 05:53:39.389029 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9k8zx" podUID="a4f75fd3-03e8-41e7-8926-f6ff21a5b681" Dec 06 05:53:40 crc kubenswrapper[4809]: I1206 05:53:40.387812 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:53:40 crc kubenswrapper[4809]: E1206 05:53:40.388033 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.388195 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.388217 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.388292 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.391144 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.391144 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.391364 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.392713 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.855466 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.901090 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sdg4b"] Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.901589 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sdg4b" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.901775 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-4pwcw"] Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.902166 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-4pwcw" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.903624 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ggvcl"] Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.903922 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-ggvcl" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.905079 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kwzrc"] Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.905414 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.909707 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.909885 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.910018 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.910171 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.910264 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.910495 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.910572 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.910645 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.910720 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.910828 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.910948 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.911044 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.911138 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.912362 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.912426 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.912366 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-kjlr8"] Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.912574 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.913036 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kjlr8" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.918341 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.918642 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.919617 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.922288 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.922445 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r7684"] Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.922814 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9rgdg"] Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.923115 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.923202 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.923229 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.923253 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r7684" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.923364 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.923616 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.923837 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9rgdg" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.937072 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-wcth2"] Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.937674 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wcth2" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.939404 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.939942 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.939945 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.940200 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dpbfw"] Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.940687 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dpbfw" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.941323 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-znmpz"] Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.941863 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-znmpz" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.943870 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.944330 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.944355 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.944474 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.944491 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.944507 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.946060 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-nsqq5"] Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.946482 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.946582 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.946650 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nsqq5" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.949632 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-m6kjl"] Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.950207 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7sxnv"] Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.950346 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-m6kjl" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.950506 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7sxnv" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.954415 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.954653 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.954812 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.955026 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.955786 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.956017 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.956214 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.958252 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.958432 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.958449 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.958604 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.972507 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.977830 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-drj7v"] Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.977853 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.978621 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.978691 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.978715 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.978890 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.980428 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.980576 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.980649 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-drj7v" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.980731 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.980576 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kr48t"] Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.981876 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kr48t" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.983043 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-7qpjs"] Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.983460 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.983638 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-tphf5"] Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.983653 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.983765 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-7qpjs" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.984720 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-tphf5" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.985048 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.986135 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.986798 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.986981 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.987086 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.987287 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.987923 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.988026 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-n6z57"] Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.988616 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-n6z57" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.989269 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/49fef68f-df3c-4bb7-80cf-f0f651797762-client-ca\") pod \"controller-manager-879f6c89f-ggvcl\" (UID: \"49fef68f-df3c-4bb7-80cf-f0f651797762\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ggvcl" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.989310 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glkvh\" (UniqueName: \"kubernetes.io/projected/8e03364a-1840-4c91-8b2b-c7e2071fc73b-kube-api-access-glkvh\") pod \"console-f9d7485db-4pwcw\" (UID: \"8e03364a-1840-4c91-8b2b-c7e2071fc73b\") " pod="openshift-console/console-f9d7485db-4pwcw" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.989318 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.989338 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c8703a5f-60fb-40f5-be0c-d1025852c05a-ca-trust-extracted\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.989391 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zb9vt\" (UniqueName: \"kubernetes.io/projected/49fef68f-df3c-4bb7-80cf-f0f651797762-kube-api-access-zb9vt\") pod \"controller-manager-879f6c89f-ggvcl\" (UID: \"49fef68f-df3c-4bb7-80cf-f0f651797762\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ggvcl" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.989415 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d30c4960-b4da-41cf-9d1c-771e0bc7b262-trusted-ca\") pod \"ingress-operator-5b745b69d9-kjlr8\" (UID: \"d30c4960-b4da-41cf-9d1c-771e0bc7b262\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kjlr8" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.989438 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d30c4960-b4da-41cf-9d1c-771e0bc7b262-bound-sa-token\") pod \"ingress-operator-5b745b69d9-kjlr8\" (UID: \"d30c4960-b4da-41cf-9d1c-771e0bc7b262\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kjlr8" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.989474 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49fef68f-df3c-4bb7-80cf-f0f651797762-config\") pod \"controller-manager-879f6c89f-ggvcl\" (UID: \"49fef68f-df3c-4bb7-80cf-f0f651797762\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ggvcl" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.989496 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8e03364a-1840-4c91-8b2b-c7e2071fc73b-console-config\") pod \"console-f9d7485db-4pwcw\" (UID: \"8e03364a-1840-4c91-8b2b-c7e2071fc73b\") " pod="openshift-console/console-f9d7485db-4pwcw" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.989516 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e03364a-1840-4c91-8b2b-c7e2071fc73b-trusted-ca-bundle\") pod \"console-f9d7485db-4pwcw\" (UID: \"8e03364a-1840-4c91-8b2b-c7e2071fc73b\") " pod="openshift-console/console-f9d7485db-4pwcw" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.989537 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c8703a5f-60fb-40f5-be0c-d1025852c05a-registry-certificates\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.989570 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c8703a5f-60fb-40f5-be0c-d1025852c05a-trusted-ca\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.989593 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2885220-fafe-48da-9ed7-98942886ab1f-config\") pod \"kube-controller-manager-operator-78b949d7b-9rgdg\" (UID: \"e2885220-fafe-48da-9ed7-98942886ab1f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9rgdg" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.989617 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c8703a5f-60fb-40f5-be0c-d1025852c05a-installation-pull-secrets\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.989641 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/49fef68f-df3c-4bb7-80cf-f0f651797762-serving-cert\") pod \"controller-manager-879f6c89f-ggvcl\" (UID: \"49fef68f-df3c-4bb7-80cf-f0f651797762\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ggvcl" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.989669 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-nff6m"] Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.989688 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8e03364a-1840-4c91-8b2b-c7e2071fc73b-oauth-serving-cert\") pod \"console-f9d7485db-4pwcw\" (UID: \"8e03364a-1840-4c91-8b2b-c7e2071fc73b\") " pod="openshift-console/console-f9d7485db-4pwcw" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.989719 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e2885220-fafe-48da-9ed7-98942886ab1f-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-9rgdg\" (UID: \"e2885220-fafe-48da-9ed7-98942886ab1f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9rgdg" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.989804 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c8703a5f-60fb-40f5-be0c-d1025852c05a-registry-tls\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.989828 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/49fef68f-df3c-4bb7-80cf-f0f651797762-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-ggvcl\" (UID: \"49fef68f-df3c-4bb7-80cf-f0f651797762\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ggvcl" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.989858 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4j955\" (UniqueName: \"kubernetes.io/projected/d30c4960-b4da-41cf-9d1c-771e0bc7b262-kube-api-access-4j955\") pod \"ingress-operator-5b745b69d9-kjlr8\" (UID: \"d30c4960-b4da-41cf-9d1c-771e0bc7b262\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kjlr8" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.989879 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8e03364a-1840-4c91-8b2b-c7e2071fc73b-console-serving-cert\") pod \"console-f9d7485db-4pwcw\" (UID: \"8e03364a-1840-4c91-8b2b-c7e2071fc73b\") " pod="openshift-console/console-f9d7485db-4pwcw" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.989926 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1718efb3-c63c-4aac-bc59-2cc85a59ba61-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-sdg4b\" (UID: \"1718efb3-c63c-4aac-bc59-2cc85a59ba61\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sdg4b" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.989959 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fjqp\" (UniqueName: \"kubernetes.io/projected/1718efb3-c63c-4aac-bc59-2cc85a59ba61-kube-api-access-6fjqp\") pod \"openshift-apiserver-operator-796bbdcf4f-sdg4b\" (UID: \"1718efb3-c63c-4aac-bc59-2cc85a59ba61\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sdg4b" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.989974 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e2885220-fafe-48da-9ed7-98942886ab1f-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-9rgdg\" (UID: \"e2885220-fafe-48da-9ed7-98942886ab1f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9rgdg" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.989994 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8e03364a-1840-4c91-8b2b-c7e2071fc73b-console-oauth-config\") pod \"console-f9d7485db-4pwcw\" (UID: \"8e03364a-1840-4c91-8b2b-c7e2071fc73b\") " pod="openshift-console/console-f9d7485db-4pwcw" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.990011 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrw5q\" (UniqueName: \"kubernetes.io/projected/cf124a04-e72c-4c67-bd76-52648b4d20c9-kube-api-access-mrw5q\") pod \"kube-storage-version-migrator-operator-b67b599dd-r7684\" (UID: \"cf124a04-e72c-4c67-bd76-52648b4d20c9\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r7684" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.990027 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d30c4960-b4da-41cf-9d1c-771e0bc7b262-metrics-tls\") pod \"ingress-operator-5b745b69d9-kjlr8\" (UID: \"d30c4960-b4da-41cf-9d1c-771e0bc7b262\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kjlr8" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.990063 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c8703a5f-60fb-40f5-be0c-d1025852c05a-bound-sa-token\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.990079 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8e03364a-1840-4c91-8b2b-c7e2071fc73b-service-ca\") pod \"console-f9d7485db-4pwcw\" (UID: \"8e03364a-1840-4c91-8b2b-c7e2071fc73b\") " pod="openshift-console/console-f9d7485db-4pwcw" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.990102 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.990117 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1718efb3-c63c-4aac-bc59-2cc85a59ba61-config\") pod \"openshift-apiserver-operator-796bbdcf4f-sdg4b\" (UID: \"1718efb3-c63c-4aac-bc59-2cc85a59ba61\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sdg4b" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.990131 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cf124a04-e72c-4c67-bd76-52648b4d20c9-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-r7684\" (UID: \"cf124a04-e72c-4c67-bd76-52648b4d20c9\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r7684" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.990207 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55jls\" (UniqueName: \"kubernetes.io/projected/c8703a5f-60fb-40f5-be0c-d1025852c05a-kube-api-access-55jls\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.990223 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf124a04-e72c-4c67-bd76-52648b4d20c9-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-r7684\" (UID: \"cf124a04-e72c-4c67-bd76-52648b4d20c9\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r7684" Dec 06 05:53:41 crc kubenswrapper[4809]: E1206 05:53:41.990404 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:42.49039481 +0000 UTC m=+147.379377752 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.991598 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-927z2"] Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.991815 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nff6m" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.993097 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.993107 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.993200 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.993298 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.993588 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.993808 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.993876 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.994004 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.994076 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.994749 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.994886 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.995899 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.997652 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wqnkb"] Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.997838 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-927z2" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.998087 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wqnkb" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.998487 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.998628 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-cd2tv"] Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.999094 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 06 05:53:41 crc kubenswrapper[4809]: I1206 05:53:41.999191 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cd2tv" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.001002 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gt9kb"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.001596 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gt9kb" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.001998 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ch964"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.002514 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ch964" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.003421 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-qqv6x"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.004008 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qqv6x" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.012377 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.012505 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.012601 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.012726 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.012889 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.013272 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.013534 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.013550 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.014204 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.014233 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.014382 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.014549 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.014644 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.014744 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.014782 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.014855 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.014896 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.014952 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.015026 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sdg4b"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.015233 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.015390 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.018401 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-c8sn6"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.018973 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-s7q7v"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.019613 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-slt9h"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.019690 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-c8sn6" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.019643 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-s7q7v" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.021049 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-slt9h" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.022312 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416665-lsfkp"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.027591 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416665-lsfkp" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.036568 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-4pwcw"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.036625 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pkq7h"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.037899 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jjf9r"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.039024 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.040781 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jjf9r" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.043520 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ggvcl"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.044870 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.045908 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pkq7h" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.046549 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.049200 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-hlvjb"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.051163 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-hlvjb" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.052484 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-g7pfc"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.053135 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-g7pfc" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.055036 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jn446"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.057548 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.058739 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-8ssrl"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.059620 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8ssrl" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.062077 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-224mv"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.065627 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-852bd"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.065837 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-224mv" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.066637 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-mxv9r"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.066845 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-852bd" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.067083 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-kjlr8"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.067111 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zj6j9"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.067310 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.067537 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dpbfw"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.067606 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-mxv9r" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.067672 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zj6j9" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.068017 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-wcth2"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.069018 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-m6kjl"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.070754 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kwzrc"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.070778 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gt9kb"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.072033 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-drj7v"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.074676 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-7qpjs"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.074720 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ch964"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.075436 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-927z2"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.076507 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7sxnv"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.077638 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9rgdg"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.078469 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wqnkb"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.079821 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-8ssrl"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.081484 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pkq7h"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.082307 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kr48t"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.083692 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.083898 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-nsqq5"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.084496 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-hlvjb"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.087653 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416665-lsfkp"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.088867 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-s7q7v"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.090591 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-qqv6x"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.090981 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.091153 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d30c4960-b4da-41cf-9d1c-771e0bc7b262-bound-sa-token\") pod \"ingress-operator-5b745b69d9-kjlr8\" (UID: \"d30c4960-b4da-41cf-9d1c-771e0bc7b262\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kjlr8" Dec 06 05:53:42 crc kubenswrapper[4809]: E1206 05:53:42.091201 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:42.591174647 +0000 UTC m=+147.480157589 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.091239 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wh944\" (UniqueName: \"kubernetes.io/projected/e5356f1d-5afd-47bd-89af-72abfa8f901f-kube-api-access-wh944\") pod \"router-default-5444994796-n6z57\" (UID: \"e5356f1d-5afd-47bd-89af-72abfa8f901f\") " pod="openshift-ingress/router-default-5444994796-n6z57" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.091352 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8e03364a-1840-4c91-8b2b-c7e2071fc73b-console-config\") pod \"console-f9d7485db-4pwcw\" (UID: \"8e03364a-1840-4c91-8b2b-c7e2071fc73b\") " pod="openshift-console/console-f9d7485db-4pwcw" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.091385 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e03364a-1840-4c91-8b2b-c7e2071fc73b-trusted-ca-bundle\") pod \"console-f9d7485db-4pwcw\" (UID: \"8e03364a-1840-4c91-8b2b-c7e2071fc73b\") " pod="openshift-console/console-f9d7485db-4pwcw" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.091410 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49fef68f-df3c-4bb7-80cf-f0f651797762-config\") pod \"controller-manager-879f6c89f-ggvcl\" (UID: \"49fef68f-df3c-4bb7-80cf-f0f651797762\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ggvcl" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.091435 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/e5356f1d-5afd-47bd-89af-72abfa8f901f-stats-auth\") pod \"router-default-5444994796-n6z57\" (UID: \"e5356f1d-5afd-47bd-89af-72abfa8f901f\") " pod="openshift-ingress/router-default-5444994796-n6z57" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.091459 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/63b0cea8-1716-4c17-9f87-f10e776cf47f-etcd-ca\") pod \"etcd-operator-b45778765-drj7v\" (UID: \"63b0cea8-1716-4c17-9f87-f10e776cf47f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-drj7v" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.091496 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c8703a5f-60fb-40f5-be0c-d1025852c05a-registry-certificates\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.091527 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhvfg\" (UniqueName: \"kubernetes.io/projected/cdf2d3c6-659f-4dc5-a21c-d29cc86ef239-kube-api-access-vhvfg\") pod \"machine-config-controller-84d6567774-wcth2\" (UID: \"cdf2d3c6-659f-4dc5-a21c-d29cc86ef239\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wcth2" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.091554 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f3a30985-f268-44de-93b9-bce0f9aebc6d-serving-cert\") pod \"console-operator-58897d9998-hlvjb\" (UID: \"f3a30985-f268-44de-93b9-bce0f9aebc6d\") " pod="openshift-console-operator/console-operator-58897d9998-hlvjb" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.091577 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d7d02f2c-c052-4320-bac4-394fb40c57d2-metrics-tls\") pod \"dns-operator-744455d44c-927z2\" (UID: \"d7d02f2c-c052-4320-bac4-394fb40c57d2\") " pod="openshift-dns-operator/dns-operator-744455d44c-927z2" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.091598 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f4008f48-e17b-492f-8fa4-b09fb9d2ce70-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-s7q7v\" (UID: \"f4008f48-e17b-492f-8fa4-b09fb9d2ce70\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-s7q7v" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.091627 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c8703a5f-60fb-40f5-be0c-d1025852c05a-trusted-ca\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.091651 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2885220-fafe-48da-9ed7-98942886ab1f-config\") pod \"kube-controller-manager-operator-78b949d7b-9rgdg\" (UID: \"e2885220-fafe-48da-9ed7-98942886ab1f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9rgdg" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.091674 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjbbp\" (UniqueName: \"kubernetes.io/projected/00bf9f05-7bc3-40e1-a2e9-1af1bf93f014-kube-api-access-pjbbp\") pod \"catalog-operator-68c6474976-pkq7h\" (UID: \"00bf9f05-7bc3-40e1-a2e9-1af1bf93f014\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pkq7h" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.091698 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2a1ed741-489c-455f-a344-404bbab66c22-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-7qpjs\" (UID: \"2a1ed741-489c-455f-a344-404bbab66c22\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7qpjs" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.091723 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8e03364a-1840-4c91-8b2b-c7e2071fc73b-oauth-serving-cert\") pod \"console-f9d7485db-4pwcw\" (UID: \"8e03364a-1840-4c91-8b2b-c7e2071fc73b\") " pod="openshift-console/console-f9d7485db-4pwcw" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.091751 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c8703a5f-60fb-40f5-be0c-d1025852c05a-installation-pull-secrets\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.091817 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/49fef68f-df3c-4bb7-80cf-f0f651797762-serving-cert\") pod \"controller-manager-879f6c89f-ggvcl\" (UID: \"49fef68f-df3c-4bb7-80cf-f0f651797762\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ggvcl" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.091848 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a225b5c-4ad5-4f08-a0cb-ae38a8fa83d4-serving-cert\") pod \"openshift-config-operator-7777fb866f-qqv6x\" (UID: \"8a225b5c-4ad5-4f08-a0cb-ae38a8fa83d4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-qqv6x" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.091870 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2a1ed741-489c-455f-a344-404bbab66c22-service-ca-bundle\") pod \"authentication-operator-69f744f599-7qpjs\" (UID: \"2a1ed741-489c-455f-a344-404bbab66c22\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7qpjs" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.091918 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e2885220-fafe-48da-9ed7-98942886ab1f-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-9rgdg\" (UID: \"e2885220-fafe-48da-9ed7-98942886ab1f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9rgdg" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.091959 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/63b0cea8-1716-4c17-9f87-f10e776cf47f-serving-cert\") pod \"etcd-operator-b45778765-drj7v\" (UID: \"63b0cea8-1716-4c17-9f87-f10e776cf47f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-drj7v" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.091986 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d844d778-6b3d-400a-8f71-db1c2e643878-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-zj6j9\" (UID: \"d844d778-6b3d-400a-8f71-db1c2e643878\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zj6j9" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.092010 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vfkx\" (UniqueName: \"kubernetes.io/projected/d844d778-6b3d-400a-8f71-db1c2e643878-kube-api-access-9vfkx\") pod \"openshift-controller-manager-operator-756b6f6bc6-zj6j9\" (UID: \"d844d778-6b3d-400a-8f71-db1c2e643878\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zj6j9" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.092053 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c8703a5f-60fb-40f5-be0c-d1025852c05a-registry-tls\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.092080 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/49fef68f-df3c-4bb7-80cf-f0f651797762-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-ggvcl\" (UID: \"49fef68f-df3c-4bb7-80cf-f0f651797762\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ggvcl" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.092144 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6nl2\" (UniqueName: \"kubernetes.io/projected/de0bb137-3c1d-4632-a7ef-6922f027a3c1-kube-api-access-v6nl2\") pod \"control-plane-machine-set-operator-78cbb6b69f-dpbfw\" (UID: \"de0bb137-3c1d-4632-a7ef-6922f027a3c1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dpbfw" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.092204 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bplj9\" (UniqueName: \"kubernetes.io/projected/c8d79e6f-40b0-4ecc-8501-454ba5cfe691-kube-api-access-bplj9\") pod \"olm-operator-6b444d44fb-gt9kb\" (UID: \"c8d79e6f-40b0-4ecc-8501-454ba5cfe691\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gt9kb" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.092230 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wfkz\" (UniqueName: \"kubernetes.io/projected/e5e4e451-e3ee-434e-8e24-0c795925a48e-kube-api-access-9wfkz\") pod \"packageserver-d55dfcdfc-jjf9r\" (UID: \"e5e4e451-e3ee-434e-8e24-0c795925a48e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jjf9r" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.092266 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25sjn\" (UniqueName: \"kubernetes.io/projected/cea05156-8322-404a-8b33-867aeaaa905e-kube-api-access-25sjn\") pod \"service-ca-operator-777779d784-mxv9r\" (UID: \"cea05156-8322-404a-8b33-867aeaaa905e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mxv9r" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.092292 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9tmj\" (UniqueName: \"kubernetes.io/projected/8a225b5c-4ad5-4f08-a0cb-ae38a8fa83d4-kube-api-access-v9tmj\") pod \"openshift-config-operator-7777fb866f-qqv6x\" (UID: \"8a225b5c-4ad5-4f08-a0cb-ae38a8fa83d4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-qqv6x" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.092315 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d844d778-6b3d-400a-8f71-db1c2e643878-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-zj6j9\" (UID: \"d844d778-6b3d-400a-8f71-db1c2e643878\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zj6j9" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.092345 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4j955\" (UniqueName: \"kubernetes.io/projected/d30c4960-b4da-41cf-9d1c-771e0bc7b262-kube-api-access-4j955\") pod \"ingress-operator-5b745b69d9-kjlr8\" (UID: \"d30c4960-b4da-41cf-9d1c-771e0bc7b262\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kjlr8" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.092407 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/de0bb137-3c1d-4632-a7ef-6922f027a3c1-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-dpbfw\" (UID: \"de0bb137-3c1d-4632-a7ef-6922f027a3c1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dpbfw" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.092436 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcpcz\" (UniqueName: \"kubernetes.io/projected/f4008f48-e17b-492f-8fa4-b09fb9d2ce70-kube-api-access-zcpcz\") pod \"multus-admission-controller-857f4d67dd-s7q7v\" (UID: \"f4008f48-e17b-492f-8fa4-b09fb9d2ce70\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-s7q7v" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.092467 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d0b4b35c-3cb0-405d-941b-1b3946b228a0-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-kr48t\" (UID: \"d0b4b35c-3cb0-405d-941b-1b3946b228a0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kr48t" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.092499 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8e03364a-1840-4c91-8b2b-c7e2071fc73b-console-serving-cert\") pod \"console-f9d7485db-4pwcw\" (UID: \"8e03364a-1840-4c91-8b2b-c7e2071fc73b\") " pod="openshift-console/console-f9d7485db-4pwcw" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.092523 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ls6xx\" (UniqueName: \"kubernetes.io/projected/26b87745-a15b-4f2d-b98e-b24afc5d512b-kube-api-access-ls6xx\") pod \"service-ca-9c57cc56f-c8sn6\" (UID: \"26b87745-a15b-4f2d-b98e-b24afc5d512b\") " pod="openshift-service-ca/service-ca-9c57cc56f-c8sn6" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.092532 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8e03364a-1840-4c91-8b2b-c7e2071fc73b-oauth-serving-cert\") pod \"console-f9d7485db-4pwcw\" (UID: \"8e03364a-1840-4c91-8b2b-c7e2071fc73b\") " pod="openshift-console/console-f9d7485db-4pwcw" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.092547 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/00bf9f05-7bc3-40e1-a2e9-1af1bf93f014-srv-cert\") pod \"catalog-operator-68c6474976-pkq7h\" (UID: \"00bf9f05-7bc3-40e1-a2e9-1af1bf93f014\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pkq7h" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.092568 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63b0cea8-1716-4c17-9f87-f10e776cf47f-config\") pod \"etcd-operator-b45778765-drj7v\" (UID: \"63b0cea8-1716-4c17-9f87-f10e776cf47f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-drj7v" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.092593 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2kwh\" (UniqueName: \"kubernetes.io/projected/2a1ed741-489c-455f-a344-404bbab66c22-kube-api-access-t2kwh\") pod \"authentication-operator-69f744f599-7qpjs\" (UID: \"2a1ed741-489c-455f-a344-404bbab66c22\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7qpjs" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.092614 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a1ed741-489c-455f-a344-404bbab66c22-config\") pod \"authentication-operator-69f744f599-7qpjs\" (UID: \"2a1ed741-489c-455f-a344-404bbab66c22\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7qpjs" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.092638 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rj9g\" (UniqueName: \"kubernetes.io/projected/63b0cea8-1716-4c17-9f87-f10e776cf47f-kube-api-access-6rj9g\") pod \"etcd-operator-b45778765-drj7v\" (UID: \"63b0cea8-1716-4c17-9f87-f10e776cf47f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-drj7v" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.092662 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8a225b5c-4ad5-4f08-a0cb-ae38a8fa83d4-available-featuregates\") pod \"openshift-config-operator-7777fb866f-qqv6x\" (UID: \"8a225b5c-4ad5-4f08-a0cb-ae38a8fa83d4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-qqv6x" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.092682 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/63b0cea8-1716-4c17-9f87-f10e776cf47f-etcd-service-ca\") pod \"etcd-operator-b45778765-drj7v\" (UID: \"63b0cea8-1716-4c17-9f87-f10e776cf47f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-drj7v" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.092710 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnqk2\" (UniqueName: \"kubernetes.io/projected/d233521e-75db-4da3-9cd3-0480a99eba90-kube-api-access-hnqk2\") pod \"migrator-59844c95c7-8ssrl\" (UID: \"d233521e-75db-4da3-9cd3-0480a99eba90\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8ssrl" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.092732 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cea05156-8322-404a-8b33-867aeaaa905e-config\") pod \"service-ca-operator-777779d784-mxv9r\" (UID: \"cea05156-8322-404a-8b33-867aeaaa905e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mxv9r" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.092753 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c8d79e6f-40b0-4ecc-8501-454ba5cfe691-profile-collector-cert\") pod \"olm-operator-6b444d44fb-gt9kb\" (UID: \"c8d79e6f-40b0-4ecc-8501-454ba5cfe691\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gt9kb" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.092777 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e5e4e451-e3ee-434e-8e24-0c795925a48e-apiservice-cert\") pod \"packageserver-d55dfcdfc-jjf9r\" (UID: \"e5e4e451-e3ee-434e-8e24-0c795925a48e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jjf9r" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.092805 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1718efb3-c63c-4aac-bc59-2cc85a59ba61-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-sdg4b\" (UID: \"1718efb3-c63c-4aac-bc59-2cc85a59ba61\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sdg4b" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.092829 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fjqp\" (UniqueName: \"kubernetes.io/projected/1718efb3-c63c-4aac-bc59-2cc85a59ba61-kube-api-access-6fjqp\") pod \"openshift-apiserver-operator-796bbdcf4f-sdg4b\" (UID: \"1718efb3-c63c-4aac-bc59-2cc85a59ba61\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sdg4b" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.092850 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e2885220-fafe-48da-9ed7-98942886ab1f-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-9rgdg\" (UID: \"e2885220-fafe-48da-9ed7-98942886ab1f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9rgdg" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.092859 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r7684"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.092873 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjvtl\" (UniqueName: \"kubernetes.io/projected/d7d02f2c-c052-4320-bac4-394fb40c57d2-kube-api-access-tjvtl\") pod \"dns-operator-744455d44c-927z2\" (UID: \"d7d02f2c-c052-4320-bac4-394fb40c57d2\") " pod="openshift-dns-operator/dns-operator-744455d44c-927z2" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.092897 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e5356f1d-5afd-47bd-89af-72abfa8f901f-service-ca-bundle\") pod \"router-default-5444994796-n6z57\" (UID: \"e5356f1d-5afd-47bd-89af-72abfa8f901f\") " pod="openshift-ingress/router-default-5444994796-n6z57" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.092920 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e5356f1d-5afd-47bd-89af-72abfa8f901f-metrics-certs\") pod \"router-default-5444994796-n6z57\" (UID: \"e5356f1d-5afd-47bd-89af-72abfa8f901f\") " pod="openshift-ingress/router-default-5444994796-n6z57" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.092964 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d30c4960-b4da-41cf-9d1c-771e0bc7b262-metrics-tls\") pod \"ingress-operator-5b745b69d9-kjlr8\" (UID: \"d30c4960-b4da-41cf-9d1c-771e0bc7b262\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kjlr8" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.092986 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8e03364a-1840-4c91-8b2b-c7e2071fc73b-console-oauth-config\") pod \"console-f9d7485db-4pwcw\" (UID: \"8e03364a-1840-4c91-8b2b-c7e2071fc73b\") " pod="openshift-console/console-f9d7485db-4pwcw" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.093010 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrw5q\" (UniqueName: \"kubernetes.io/projected/cf124a04-e72c-4c67-bd76-52648b4d20c9-kube-api-access-mrw5q\") pod \"kube-storage-version-migrator-operator-b67b599dd-r7684\" (UID: \"cf124a04-e72c-4c67-bd76-52648b4d20c9\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r7684" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.093032 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f3a30985-f268-44de-93b9-bce0f9aebc6d-trusted-ca\") pod \"console-operator-58897d9998-hlvjb\" (UID: \"f3a30985-f268-44de-93b9-bce0f9aebc6d\") " pod="openshift-console-operator/console-operator-58897d9998-hlvjb" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.093058 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3a30985-f268-44de-93b9-bce0f9aebc6d-config\") pod \"console-operator-58897d9998-hlvjb\" (UID: \"f3a30985-f268-44de-93b9-bce0f9aebc6d\") " pod="openshift-console-operator/console-operator-58897d9998-hlvjb" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.093089 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/d0b4b35c-3cb0-405d-941b-1b3946b228a0-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-kr48t\" (UID: \"d0b4b35c-3cb0-405d-941b-1b3946b228a0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kr48t" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.093114 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/00bf9f05-7bc3-40e1-a2e9-1af1bf93f014-profile-collector-cert\") pod \"catalog-operator-68c6474976-pkq7h\" (UID: \"00bf9f05-7bc3-40e1-a2e9-1af1bf93f014\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pkq7h" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.093140 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8e03364a-1840-4c91-8b2b-c7e2071fc73b-service-ca\") pod \"console-f9d7485db-4pwcw\" (UID: \"8e03364a-1840-4c91-8b2b-c7e2071fc73b\") " pod="openshift-console/console-f9d7485db-4pwcw" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.093217 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.093589 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c8703a5f-60fb-40f5-be0c-d1025852c05a-bound-sa-token\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.093626 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d0b4b35c-3cb0-405d-941b-1b3946b228a0-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-kr48t\" (UID: \"d0b4b35c-3cb0-405d-941b-1b3946b228a0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kr48t" Dec 06 05:53:42 crc kubenswrapper[4809]: E1206 05:53:42.093901 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:42.593888312 +0000 UTC m=+147.482871314 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.094740 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-nff6m"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.097330 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-c8sn6"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.098892 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c8703a5f-60fb-40f5-be0c-d1025852c05a-registry-tls\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.099129 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1718efb3-c63c-4aac-bc59-2cc85a59ba61-config\") pod \"openshift-apiserver-operator-796bbdcf4f-sdg4b\" (UID: \"1718efb3-c63c-4aac-bc59-2cc85a59ba61\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sdg4b" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.099144 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2885220-fafe-48da-9ed7-98942886ab1f-config\") pod \"kube-controller-manager-operator-78b949d7b-9rgdg\" (UID: \"e2885220-fafe-48da-9ed7-98942886ab1f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9rgdg" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.099186 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cf124a04-e72c-4c67-bd76-52648b4d20c9-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-r7684\" (UID: \"cf124a04-e72c-4c67-bd76-52648b4d20c9\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r7684" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.099218 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59tgt\" (UniqueName: \"kubernetes.io/projected/8e290142-b2fc-492e-8763-ff41224579fc-kube-api-access-59tgt\") pod \"downloads-7954f5f757-224mv\" (UID: \"8e290142-b2fc-492e-8763-ff41224579fc\") " pod="openshift-console/downloads-7954f5f757-224mv" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.099248 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e03364a-1840-4c91-8b2b-c7e2071fc73b-trusted-ca-bundle\") pod \"console-f9d7485db-4pwcw\" (UID: \"8e03364a-1840-4c91-8b2b-c7e2071fc73b\") " pod="openshift-console/console-f9d7485db-4pwcw" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.099260 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e5e4e451-e3ee-434e-8e24-0c795925a48e-webhook-cert\") pod \"packageserver-d55dfcdfc-jjf9r\" (UID: \"e5e4e451-e3ee-434e-8e24-0c795925a48e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jjf9r" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.099286 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/63b0cea8-1716-4c17-9f87-f10e776cf47f-etcd-client\") pod \"etcd-operator-b45778765-drj7v\" (UID: \"63b0cea8-1716-4c17-9f87-f10e776cf47f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-drj7v" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.099778 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8e03364a-1840-4c91-8b2b-c7e2071fc73b-console-config\") pod \"console-f9d7485db-4pwcw\" (UID: \"8e03364a-1840-4c91-8b2b-c7e2071fc73b\") " pod="openshift-console/console-f9d7485db-4pwcw" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.099873 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1718efb3-c63c-4aac-bc59-2cc85a59ba61-config\") pod \"openshift-apiserver-operator-796bbdcf4f-sdg4b\" (UID: \"1718efb3-c63c-4aac-bc59-2cc85a59ba61\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sdg4b" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.100027 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c8703a5f-60fb-40f5-be0c-d1025852c05a-trusted-ca\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.100160 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55jls\" (UniqueName: \"kubernetes.io/projected/c8703a5f-60fb-40f5-be0c-d1025852c05a-kube-api-access-55jls\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.100240 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf124a04-e72c-4c67-bd76-52648b4d20c9-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-r7684\" (UID: \"cf124a04-e72c-4c67-bd76-52648b4d20c9\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r7684" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.100274 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/b119ee35-1a33-4a34-9fca-62ad465a20f7-node-bootstrap-token\") pod \"machine-config-server-g7pfc\" (UID: \"b119ee35-1a33-4a34-9fca-62ad465a20f7\") " pod="openshift-machine-config-operator/machine-config-server-g7pfc" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.100787 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qw5pv\" (UniqueName: \"kubernetes.io/projected/b119ee35-1a33-4a34-9fca-62ad465a20f7-kube-api-access-qw5pv\") pod \"machine-config-server-g7pfc\" (UID: \"b119ee35-1a33-4a34-9fca-62ad465a20f7\") " pod="openshift-machine-config-operator/machine-config-server-g7pfc" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.101018 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/26b87745-a15b-4f2d-b98e-b24afc5d512b-signing-key\") pod \"service-ca-9c57cc56f-c8sn6\" (UID: \"26b87745-a15b-4f2d-b98e-b24afc5d512b\") " pod="openshift-service-ca/service-ca-9c57cc56f-c8sn6" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.101081 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cdf2d3c6-659f-4dc5-a21c-d29cc86ef239-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-wcth2\" (UID: \"cdf2d3c6-659f-4dc5-a21c-d29cc86ef239\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wcth2" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.101119 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2a1ed741-489c-455f-a344-404bbab66c22-serving-cert\") pod \"authentication-operator-69f744f599-7qpjs\" (UID: \"2a1ed741-489c-455f-a344-404bbab66c22\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7qpjs" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.101358 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cea05156-8322-404a-8b33-867aeaaa905e-serving-cert\") pod \"service-ca-operator-777779d784-mxv9r\" (UID: \"cea05156-8322-404a-8b33-867aeaaa905e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mxv9r" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.101434 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c8d79e6f-40b0-4ecc-8501-454ba5cfe691-srv-cert\") pod \"olm-operator-6b444d44fb-gt9kb\" (UID: \"c8d79e6f-40b0-4ecc-8501-454ba5cfe691\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gt9kb" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.101491 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qt5sg\" (UniqueName: \"kubernetes.io/projected/f3a30985-f268-44de-93b9-bce0f9aebc6d-kube-api-access-qt5sg\") pod \"console-operator-58897d9998-hlvjb\" (UID: \"f3a30985-f268-44de-93b9-bce0f9aebc6d\") " pod="openshift-console-operator/console-operator-58897d9998-hlvjb" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.101833 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf124a04-e72c-4c67-bd76-52648b4d20c9-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-r7684\" (UID: \"cf124a04-e72c-4c67-bd76-52648b4d20c9\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r7684" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.101986 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glkvh\" (UniqueName: \"kubernetes.io/projected/8e03364a-1840-4c91-8b2b-c7e2071fc73b-kube-api-access-glkvh\") pod \"console-f9d7485db-4pwcw\" (UID: \"8e03364a-1840-4c91-8b2b-c7e2071fc73b\") " pod="openshift-console/console-f9d7485db-4pwcw" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.102059 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/49fef68f-df3c-4bb7-80cf-f0f651797762-client-ca\") pod \"controller-manager-879f6c89f-ggvcl\" (UID: \"49fef68f-df3c-4bb7-80cf-f0f651797762\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ggvcl" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.102183 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8e03364a-1840-4c91-8b2b-c7e2071fc73b-service-ca\") pod \"console-f9d7485db-4pwcw\" (UID: \"8e03364a-1840-4c91-8b2b-c7e2071fc73b\") " pod="openshift-console/console-f9d7485db-4pwcw" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.102366 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c8703a5f-60fb-40f5-be0c-d1025852c05a-registry-certificates\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.102497 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c8703a5f-60fb-40f5-be0c-d1025852c05a-ca-trust-extracted\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.102577 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/b119ee35-1a33-4a34-9fca-62ad465a20f7-certs\") pod \"machine-config-server-g7pfc\" (UID: \"b119ee35-1a33-4a34-9fca-62ad465a20f7\") " pod="openshift-machine-config-operator/machine-config-server-g7pfc" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.102675 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zb9vt\" (UniqueName: \"kubernetes.io/projected/49fef68f-df3c-4bb7-80cf-f0f651797762-kube-api-access-zb9vt\") pod \"controller-manager-879f6c89f-ggvcl\" (UID: \"49fef68f-df3c-4bb7-80cf-f0f651797762\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ggvcl" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.102793 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d30c4960-b4da-41cf-9d1c-771e0bc7b262-trusted-ca\") pod \"ingress-operator-5b745b69d9-kjlr8\" (UID: \"d30c4960-b4da-41cf-9d1c-771e0bc7b262\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kjlr8" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.102862 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9g7x\" (UniqueName: \"kubernetes.io/projected/d0b4b35c-3cb0-405d-941b-1b3946b228a0-kube-api-access-c9g7x\") pod \"cluster-image-registry-operator-dc59b4c8b-kr48t\" (UID: \"d0b4b35c-3cb0-405d-941b-1b3946b228a0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kr48t" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.103021 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/e5356f1d-5afd-47bd-89af-72abfa8f901f-default-certificate\") pod \"router-default-5444994796-n6z57\" (UID: \"e5356f1d-5afd-47bd-89af-72abfa8f901f\") " pod="openshift-ingress/router-default-5444994796-n6z57" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.103065 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-tphf5"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.103251 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/26b87745-a15b-4f2d-b98e-b24afc5d512b-signing-cabundle\") pod \"service-ca-9c57cc56f-c8sn6\" (UID: \"26b87745-a15b-4f2d-b98e-b24afc5d512b\") " pod="openshift-service-ca/service-ca-9c57cc56f-c8sn6" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.103336 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/cdf2d3c6-659f-4dc5-a21c-d29cc86ef239-proxy-tls\") pod \"machine-config-controller-84d6567774-wcth2\" (UID: \"cdf2d3c6-659f-4dc5-a21c-d29cc86ef239\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wcth2" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.103445 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/e5e4e451-e3ee-434e-8e24-0c795925a48e-tmpfs\") pod \"packageserver-d55dfcdfc-jjf9r\" (UID: \"e5e4e451-e3ee-434e-8e24-0c795925a48e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jjf9r" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.103501 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/49fef68f-df3c-4bb7-80cf-f0f651797762-client-ca\") pod \"controller-manager-879f6c89f-ggvcl\" (UID: \"49fef68f-df3c-4bb7-80cf-f0f651797762\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ggvcl" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.104081 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c8703a5f-60fb-40f5-be0c-d1025852c05a-ca-trust-extracted\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.104460 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49fef68f-df3c-4bb7-80cf-f0f651797762-config\") pod \"controller-manager-879f6c89f-ggvcl\" (UID: \"49fef68f-df3c-4bb7-80cf-f0f651797762\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ggvcl" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.105040 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.105747 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/49fef68f-df3c-4bb7-80cf-f0f651797762-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-ggvcl\" (UID: \"49fef68f-df3c-4bb7-80cf-f0f651797762\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ggvcl" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.105786 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-cd2tv"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.109610 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-p2lkt"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.109748 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d30c4960-b4da-41cf-9d1c-771e0bc7b262-trusted-ca\") pod \"ingress-operator-5b745b69d9-kjlr8\" (UID: \"d30c4960-b4da-41cf-9d1c-771e0bc7b262\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kjlr8" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.112648 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e2885220-fafe-48da-9ed7-98942886ab1f-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-9rgdg\" (UID: \"e2885220-fafe-48da-9ed7-98942886ab1f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9rgdg" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.113153 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-mxv9r"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.113265 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-p2lkt" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.115147 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-852bd"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.120487 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jjf9r"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.123536 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8e03364a-1840-4c91-8b2b-c7e2071fc73b-console-oauth-config\") pod \"console-f9d7485db-4pwcw\" (UID: \"8e03364a-1840-4c91-8b2b-c7e2071fc73b\") " pod="openshift-console/console-f9d7485db-4pwcw" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.126373 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.129992 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-slt9h"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.145816 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8e03364a-1840-4c91-8b2b-c7e2071fc73b-console-serving-cert\") pod \"console-f9d7485db-4pwcw\" (UID: \"8e03364a-1840-4c91-8b2b-c7e2071fc73b\") " pod="openshift-console/console-f9d7485db-4pwcw" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.146290 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1718efb3-c63c-4aac-bc59-2cc85a59ba61-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-sdg4b\" (UID: \"1718efb3-c63c-4aac-bc59-2cc85a59ba61\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sdg4b" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.148790 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d30c4960-b4da-41cf-9d1c-771e0bc7b262-metrics-tls\") pod \"ingress-operator-5b745b69d9-kjlr8\" (UID: \"d30c4960-b4da-41cf-9d1c-771e0bc7b262\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kjlr8" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.152365 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cf124a04-e72c-4c67-bd76-52648b4d20c9-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-r7684\" (UID: \"cf124a04-e72c-4c67-bd76-52648b4d20c9\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r7684" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.152453 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jn446"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.154647 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/49fef68f-df3c-4bb7-80cf-f0f651797762-serving-cert\") pod \"controller-manager-879f6c89f-ggvcl\" (UID: \"49fef68f-df3c-4bb7-80cf-f0f651797762\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ggvcl" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.157719 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-224mv"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.159658 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-p2lkt"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.161128 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zj6j9"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.165733 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.166086 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.166197 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-czhdd"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.167492 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c8703a5f-60fb-40f5-be0c-d1025852c05a-installation-pull-secrets\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.171203 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-czhdd" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.187318 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.187532 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-s7px9"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.188434 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-czhdd"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.188455 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-s7px9"] Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.188518 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-s7px9" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.203993 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:42 crc kubenswrapper[4809]: E1206 05:53:42.204241 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:42.704167729 +0000 UTC m=+147.593150681 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.204404 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/de0bb137-3c1d-4632-a7ef-6922f027a3c1-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-dpbfw\" (UID: \"de0bb137-3c1d-4632-a7ef-6922f027a3c1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dpbfw" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.205175 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcpcz\" (UniqueName: \"kubernetes.io/projected/f4008f48-e17b-492f-8fa4-b09fb9d2ce70-kube-api-access-zcpcz\") pod \"multus-admission-controller-857f4d67dd-s7q7v\" (UID: \"f4008f48-e17b-492f-8fa4-b09fb9d2ce70\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-s7q7v" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.204775 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.205269 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d0b4b35c-3cb0-405d-941b-1b3946b228a0-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-kr48t\" (UID: \"d0b4b35c-3cb0-405d-941b-1b3946b228a0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kr48t" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.205596 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ls6xx\" (UniqueName: \"kubernetes.io/projected/26b87745-a15b-4f2d-b98e-b24afc5d512b-kube-api-access-ls6xx\") pod \"service-ca-9c57cc56f-c8sn6\" (UID: \"26b87745-a15b-4f2d-b98e-b24afc5d512b\") " pod="openshift-service-ca/service-ca-9c57cc56f-c8sn6" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.205680 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/00bf9f05-7bc3-40e1-a2e9-1af1bf93f014-srv-cert\") pod \"catalog-operator-68c6474976-pkq7h\" (UID: \"00bf9f05-7bc3-40e1-a2e9-1af1bf93f014\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pkq7h" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.205908 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63b0cea8-1716-4c17-9f87-f10e776cf47f-config\") pod \"etcd-operator-b45778765-drj7v\" (UID: \"63b0cea8-1716-4c17-9f87-f10e776cf47f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-drj7v" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.206902 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63b0cea8-1716-4c17-9f87-f10e776cf47f-config\") pod \"etcd-operator-b45778765-drj7v\" (UID: \"63b0cea8-1716-4c17-9f87-f10e776cf47f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-drj7v" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.207979 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/de0bb137-3c1d-4632-a7ef-6922f027a3c1-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-dpbfw\" (UID: \"de0bb137-3c1d-4632-a7ef-6922f027a3c1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dpbfw" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.208093 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a1ed741-489c-455f-a344-404bbab66c22-config\") pod \"authentication-operator-69f744f599-7qpjs\" (UID: \"2a1ed741-489c-455f-a344-404bbab66c22\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7qpjs" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.208124 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2kwh\" (UniqueName: \"kubernetes.io/projected/2a1ed741-489c-455f-a344-404bbab66c22-kube-api-access-t2kwh\") pod \"authentication-operator-69f744f599-7qpjs\" (UID: \"2a1ed741-489c-455f-a344-404bbab66c22\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7qpjs" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.208167 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rj9g\" (UniqueName: \"kubernetes.io/projected/63b0cea8-1716-4c17-9f87-f10e776cf47f-kube-api-access-6rj9g\") pod \"etcd-operator-b45778765-drj7v\" (UID: \"63b0cea8-1716-4c17-9f87-f10e776cf47f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-drj7v" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.208193 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8a225b5c-4ad5-4f08-a0cb-ae38a8fa83d4-available-featuregates\") pod \"openshift-config-operator-7777fb866f-qqv6x\" (UID: \"8a225b5c-4ad5-4f08-a0cb-ae38a8fa83d4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-qqv6x" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.208248 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/63b0cea8-1716-4c17-9f87-f10e776cf47f-etcd-service-ca\") pod \"etcd-operator-b45778765-drj7v\" (UID: \"63b0cea8-1716-4c17-9f87-f10e776cf47f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-drj7v" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.208278 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnqk2\" (UniqueName: \"kubernetes.io/projected/d233521e-75db-4da3-9cd3-0480a99eba90-kube-api-access-hnqk2\") pod \"migrator-59844c95c7-8ssrl\" (UID: \"d233521e-75db-4da3-9cd3-0480a99eba90\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8ssrl" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.208300 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cea05156-8322-404a-8b33-867aeaaa905e-config\") pod \"service-ca-operator-777779d784-mxv9r\" (UID: \"cea05156-8322-404a-8b33-867aeaaa905e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mxv9r" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.208342 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c8d79e6f-40b0-4ecc-8501-454ba5cfe691-profile-collector-cert\") pod \"olm-operator-6b444d44fb-gt9kb\" (UID: \"c8d79e6f-40b0-4ecc-8501-454ba5cfe691\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gt9kb" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.208367 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e5e4e451-e3ee-434e-8e24-0c795925a48e-apiservice-cert\") pod \"packageserver-d55dfcdfc-jjf9r\" (UID: \"e5e4e451-e3ee-434e-8e24-0c795925a48e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jjf9r" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.208422 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjvtl\" (UniqueName: \"kubernetes.io/projected/d7d02f2c-c052-4320-bac4-394fb40c57d2-kube-api-access-tjvtl\") pod \"dns-operator-744455d44c-927z2\" (UID: \"d7d02f2c-c052-4320-bac4-394fb40c57d2\") " pod="openshift-dns-operator/dns-operator-744455d44c-927z2" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.208445 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e5356f1d-5afd-47bd-89af-72abfa8f901f-service-ca-bundle\") pod \"router-default-5444994796-n6z57\" (UID: \"e5356f1d-5afd-47bd-89af-72abfa8f901f\") " pod="openshift-ingress/router-default-5444994796-n6z57" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.208487 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e5356f1d-5afd-47bd-89af-72abfa8f901f-metrics-certs\") pod \"router-default-5444994796-n6z57\" (UID: \"e5356f1d-5afd-47bd-89af-72abfa8f901f\") " pod="openshift-ingress/router-default-5444994796-n6z57" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.208519 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f3a30985-f268-44de-93b9-bce0f9aebc6d-trusted-ca\") pod \"console-operator-58897d9998-hlvjb\" (UID: \"f3a30985-f268-44de-93b9-bce0f9aebc6d\") " pod="openshift-console-operator/console-operator-58897d9998-hlvjb" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.208542 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3a30985-f268-44de-93b9-bce0f9aebc6d-config\") pod \"console-operator-58897d9998-hlvjb\" (UID: \"f3a30985-f268-44de-93b9-bce0f9aebc6d\") " pod="openshift-console-operator/console-operator-58897d9998-hlvjb" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.208589 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/d0b4b35c-3cb0-405d-941b-1b3946b228a0-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-kr48t\" (UID: \"d0b4b35c-3cb0-405d-941b-1b3946b228a0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kr48t" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.208612 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/00bf9f05-7bc3-40e1-a2e9-1af1bf93f014-profile-collector-cert\") pod \"catalog-operator-68c6474976-pkq7h\" (UID: \"00bf9f05-7bc3-40e1-a2e9-1af1bf93f014\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pkq7h" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.208665 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.208695 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d0b4b35c-3cb0-405d-941b-1b3946b228a0-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-kr48t\" (UID: \"d0b4b35c-3cb0-405d-941b-1b3946b228a0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kr48t" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.208743 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59tgt\" (UniqueName: \"kubernetes.io/projected/8e290142-b2fc-492e-8763-ff41224579fc-kube-api-access-59tgt\") pod \"downloads-7954f5f757-224mv\" (UID: \"8e290142-b2fc-492e-8763-ff41224579fc\") " pod="openshift-console/downloads-7954f5f757-224mv" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.208772 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e5e4e451-e3ee-434e-8e24-0c795925a48e-webhook-cert\") pod \"packageserver-d55dfcdfc-jjf9r\" (UID: \"e5e4e451-e3ee-434e-8e24-0c795925a48e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jjf9r" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.208815 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/63b0cea8-1716-4c17-9f87-f10e776cf47f-etcd-client\") pod \"etcd-operator-b45778765-drj7v\" (UID: \"63b0cea8-1716-4c17-9f87-f10e776cf47f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-drj7v" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.208858 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/b119ee35-1a33-4a34-9fca-62ad465a20f7-node-bootstrap-token\") pod \"machine-config-server-g7pfc\" (UID: \"b119ee35-1a33-4a34-9fca-62ad465a20f7\") " pod="openshift-machine-config-operator/machine-config-server-g7pfc" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.208898 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qw5pv\" (UniqueName: \"kubernetes.io/projected/b119ee35-1a33-4a34-9fca-62ad465a20f7-kube-api-access-qw5pv\") pod \"machine-config-server-g7pfc\" (UID: \"b119ee35-1a33-4a34-9fca-62ad465a20f7\") " pod="openshift-machine-config-operator/machine-config-server-g7pfc" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.208925 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/26b87745-a15b-4f2d-b98e-b24afc5d512b-signing-key\") pod \"service-ca-9c57cc56f-c8sn6\" (UID: \"26b87745-a15b-4f2d-b98e-b24afc5d512b\") " pod="openshift-service-ca/service-ca-9c57cc56f-c8sn6" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.208994 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cdf2d3c6-659f-4dc5-a21c-d29cc86ef239-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-wcth2\" (UID: \"cdf2d3c6-659f-4dc5-a21c-d29cc86ef239\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wcth2" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.209020 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2a1ed741-489c-455f-a344-404bbab66c22-serving-cert\") pod \"authentication-operator-69f744f599-7qpjs\" (UID: \"2a1ed741-489c-455f-a344-404bbab66c22\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7qpjs" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.209078 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cea05156-8322-404a-8b33-867aeaaa905e-serving-cert\") pod \"service-ca-operator-777779d784-mxv9r\" (UID: \"cea05156-8322-404a-8b33-867aeaaa905e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mxv9r" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.209099 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c8d79e6f-40b0-4ecc-8501-454ba5cfe691-srv-cert\") pod \"olm-operator-6b444d44fb-gt9kb\" (UID: \"c8d79e6f-40b0-4ecc-8501-454ba5cfe691\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gt9kb" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.209119 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qt5sg\" (UniqueName: \"kubernetes.io/projected/f3a30985-f268-44de-93b9-bce0f9aebc6d-kube-api-access-qt5sg\") pod \"console-operator-58897d9998-hlvjb\" (UID: \"f3a30985-f268-44de-93b9-bce0f9aebc6d\") " pod="openshift-console-operator/console-operator-58897d9998-hlvjb" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.209188 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/b119ee35-1a33-4a34-9fca-62ad465a20f7-certs\") pod \"machine-config-server-g7pfc\" (UID: \"b119ee35-1a33-4a34-9fca-62ad465a20f7\") " pod="openshift-machine-config-operator/machine-config-server-g7pfc" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.209234 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9g7x\" (UniqueName: \"kubernetes.io/projected/d0b4b35c-3cb0-405d-941b-1b3946b228a0-kube-api-access-c9g7x\") pod \"cluster-image-registry-operator-dc59b4c8b-kr48t\" (UID: \"d0b4b35c-3cb0-405d-941b-1b3946b228a0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kr48t" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.209258 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/e5356f1d-5afd-47bd-89af-72abfa8f901f-default-certificate\") pod \"router-default-5444994796-n6z57\" (UID: \"e5356f1d-5afd-47bd-89af-72abfa8f901f\") " pod="openshift-ingress/router-default-5444994796-n6z57" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.209303 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/26b87745-a15b-4f2d-b98e-b24afc5d512b-signing-cabundle\") pod \"service-ca-9c57cc56f-c8sn6\" (UID: \"26b87745-a15b-4f2d-b98e-b24afc5d512b\") " pod="openshift-service-ca/service-ca-9c57cc56f-c8sn6" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.209323 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/cdf2d3c6-659f-4dc5-a21c-d29cc86ef239-proxy-tls\") pod \"machine-config-controller-84d6567774-wcth2\" (UID: \"cdf2d3c6-659f-4dc5-a21c-d29cc86ef239\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wcth2" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.209343 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/e5e4e451-e3ee-434e-8e24-0c795925a48e-tmpfs\") pod \"packageserver-d55dfcdfc-jjf9r\" (UID: \"e5e4e451-e3ee-434e-8e24-0c795925a48e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jjf9r" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.209391 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wh944\" (UniqueName: \"kubernetes.io/projected/e5356f1d-5afd-47bd-89af-72abfa8f901f-kube-api-access-wh944\") pod \"router-default-5444994796-n6z57\" (UID: \"e5356f1d-5afd-47bd-89af-72abfa8f901f\") " pod="openshift-ingress/router-default-5444994796-n6z57" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.209461 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/e5356f1d-5afd-47bd-89af-72abfa8f901f-stats-auth\") pod \"router-default-5444994796-n6z57\" (UID: \"e5356f1d-5afd-47bd-89af-72abfa8f901f\") " pod="openshift-ingress/router-default-5444994796-n6z57" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.209488 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/63b0cea8-1716-4c17-9f87-f10e776cf47f-etcd-ca\") pod \"etcd-operator-b45778765-drj7v\" (UID: \"63b0cea8-1716-4c17-9f87-f10e776cf47f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-drj7v" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.209512 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhvfg\" (UniqueName: \"kubernetes.io/projected/cdf2d3c6-659f-4dc5-a21c-d29cc86ef239-kube-api-access-vhvfg\") pod \"machine-config-controller-84d6567774-wcth2\" (UID: \"cdf2d3c6-659f-4dc5-a21c-d29cc86ef239\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wcth2" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.209551 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f3a30985-f268-44de-93b9-bce0f9aebc6d-serving-cert\") pod \"console-operator-58897d9998-hlvjb\" (UID: \"f3a30985-f268-44de-93b9-bce0f9aebc6d\") " pod="openshift-console-operator/console-operator-58897d9998-hlvjb" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.209571 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d7d02f2c-c052-4320-bac4-394fb40c57d2-metrics-tls\") pod \"dns-operator-744455d44c-927z2\" (UID: \"d7d02f2c-c052-4320-bac4-394fb40c57d2\") " pod="openshift-dns-operator/dns-operator-744455d44c-927z2" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.209592 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f4008f48-e17b-492f-8fa4-b09fb9d2ce70-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-s7q7v\" (UID: \"f4008f48-e17b-492f-8fa4-b09fb9d2ce70\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-s7q7v" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.209643 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjbbp\" (UniqueName: \"kubernetes.io/projected/00bf9f05-7bc3-40e1-a2e9-1af1bf93f014-kube-api-access-pjbbp\") pod \"catalog-operator-68c6474976-pkq7h\" (UID: \"00bf9f05-7bc3-40e1-a2e9-1af1bf93f014\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pkq7h" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.209665 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2a1ed741-489c-455f-a344-404bbab66c22-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-7qpjs\" (UID: \"2a1ed741-489c-455f-a344-404bbab66c22\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7qpjs" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.209708 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a225b5c-4ad5-4f08-a0cb-ae38a8fa83d4-serving-cert\") pod \"openshift-config-operator-7777fb866f-qqv6x\" (UID: \"8a225b5c-4ad5-4f08-a0cb-ae38a8fa83d4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-qqv6x" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.209741 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2a1ed741-489c-455f-a344-404bbab66c22-service-ca-bundle\") pod \"authentication-operator-69f744f599-7qpjs\" (UID: \"2a1ed741-489c-455f-a344-404bbab66c22\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7qpjs" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.209805 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/63b0cea8-1716-4c17-9f87-f10e776cf47f-serving-cert\") pod \"etcd-operator-b45778765-drj7v\" (UID: \"63b0cea8-1716-4c17-9f87-f10e776cf47f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-drj7v" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.209835 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d844d778-6b3d-400a-8f71-db1c2e643878-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-zj6j9\" (UID: \"d844d778-6b3d-400a-8f71-db1c2e643878\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zj6j9" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.209884 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vfkx\" (UniqueName: \"kubernetes.io/projected/d844d778-6b3d-400a-8f71-db1c2e643878-kube-api-access-9vfkx\") pod \"openshift-controller-manager-operator-756b6f6bc6-zj6j9\" (UID: \"d844d778-6b3d-400a-8f71-db1c2e643878\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zj6j9" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.209905 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e5356f1d-5afd-47bd-89af-72abfa8f901f-service-ca-bundle\") pod \"router-default-5444994796-n6z57\" (UID: \"e5356f1d-5afd-47bd-89af-72abfa8f901f\") " pod="openshift-ingress/router-default-5444994796-n6z57" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.209970 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6nl2\" (UniqueName: \"kubernetes.io/projected/de0bb137-3c1d-4632-a7ef-6922f027a3c1-kube-api-access-v6nl2\") pod \"control-plane-machine-set-operator-78cbb6b69f-dpbfw\" (UID: \"de0bb137-3c1d-4632-a7ef-6922f027a3c1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dpbfw" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.209996 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bplj9\" (UniqueName: \"kubernetes.io/projected/c8d79e6f-40b0-4ecc-8501-454ba5cfe691-kube-api-access-bplj9\") pod \"olm-operator-6b444d44fb-gt9kb\" (UID: \"c8d79e6f-40b0-4ecc-8501-454ba5cfe691\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gt9kb" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.210019 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wfkz\" (UniqueName: \"kubernetes.io/projected/e5e4e451-e3ee-434e-8e24-0c795925a48e-kube-api-access-9wfkz\") pod \"packageserver-d55dfcdfc-jjf9r\" (UID: \"e5e4e451-e3ee-434e-8e24-0c795925a48e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jjf9r" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.210073 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25sjn\" (UniqueName: \"kubernetes.io/projected/cea05156-8322-404a-8b33-867aeaaa905e-kube-api-access-25sjn\") pod \"service-ca-operator-777779d784-mxv9r\" (UID: \"cea05156-8322-404a-8b33-867aeaaa905e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mxv9r" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.210097 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9tmj\" (UniqueName: \"kubernetes.io/projected/8a225b5c-4ad5-4f08-a0cb-ae38a8fa83d4-kube-api-access-v9tmj\") pod \"openshift-config-operator-7777fb866f-qqv6x\" (UID: \"8a225b5c-4ad5-4f08-a0cb-ae38a8fa83d4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-qqv6x" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.210148 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d844d778-6b3d-400a-8f71-db1c2e643878-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-zj6j9\" (UID: \"d844d778-6b3d-400a-8f71-db1c2e643878\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zj6j9" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.210655 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a1ed741-489c-455f-a344-404bbab66c22-config\") pod \"authentication-operator-69f744f599-7qpjs\" (UID: \"2a1ed741-489c-455f-a344-404bbab66c22\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7qpjs" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.211236 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8a225b5c-4ad5-4f08-a0cb-ae38a8fa83d4-available-featuregates\") pod \"openshift-config-operator-7777fb866f-qqv6x\" (UID: \"8a225b5c-4ad5-4f08-a0cb-ae38a8fa83d4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-qqv6x" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.211814 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/63b0cea8-1716-4c17-9f87-f10e776cf47f-etcd-service-ca\") pod \"etcd-operator-b45778765-drj7v\" (UID: \"63b0cea8-1716-4c17-9f87-f10e776cf47f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-drj7v" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.213536 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2a1ed741-489c-455f-a344-404bbab66c22-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-7qpjs\" (UID: \"2a1ed741-489c-455f-a344-404bbab66c22\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7qpjs" Dec 06 05:53:42 crc kubenswrapper[4809]: E1206 05:53:42.213736 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:42.713717799 +0000 UTC m=+147.602700791 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.213859 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/63b0cea8-1716-4c17-9f87-f10e776cf47f-etcd-ca\") pod \"etcd-operator-b45778765-drj7v\" (UID: \"63b0cea8-1716-4c17-9f87-f10e776cf47f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-drj7v" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.213896 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2a1ed741-489c-455f-a344-404bbab66c22-service-ca-bundle\") pod \"authentication-operator-69f744f599-7qpjs\" (UID: \"2a1ed741-489c-455f-a344-404bbab66c22\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7qpjs" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.214898 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/e5e4e451-e3ee-434e-8e24-0c795925a48e-tmpfs\") pod \"packageserver-d55dfcdfc-jjf9r\" (UID: \"e5e4e451-e3ee-434e-8e24-0c795925a48e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jjf9r" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.215359 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cdf2d3c6-659f-4dc5-a21c-d29cc86ef239-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-wcth2\" (UID: \"cdf2d3c6-659f-4dc5-a21c-d29cc86ef239\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wcth2" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.215383 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d0b4b35c-3cb0-405d-941b-1b3946b228a0-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-kr48t\" (UID: \"d0b4b35c-3cb0-405d-941b-1b3946b228a0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kr48t" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.217048 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d7d02f2c-c052-4320-bac4-394fb40c57d2-metrics-tls\") pod \"dns-operator-744455d44c-927z2\" (UID: \"d7d02f2c-c052-4320-bac4-394fb40c57d2\") " pod="openshift-dns-operator/dns-operator-744455d44c-927z2" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.217475 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2a1ed741-489c-455f-a344-404bbab66c22-serving-cert\") pod \"authentication-operator-69f744f599-7qpjs\" (UID: \"2a1ed741-489c-455f-a344-404bbab66c22\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7qpjs" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.218326 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/e5356f1d-5afd-47bd-89af-72abfa8f901f-default-certificate\") pod \"router-default-5444994796-n6z57\" (UID: \"e5356f1d-5afd-47bd-89af-72abfa8f901f\") " pod="openshift-ingress/router-default-5444994796-n6z57" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.218710 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/63b0cea8-1716-4c17-9f87-f10e776cf47f-etcd-client\") pod \"etcd-operator-b45778765-drj7v\" (UID: \"63b0cea8-1716-4c17-9f87-f10e776cf47f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-drj7v" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.218858 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/e5356f1d-5afd-47bd-89af-72abfa8f901f-stats-auth\") pod \"router-default-5444994796-n6z57\" (UID: \"e5356f1d-5afd-47bd-89af-72abfa8f901f\") " pod="openshift-ingress/router-default-5444994796-n6z57" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.219427 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/63b0cea8-1716-4c17-9f87-f10e776cf47f-serving-cert\") pod \"etcd-operator-b45778765-drj7v\" (UID: \"63b0cea8-1716-4c17-9f87-f10e776cf47f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-drj7v" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.219736 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/cdf2d3c6-659f-4dc5-a21c-d29cc86ef239-proxy-tls\") pod \"machine-config-controller-84d6567774-wcth2\" (UID: \"cdf2d3c6-659f-4dc5-a21c-d29cc86ef239\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wcth2" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.221423 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/d0b4b35c-3cb0-405d-941b-1b3946b228a0-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-kr48t\" (UID: \"d0b4b35c-3cb0-405d-941b-1b3946b228a0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kr48t" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.223505 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e5356f1d-5afd-47bd-89af-72abfa8f901f-metrics-certs\") pod \"router-default-5444994796-n6z57\" (UID: \"e5356f1d-5afd-47bd-89af-72abfa8f901f\") " pod="openshift-ingress/router-default-5444994796-n6z57" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.223866 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.244298 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.263820 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.284160 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.304846 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.311460 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:42 crc kubenswrapper[4809]: E1206 05:53:42.311592 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:42.811556466 +0000 UTC m=+147.700539408 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.312075 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:42 crc kubenswrapper[4809]: E1206 05:53:42.312384 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:42.812376029 +0000 UTC m=+147.701358971 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.323656 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.344029 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.364332 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.383961 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.387442 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.404370 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.415175 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:42 crc kubenswrapper[4809]: E1206 05:53:42.415733 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:42.915716277 +0000 UTC m=+147.804699219 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.423783 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.444777 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.458024 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c8d79e6f-40b0-4ecc-8501-454ba5cfe691-srv-cert\") pod \"olm-operator-6b444d44fb-gt9kb\" (UID: \"c8d79e6f-40b0-4ecc-8501-454ba5cfe691\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gt9kb" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.463561 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.466351 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/00bf9f05-7bc3-40e1-a2e9-1af1bf93f014-profile-collector-cert\") pod \"catalog-operator-68c6474976-pkq7h\" (UID: \"00bf9f05-7bc3-40e1-a2e9-1af1bf93f014\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pkq7h" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.474265 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c8d79e6f-40b0-4ecc-8501-454ba5cfe691-profile-collector-cert\") pod \"olm-operator-6b444d44fb-gt9kb\" (UID: \"c8d79e6f-40b0-4ecc-8501-454ba5cfe691\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gt9kb" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.483762 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.503915 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.516784 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:42 crc kubenswrapper[4809]: E1206 05:53:42.517153 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:43.017140531 +0000 UTC m=+147.906123463 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.523645 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.544049 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.564890 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.579285 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a225b5c-4ad5-4f08-a0cb-ae38a8fa83d4-serving-cert\") pod \"openshift-config-operator-7777fb866f-qqv6x\" (UID: \"8a225b5c-4ad5-4f08-a0cb-ae38a8fa83d4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-qqv6x" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.584367 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.618062 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:42 crc kubenswrapper[4809]: E1206 05:53:42.618252 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:43.118233557 +0000 UTC m=+148.007216499 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.618587 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:42 crc kubenswrapper[4809]: E1206 05:53:42.618915 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:43.118906926 +0000 UTC m=+148.007889858 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.624169 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.644037 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.660316 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/26b87745-a15b-4f2d-b98e-b24afc5d512b-signing-key\") pod \"service-ca-9c57cc56f-c8sn6\" (UID: \"26b87745-a15b-4f2d-b98e-b24afc5d512b\") " pod="openshift-service-ca/service-ca-9c57cc56f-c8sn6" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.665380 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.678237 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f4008f48-e17b-492f-8fa4-b09fb9d2ce70-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-s7q7v\" (UID: \"f4008f48-e17b-492f-8fa4-b09fb9d2ce70\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-s7q7v" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.684322 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.694809 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/26b87745-a15b-4f2d-b98e-b24afc5d512b-signing-cabundle\") pod \"service-ca-9c57cc56f-c8sn6\" (UID: \"26b87745-a15b-4f2d-b98e-b24afc5d512b\") " pod="openshift-service-ca/service-ca-9c57cc56f-c8sn6" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.705099 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.720304 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:42 crc kubenswrapper[4809]: E1206 05:53:42.720559 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:43.220532425 +0000 UTC m=+148.109515367 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.720717 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:42 crc kubenswrapper[4809]: E1206 05:53:42.721370 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:43.221361038 +0000 UTC m=+148.110343980 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.725715 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.744717 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.764683 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.785040 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.803916 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.821781 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:42 crc kubenswrapper[4809]: E1206 05:53:42.822016 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:43.321978901 +0000 UTC m=+148.210961873 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.822584 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:42 crc kubenswrapper[4809]: E1206 05:53:42.822956 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:43.322919737 +0000 UTC m=+148.211902719 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.834445 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.844189 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.865234 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.884371 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.903699 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.917874 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e5e4e451-e3ee-434e-8e24-0c795925a48e-apiservice-cert\") pod \"packageserver-d55dfcdfc-jjf9r\" (UID: \"e5e4e451-e3ee-434e-8e24-0c795925a48e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jjf9r" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.922333 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e5e4e451-e3ee-434e-8e24-0c795925a48e-webhook-cert\") pod \"packageserver-d55dfcdfc-jjf9r\" (UID: \"e5e4e451-e3ee-434e-8e24-0c795925a48e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jjf9r" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.924009 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:42 crc kubenswrapper[4809]: E1206 05:53:42.924390 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:43.424358112 +0000 UTC m=+148.313341094 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.924414 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.924647 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:42 crc kubenswrapper[4809]: E1206 05:53:42.925264 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:43.425246927 +0000 UTC m=+148.314229899 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.931500 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/00bf9f05-7bc3-40e1-a2e9-1af1bf93f014-srv-cert\") pod \"catalog-operator-68c6474976-pkq7h\" (UID: \"00bf9f05-7bc3-40e1-a2e9-1af1bf93f014\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pkq7h" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.943848 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.964261 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.983790 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 06 05:53:42 crc kubenswrapper[4809]: I1206 05:53:42.997446 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f3a30985-f268-44de-93b9-bce0f9aebc6d-serving-cert\") pod \"console-operator-58897d9998-hlvjb\" (UID: \"f3a30985-f268-44de-93b9-bce0f9aebc6d\") " pod="openshift-console-operator/console-operator-58897d9998-hlvjb" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.004279 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.014046 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3a30985-f268-44de-93b9-bce0f9aebc6d-config\") pod \"console-operator-58897d9998-hlvjb\" (UID: \"f3a30985-f268-44de-93b9-bce0f9aebc6d\") " pod="openshift-console-operator/console-operator-58897d9998-hlvjb" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.023650 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.025962 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:43 crc kubenswrapper[4809]: E1206 05:53:43.026094 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:43.526074796 +0000 UTC m=+148.415057748 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.026299 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:43 crc kubenswrapper[4809]: E1206 05:53:43.026637 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:43.52662711 +0000 UTC m=+148.415610052 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.058691 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.062270 4809 request.go:700] Waited for 1.008883042s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/secrets?fieldSelector=metadata.name%3Dmachine-config-server-tls&limit=500&resourceVersion=0 Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.063274 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.064520 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f3a30985-f268-44de-93b9-bce0f9aebc6d-trusted-ca\") pod \"console-operator-58897d9998-hlvjb\" (UID: \"f3a30985-f268-44de-93b9-bce0f9aebc6d\") " pod="openshift-console-operator/console-operator-58897d9998-hlvjb" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.068186 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/b119ee35-1a33-4a34-9fca-62ad465a20f7-certs\") pod \"machine-config-server-g7pfc\" (UID: \"b119ee35-1a33-4a34-9fca-62ad465a20f7\") " pod="openshift-machine-config-operator/machine-config-server-g7pfc" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.084033 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.088788 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/b119ee35-1a33-4a34-9fca-62ad465a20f7-node-bootstrap-token\") pod \"machine-config-server-g7pfc\" (UID: \"b119ee35-1a33-4a34-9fca-62ad465a20f7\") " pod="openshift-machine-config-operator/machine-config-server-g7pfc" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.104266 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.123592 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.127501 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:43 crc kubenswrapper[4809]: E1206 05:53:43.128799 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:43.628528229 +0000 UTC m=+148.517511191 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.144240 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.164525 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.183774 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 06 05:53:43 crc kubenswrapper[4809]: E1206 05:53:43.210586 4809 configmap.go:193] Couldn't get configMap openshift-controller-manager-operator/openshift-controller-manager-operator-config: failed to sync configmap cache: timed out waiting for the condition Dec 06 05:53:43 crc kubenswrapper[4809]: E1206 05:53:43.210693 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/d844d778-6b3d-400a-8f71-db1c2e643878-config podName:d844d778-6b3d-400a-8f71-db1c2e643878 nodeName:}" failed. No retries permitted until 2025-12-06 05:53:43.710667995 +0000 UTC m=+148.599650957 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/d844d778-6b3d-400a-8f71-db1c2e643878-config") pod "openshift-controller-manager-operator-756b6f6bc6-zj6j9" (UID: "d844d778-6b3d-400a-8f71-db1c2e643878") : failed to sync configmap cache: timed out waiting for the condition Dec 06 05:53:43 crc kubenswrapper[4809]: E1206 05:53:43.212039 4809 configmap.go:193] Couldn't get configMap openshift-service-ca-operator/service-ca-operator-config: failed to sync configmap cache: timed out waiting for the condition Dec 06 05:53:43 crc kubenswrapper[4809]: E1206 05:53:43.212117 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/cea05156-8322-404a-8b33-867aeaaa905e-config podName:cea05156-8322-404a-8b33-867aeaaa905e nodeName:}" failed. No retries permitted until 2025-12-06 05:53:43.712099075 +0000 UTC m=+148.601082017 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/cea05156-8322-404a-8b33-867aeaaa905e-config") pod "service-ca-operator-777779d784-mxv9r" (UID: "cea05156-8322-404a-8b33-867aeaaa905e") : failed to sync configmap cache: timed out waiting for the condition Dec 06 05:53:43 crc kubenswrapper[4809]: E1206 05:53:43.215504 4809 secret.go:188] Couldn't get secret openshift-service-ca-operator/serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 06 05:53:43 crc kubenswrapper[4809]: E1206 05:53:43.215589 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cea05156-8322-404a-8b33-867aeaaa905e-serving-cert podName:cea05156-8322-404a-8b33-867aeaaa905e nodeName:}" failed. No retries permitted until 2025-12-06 05:53:43.71556743 +0000 UTC m=+148.604550432 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/cea05156-8322-404a-8b33-867aeaaa905e-serving-cert") pod "service-ca-operator-777779d784-mxv9r" (UID: "cea05156-8322-404a-8b33-867aeaaa905e") : failed to sync secret cache: timed out waiting for the condition Dec 06 05:53:43 crc kubenswrapper[4809]: E1206 05:53:43.215629 4809 secret.go:188] Couldn't get secret openshift-controller-manager-operator/openshift-controller-manager-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 06 05:53:43 crc kubenswrapper[4809]: E1206 05:53:43.215676 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d844d778-6b3d-400a-8f71-db1c2e643878-serving-cert podName:d844d778-6b3d-400a-8f71-db1c2e643878 nodeName:}" failed. No retries permitted until 2025-12-06 05:53:43.715661942 +0000 UTC m=+148.604644884 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/d844d778-6b3d-400a-8f71-db1c2e643878-serving-cert") pod "openshift-controller-manager-operator-756b6f6bc6-zj6j9" (UID: "d844d778-6b3d-400a-8f71-db1c2e643878") : failed to sync secret cache: timed out waiting for the condition Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.220340 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.229682 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:43 crc kubenswrapper[4809]: E1206 05:53:43.230049 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:43.730028235 +0000 UTC m=+148.619011187 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.230727 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.244413 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.264014 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.284202 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.303963 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.330426 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.330995 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:43 crc kubenswrapper[4809]: E1206 05:53:43.331112 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:43.83109529 +0000 UTC m=+148.720078232 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.331191 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.331318 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.331467 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.331726 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:43 crc kubenswrapper[4809]: E1206 05:53:43.332104 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:43.832091598 +0000 UTC m=+148.721074530 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.339488 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.343432 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.364110 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.383671 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.403641 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.423347 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.433175 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:43 crc kubenswrapper[4809]: E1206 05:53:43.433284 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:43.933263095 +0000 UTC m=+148.822246057 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.433407 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.433587 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:43 crc kubenswrapper[4809]: E1206 05:53:43.434229 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:43.934202461 +0000 UTC m=+148.823185483 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.437334 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.444659 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.464091 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.484568 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.504691 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.504704 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.520237 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.524607 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.534921 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:43 crc kubenswrapper[4809]: E1206 05:53:43.535073 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:44.03505087 +0000 UTC m=+148.924033812 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.535208 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:43 crc kubenswrapper[4809]: E1206 05:53:43.535567 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:44.035557795 +0000 UTC m=+148.924540847 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.544563 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.563835 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.583718 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.604477 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.625464 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.636110 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:43 crc kubenswrapper[4809]: E1206 05:53:43.636249 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:44.136232438 +0000 UTC m=+149.025215380 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.636488 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:43 crc kubenswrapper[4809]: E1206 05:53:43.637093 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:44.137083432 +0000 UTC m=+149.026066374 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.645337 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.664348 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.684468 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.706612 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.724576 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.737107 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:43 crc kubenswrapper[4809]: E1206 05:53:43.737271 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:44.237227222 +0000 UTC m=+149.126210164 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.737321 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cea05156-8322-404a-8b33-867aeaaa905e-config\") pod \"service-ca-operator-777779d784-mxv9r\" (UID: \"cea05156-8322-404a-8b33-867aeaaa905e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mxv9r" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.737403 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.737466 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cea05156-8322-404a-8b33-867aeaaa905e-serving-cert\") pod \"service-ca-operator-777779d784-mxv9r\" (UID: \"cea05156-8322-404a-8b33-867aeaaa905e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mxv9r" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.737705 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d844d778-6b3d-400a-8f71-db1c2e643878-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-zj6j9\" (UID: \"d844d778-6b3d-400a-8f71-db1c2e643878\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zj6j9" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.737830 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d844d778-6b3d-400a-8f71-db1c2e643878-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-zj6j9\" (UID: \"d844d778-6b3d-400a-8f71-db1c2e643878\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zj6j9" Dec 06 05:53:43 crc kubenswrapper[4809]: E1206 05:53:43.737992 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:44.237976482 +0000 UTC m=+149.126959424 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.738447 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cea05156-8322-404a-8b33-867aeaaa905e-config\") pod \"service-ca-operator-777779d784-mxv9r\" (UID: \"cea05156-8322-404a-8b33-867aeaaa905e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mxv9r" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.742789 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cea05156-8322-404a-8b33-867aeaaa905e-serving-cert\") pod \"service-ca-operator-777779d784-mxv9r\" (UID: \"cea05156-8322-404a-8b33-867aeaaa905e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mxv9r" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.743128 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d844d778-6b3d-400a-8f71-db1c2e643878-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-zj6j9\" (UID: \"d844d778-6b3d-400a-8f71-db1c2e643878\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zj6j9" Dec 06 05:53:43 crc kubenswrapper[4809]: W1206 05:53:43.744317 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-fde7958ba433cbc943797f2e0d1c63d508bbbde3f37080b29a038ca15f722c50 WatchSource:0}: Error finding container fde7958ba433cbc943797f2e0d1c63d508bbbde3f37080b29a038ca15f722c50: Status 404 returned error can't find the container with id fde7958ba433cbc943797f2e0d1c63d508bbbde3f37080b29a038ca15f722c50 Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.744438 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.748599 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d844d778-6b3d-400a-8f71-db1c2e643878-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-zj6j9\" (UID: \"d844d778-6b3d-400a-8f71-db1c2e643878\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zj6j9" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.763729 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.817147 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d30c4960-b4da-41cf-9d1c-771e0bc7b262-bound-sa-token\") pod \"ingress-operator-5b745b69d9-kjlr8\" (UID: \"d30c4960-b4da-41cf-9d1c-771e0bc7b262\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kjlr8" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.837285 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4j955\" (UniqueName: \"kubernetes.io/projected/d30c4960-b4da-41cf-9d1c-771e0bc7b262-kube-api-access-4j955\") pod \"ingress-operator-5b745b69d9-kjlr8\" (UID: \"d30c4960-b4da-41cf-9d1c-771e0bc7b262\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kjlr8" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.839095 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:43 crc kubenswrapper[4809]: E1206 05:53:43.839243 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:44.339224542 +0000 UTC m=+149.228207484 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.839568 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:43 crc kubenswrapper[4809]: E1206 05:53:43.839862 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:44.33985067 +0000 UTC m=+149.228833612 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.857056 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e2885220-fafe-48da-9ed7-98942886ab1f-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-9rgdg\" (UID: \"e2885220-fafe-48da-9ed7-98942886ab1f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9rgdg" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.878336 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fjqp\" (UniqueName: \"kubernetes.io/projected/1718efb3-c63c-4aac-bc59-2cc85a59ba61-kube-api-access-6fjqp\") pod \"openshift-apiserver-operator-796bbdcf4f-sdg4b\" (UID: \"1718efb3-c63c-4aac-bc59-2cc85a59ba61\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sdg4b" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.899759 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c8703a5f-60fb-40f5-be0c-d1025852c05a-bound-sa-token\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.920735 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55jls\" (UniqueName: \"kubernetes.io/projected/c8703a5f-60fb-40f5-be0c-d1025852c05a-kube-api-access-55jls\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.940486 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:43 crc kubenswrapper[4809]: E1206 05:53:43.940636 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:44.440607296 +0000 UTC m=+149.329590248 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.941408 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:43 crc kubenswrapper[4809]: E1206 05:53:43.941824 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:44.441811749 +0000 UTC m=+149.330794691 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.942445 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrw5q\" (UniqueName: \"kubernetes.io/projected/cf124a04-e72c-4c67-bd76-52648b4d20c9-kube-api-access-mrw5q\") pod \"kube-storage-version-migrator-operator-b67b599dd-r7684\" (UID: \"cf124a04-e72c-4c67-bd76-52648b4d20c9\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r7684" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.962805 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glkvh\" (UniqueName: \"kubernetes.io/projected/8e03364a-1840-4c91-8b2b-c7e2071fc73b-kube-api-access-glkvh\") pod \"console-f9d7485db-4pwcw\" (UID: \"8e03364a-1840-4c91-8b2b-c7e2071fc73b\") " pod="openshift-console/console-f9d7485db-4pwcw" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.979980 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zb9vt\" (UniqueName: \"kubernetes.io/projected/49fef68f-df3c-4bb7-80cf-f0f651797762-kube-api-access-zb9vt\") pod \"controller-manager-879f6c89f-ggvcl\" (UID: \"49fef68f-df3c-4bb7-80cf-f0f651797762\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ggvcl" Dec 06 05:53:43 crc kubenswrapper[4809]: I1206 05:53:43.983809 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.003351 4809 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.024204 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.028092 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sdg4b" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.039636 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-4pwcw" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.042518 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:44 crc kubenswrapper[4809]: E1206 05:53:44.042680 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:44.542643458 +0000 UTC m=+149.431626390 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.043001 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:44 crc kubenswrapper[4809]: E1206 05:53:44.043414 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:44.543377808 +0000 UTC m=+149.432360750 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.044121 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.051960 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-ggvcl" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.062335 4809 request.go:700] Waited for 1.89075462s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress-canary/secrets?fieldSelector=metadata.name%3Ddefault-dockercfg-2llfx&limit=500&resourceVersion=0 Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.064392 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.084611 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.094792 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kjlr8" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.108256 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r7684" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.110608 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.124203 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.129637 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9rgdg" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.129880 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"522a40c2a6991d0ac8155e305d37d9297d253479d1132ce29767258105b3daf2"} Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.130662 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"fde7958ba433cbc943797f2e0d1c63d508bbbde3f37080b29a038ca15f722c50"} Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.149739 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:44 crc kubenswrapper[4809]: E1206 05:53:44.150244 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:44.650225612 +0000 UTC m=+149.539208554 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.151434 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.165202 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.202032 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcpcz\" (UniqueName: \"kubernetes.io/projected/f4008f48-e17b-492f-8fa4-b09fb9d2ce70-kube-api-access-zcpcz\") pod \"multus-admission-controller-857f4d67dd-s7q7v\" (UID: \"f4008f48-e17b-492f-8fa4-b09fb9d2ce70\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-s7q7v" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.223518 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d0b4b35c-3cb0-405d-941b-1b3946b228a0-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-kr48t\" (UID: \"d0b4b35c-3cb0-405d-941b-1b3946b228a0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kr48t" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.247723 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ls6xx\" (UniqueName: \"kubernetes.io/projected/26b87745-a15b-4f2d-b98e-b24afc5d512b-kube-api-access-ls6xx\") pod \"service-ca-9c57cc56f-c8sn6\" (UID: \"26b87745-a15b-4f2d-b98e-b24afc5d512b\") " pod="openshift-service-ca/service-ca-9c57cc56f-c8sn6" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.252013 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:44 crc kubenswrapper[4809]: E1206 05:53:44.252309 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:44.752298223 +0000 UTC m=+149.641281165 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.260275 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2kwh\" (UniqueName: \"kubernetes.io/projected/2a1ed741-489c-455f-a344-404bbab66c22-kube-api-access-t2kwh\") pod \"authentication-operator-69f744f599-7qpjs\" (UID: \"2a1ed741-489c-455f-a344-404bbab66c22\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7qpjs" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.276536 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-4pwcw"] Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.281919 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rj9g\" (UniqueName: \"kubernetes.io/projected/63b0cea8-1716-4c17-9f87-f10e776cf47f-kube-api-access-6rj9g\") pod \"etcd-operator-b45778765-drj7v\" (UID: \"63b0cea8-1716-4c17-9f87-f10e776cf47f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-drj7v" Dec 06 05:53:44 crc kubenswrapper[4809]: W1206 05:53:44.286052 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8e03364a_1840_4c91_8b2b_c7e2071fc73b.slice/crio-16846d1b1f5b3efa865f1863bb744eaa55819731e81fe580e571a1e290114936 WatchSource:0}: Error finding container 16846d1b1f5b3efa865f1863bb744eaa55819731e81fe580e571a1e290114936: Status 404 returned error can't find the container with id 16846d1b1f5b3efa865f1863bb744eaa55819731e81fe580e571a1e290114936 Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.296480 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-7qpjs" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.298014 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnqk2\" (UniqueName: \"kubernetes.io/projected/d233521e-75db-4da3-9cd3-0480a99eba90-kube-api-access-hnqk2\") pod \"migrator-59844c95c7-8ssrl\" (UID: \"d233521e-75db-4da3-9cd3-0480a99eba90\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8ssrl" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.324003 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjvtl\" (UniqueName: \"kubernetes.io/projected/d7d02f2c-c052-4320-bac4-394fb40c57d2-kube-api-access-tjvtl\") pod \"dns-operator-744455d44c-927z2\" (UID: \"d7d02f2c-c052-4320-bac4-394fb40c57d2\") " pod="openshift-dns-operator/dns-operator-744455d44c-927z2" Dec 06 05:53:44 crc kubenswrapper[4809]: E1206 05:53:44.331977 4809 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: failed to sync configmap cache: timed out waiting for the condition Dec 06 05:53:44 crc kubenswrapper[4809]: E1206 05:53:44.332057 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 05:55:46.332040205 +0000 UTC m=+271.221023147 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : failed to sync configmap cache: timed out waiting for the condition Dec 06 05:53:44 crc kubenswrapper[4809]: E1206 05:53:44.331977 4809 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: failed to sync secret cache: timed out waiting for the condition Dec 06 05:53:44 crc kubenswrapper[4809]: E1206 05:53:44.332105 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 05:55:46.332099007 +0000 UTC m=+271.221081959 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : failed to sync secret cache: timed out waiting for the condition Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.335129 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-927z2" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.352297 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vfkx\" (UniqueName: \"kubernetes.io/projected/d844d778-6b3d-400a-8f71-db1c2e643878-kube-api-access-9vfkx\") pod \"openshift-controller-manager-operator-756b6f6bc6-zj6j9\" (UID: \"d844d778-6b3d-400a-8f71-db1c2e643878\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zj6j9" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.353020 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:44 crc kubenswrapper[4809]: E1206 05:53:44.353873 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:44.853851492 +0000 UTC m=+149.742834434 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.359551 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhvfg\" (UniqueName: \"kubernetes.io/projected/cdf2d3c6-659f-4dc5-a21c-d29cc86ef239-kube-api-access-vhvfg\") pod \"machine-config-controller-84d6567774-wcth2\" (UID: \"cdf2d3c6-659f-4dc5-a21c-d29cc86ef239\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wcth2" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.383483 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-c8sn6" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.387333 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wh944\" (UniqueName: \"kubernetes.io/projected/e5356f1d-5afd-47bd-89af-72abfa8f901f-kube-api-access-wh944\") pod \"router-default-5444994796-n6z57\" (UID: \"e5356f1d-5afd-47bd-89af-72abfa8f901f\") " pod="openshift-ingress/router-default-5444994796-n6z57" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.392015 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9rgdg"] Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.393833 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-s7q7v" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.400586 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjbbp\" (UniqueName: \"kubernetes.io/projected/00bf9f05-7bc3-40e1-a2e9-1af1bf93f014-kube-api-access-pjbbp\") pod \"catalog-operator-68c6474976-pkq7h\" (UID: \"00bf9f05-7bc3-40e1-a2e9-1af1bf93f014\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pkq7h" Dec 06 05:53:44 crc kubenswrapper[4809]: W1206 05:53:44.403693 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode2885220_fafe_48da_9ed7_98942886ab1f.slice/crio-26e8dc55f563cb88ca71eeb0cffb270945a93248d0da5997f3c6e2162488cc7f WatchSource:0}: Error finding container 26e8dc55f563cb88ca71eeb0cffb270945a93248d0da5997f3c6e2162488cc7f: Status 404 returned error can't find the container with id 26e8dc55f563cb88ca71eeb0cffb270945a93248d0da5997f3c6e2162488cc7f Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.419728 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qt5sg\" (UniqueName: \"kubernetes.io/projected/f3a30985-f268-44de-93b9-bce0f9aebc6d-kube-api-access-qt5sg\") pod \"console-operator-58897d9998-hlvjb\" (UID: \"f3a30985-f268-44de-93b9-bce0f9aebc6d\") " pod="openshift-console-operator/console-operator-58897d9998-hlvjb" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.429824 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pkq7h" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.437149 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r7684"] Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.440155 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-hlvjb" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.444306 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9g7x\" (UniqueName: \"kubernetes.io/projected/d0b4b35c-3cb0-405d-941b-1b3946b228a0-kube-api-access-c9g7x\") pod \"cluster-image-registry-operator-dc59b4c8b-kr48t\" (UID: \"d0b4b35c-3cb0-405d-941b-1b3946b228a0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kr48t" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.446325 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wcth2" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.455197 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:44 crc kubenswrapper[4809]: E1206 05:53:44.455637 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:44.955623677 +0000 UTC m=+149.844606619 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.460546 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59tgt\" (UniqueName: \"kubernetes.io/projected/8e290142-b2fc-492e-8763-ff41224579fc-kube-api-access-59tgt\") pod \"downloads-7954f5f757-224mv\" (UID: \"8e290142-b2fc-492e-8763-ff41224579fc\") " pod="openshift-console/downloads-7954f5f757-224mv" Dec 06 05:53:44 crc kubenswrapper[4809]: W1206 05:53:44.467188 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcf124a04_e72c_4c67_bd76_52648b4d20c9.slice/crio-cd2b6959582e85000cb9d7715d3869c8fa94499e78f418a632ebf21fd345f94f WatchSource:0}: Error finding container cd2b6959582e85000cb9d7715d3869c8fa94499e78f418a632ebf21fd345f94f: Status 404 returned error can't find the container with id cd2b6959582e85000cb9d7715d3869c8fa94499e78f418a632ebf21fd345f94f Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.473490 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8ssrl" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.478767 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qw5pv\" (UniqueName: \"kubernetes.io/projected/b119ee35-1a33-4a34-9fca-62ad465a20f7-kube-api-access-qw5pv\") pod \"machine-config-server-g7pfc\" (UID: \"b119ee35-1a33-4a34-9fca-62ad465a20f7\") " pod="openshift-machine-config-operator/machine-config-server-g7pfc" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.482374 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-224mv" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.510879 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sdg4b"] Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.513173 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-kjlr8"] Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.513661 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wfkz\" (UniqueName: \"kubernetes.io/projected/e5e4e451-e3ee-434e-8e24-0c795925a48e-kube-api-access-9wfkz\") pod \"packageserver-d55dfcdfc-jjf9r\" (UID: \"e5e4e451-e3ee-434e-8e24-0c795925a48e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jjf9r" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.517889 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zj6j9" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.521062 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ggvcl"] Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.521219 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6nl2\" (UniqueName: \"kubernetes.io/projected/de0bb137-3c1d-4632-a7ef-6922f027a3c1-kube-api-access-v6nl2\") pod \"control-plane-machine-set-operator-78cbb6b69f-dpbfw\" (UID: \"de0bb137-3c1d-4632-a7ef-6922f027a3c1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dpbfw" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.544569 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bplj9\" (UniqueName: \"kubernetes.io/projected/c8d79e6f-40b0-4ecc-8501-454ba5cfe691-kube-api-access-bplj9\") pod \"olm-operator-6b444d44fb-gt9kb\" (UID: \"c8d79e6f-40b0-4ecc-8501-454ba5cfe691\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gt9kb" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.557577 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:44 crc kubenswrapper[4809]: E1206 05:53:44.558164 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:45.058145012 +0000 UTC m=+149.947127954 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.558791 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25sjn\" (UniqueName: \"kubernetes.io/projected/cea05156-8322-404a-8b33-867aeaaa905e-kube-api-access-25sjn\") pod \"service-ca-operator-777779d784-mxv9r\" (UID: \"cea05156-8322-404a-8b33-867aeaaa905e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mxv9r" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.571379 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-drj7v" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.582294 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9tmj\" (UniqueName: \"kubernetes.io/projected/8a225b5c-4ad5-4f08-a0cb-ae38a8fa83d4-kube-api-access-v9tmj\") pod \"openshift-config-operator-7777fb866f-qqv6x\" (UID: \"8a225b5c-4ad5-4f08-a0cb-ae38a8fa83d4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-qqv6x" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.584398 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.587968 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kr48t" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.593529 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-7qpjs"] Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.605037 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.615310 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-c8sn6"] Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.616265 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-n6z57" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.630744 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-927z2"] Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.659793 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gt9kb" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.661385 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/14f07a0b-5e47-4636-9f17-cdf6f327b39d-audit-dir\") pod \"apiserver-7bbb656c7d-cd2tv\" (UID: \"14f07a0b-5e47-4636-9f17-cdf6f327b39d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cd2tv" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.661472 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1f78456e-8b1a-472d-beb4-aabc4a911e01-auth-proxy-config\") pod \"machine-approver-56656f9798-znmpz\" (UID: \"1f78456e-8b1a-472d-beb4-aabc4a911e01\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-znmpz" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.661520 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0a9baebe-98e6-44be-ad16-882e47f09248-audit-dir\") pod \"oauth-openshift-558db77b4-jn446\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.661544 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-jn446\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.661568 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f6465618-cc08-4930-a290-0d1ffe135c98-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-wqnkb\" (UID: \"f6465618-cc08-4930-a290-0d1ffe135c98\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wqnkb" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.661609 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c892e996-27b7-4b8d-bd92-7eac2df7702b-secret-volume\") pod \"collect-profiles-29416665-lsfkp\" (UID: \"c892e996-27b7-4b8d-bd92-7eac2df7702b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416665-lsfkp" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.661639 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8063bdc5-b0d2-4a8b-91d2-12c3e1e5f938-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7sxnv\" (UID: \"8063bdc5-b0d2-4a8b-91d2-12c3e1e5f938\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7sxnv" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.661708 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fv8gt\" (UniqueName: \"kubernetes.io/projected/1f78456e-8b1a-472d-beb4-aabc4a911e01-kube-api-access-fv8gt\") pod \"machine-approver-56656f9798-znmpz\" (UID: \"1f78456e-8b1a-472d-beb4-aabc4a911e01\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-znmpz" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.661738 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/14f07a0b-5e47-4636-9f17-cdf6f327b39d-serving-cert\") pod \"apiserver-7bbb656c7d-cd2tv\" (UID: \"14f07a0b-5e47-4636-9f17-cdf6f327b39d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cd2tv" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.661955 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6465618-cc08-4930-a290-0d1ffe135c98-config\") pod \"kube-apiserver-operator-766d6c64bb-wqnkb\" (UID: \"f6465618-cc08-4930-a290-0d1ffe135c98\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wqnkb" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.662072 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8063bdc5-b0d2-4a8b-91d2-12c3e1e5f938-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7sxnv\" (UID: \"8063bdc5-b0d2-4a8b-91d2-12c3e1e5f938\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7sxnv" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.662143 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/14f07a0b-5e47-4636-9f17-cdf6f327b39d-encryption-config\") pod \"apiserver-7bbb656c7d-cd2tv\" (UID: \"14f07a0b-5e47-4636-9f17-cdf6f327b39d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cd2tv" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.662176 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-jn446\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.662207 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-jn446\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.662232 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09a7af3d-6bf2-4b7a-93d9-cecc73d993a1-config\") pod \"apiserver-76f77b778f-tphf5\" (UID: \"09a7af3d-6bf2-4b7a-93d9-cecc73d993a1\") " pod="openshift-apiserver/apiserver-76f77b778f-tphf5" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.662254 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/87db7503-6e3e-49b1-bc84-5230a444e8f6-auth-proxy-config\") pod \"machine-config-operator-74547568cd-nsqq5\" (UID: \"87db7503-6e3e-49b1-bc84-5230a444e8f6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nsqq5" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.662287 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45f3dc48-d71b-47f8-8567-90df740103b5-config\") pod \"route-controller-manager-6576b87f9c-nff6m\" (UID: \"45f3dc48-d71b-47f8-8567-90df740103b5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nff6m" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.662325 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8063bdc5-b0d2-4a8b-91d2-12c3e1e5f938-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7sxnv\" (UID: \"8063bdc5-b0d2-4a8b-91d2-12c3e1e5f938\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7sxnv" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.662357 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fhld\" (UniqueName: \"kubernetes.io/projected/45f3dc48-d71b-47f8-8567-90df740103b5-kube-api-access-7fhld\") pod \"route-controller-manager-6576b87f9c-nff6m\" (UID: \"45f3dc48-d71b-47f8-8567-90df740103b5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nff6m" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.662384 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-jn446\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.662413 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cztt\" (UniqueName: \"kubernetes.io/projected/14f07a0b-5e47-4636-9f17-cdf6f327b39d-kube-api-access-7cztt\") pod \"apiserver-7bbb656c7d-cd2tv\" (UID: \"14f07a0b-5e47-4636-9f17-cdf6f327b39d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cd2tv" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.662543 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/09a7af3d-6bf2-4b7a-93d9-cecc73d993a1-audit-dir\") pod \"apiserver-76f77b778f-tphf5\" (UID: \"09a7af3d-6bf2-4b7a-93d9-cecc73d993a1\") " pod="openshift-apiserver/apiserver-76f77b778f-tphf5" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.662571 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5q7r\" (UniqueName: \"kubernetes.io/projected/ee96792c-e168-4c58-a477-cd88c14da612-kube-api-access-c5q7r\") pod \"marketplace-operator-79b997595-slt9h\" (UID: \"ee96792c-e168-4c58-a477-cd88c14da612\") " pod="openshift-marketplace/marketplace-operator-79b997595-slt9h" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.662599 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktzqv\" (UniqueName: \"kubernetes.io/projected/6b48dffc-c51b-48e2-b22b-e809176c8830-kube-api-access-ktzqv\") pod \"machine-api-operator-5694c8668f-852bd\" (UID: \"6b48dffc-c51b-48e2-b22b-e809176c8830\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-852bd" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.662623 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6b48dffc-c51b-48e2-b22b-e809176c8830-images\") pod \"machine-api-operator-5694c8668f-852bd\" (UID: \"6b48dffc-c51b-48e2-b22b-e809176c8830\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-852bd" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.662645 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/09a7af3d-6bf2-4b7a-93d9-cecc73d993a1-image-import-ca\") pod \"apiserver-76f77b778f-tphf5\" (UID: \"09a7af3d-6bf2-4b7a-93d9-cecc73d993a1\") " pod="openshift-apiserver/apiserver-76f77b778f-tphf5" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.662672 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/1f78456e-8b1a-472d-beb4-aabc4a911e01-machine-approver-tls\") pod \"machine-approver-56656f9798-znmpz\" (UID: \"1f78456e-8b1a-472d-beb4-aabc4a911e01\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-znmpz" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.662739 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/45f3dc48-d71b-47f8-8567-90df740103b5-client-ca\") pod \"route-controller-manager-6576b87f9c-nff6m\" (UID: \"45f3dc48-d71b-47f8-8567-90df740103b5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nff6m" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.662819 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f78456e-8b1a-472d-beb4-aabc4a911e01-config\") pod \"machine-approver-56656f9798-znmpz\" (UID: \"1f78456e-8b1a-472d-beb4-aabc4a911e01\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-znmpz" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.662852 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f6465618-cc08-4930-a290-0d1ffe135c98-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-wqnkb\" (UID: \"f6465618-cc08-4930-a290-0d1ffe135c98\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wqnkb" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.663621 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ee96792c-e168-4c58-a477-cd88c14da612-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-slt9h\" (UID: \"ee96792c-e168-4c58-a477-cd88c14da612\") " pod="openshift-marketplace/marketplace-operator-79b997595-slt9h" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.663681 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/14f07a0b-5e47-4636-9f17-cdf6f327b39d-etcd-client\") pod \"apiserver-7bbb656c7d-cd2tv\" (UID: \"14f07a0b-5e47-4636-9f17-cdf6f327b39d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cd2tv" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.663880 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grzz4\" (UniqueName: \"kubernetes.io/projected/b6b25f1e-33d4-481d-8c73-24b95c61364f-kube-api-access-grzz4\") pod \"package-server-manager-789f6589d5-ch964\" (UID: \"b6b25f1e-33d4-481d-8c73-24b95c61364f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ch964" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.663917 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4g75z\" (UniqueName: \"kubernetes.io/projected/87db7503-6e3e-49b1-bc84-5230a444e8f6-kube-api-access-4g75z\") pod \"machine-config-operator-74547568cd-nsqq5\" (UID: \"87db7503-6e3e-49b1-bc84-5230a444e8f6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nsqq5" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.664009 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/14f07a0b-5e47-4636-9f17-cdf6f327b39d-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-cd2tv\" (UID: \"14f07a0b-5e47-4636-9f17-cdf6f327b39d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cd2tv" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.664046 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ee96792c-e168-4c58-a477-cd88c14da612-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-slt9h\" (UID: \"ee96792c-e168-4c58-a477-cd88c14da612\") " pod="openshift-marketplace/marketplace-operator-79b997595-slt9h" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.664071 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/09a7af3d-6bf2-4b7a-93d9-cecc73d993a1-audit\") pod \"apiserver-76f77b778f-tphf5\" (UID: \"09a7af3d-6bf2-4b7a-93d9-cecc73d993a1\") " pod="openshift-apiserver/apiserver-76f77b778f-tphf5" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.664397 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nczzr\" (UniqueName: \"kubernetes.io/projected/0a9baebe-98e6-44be-ad16-882e47f09248-kube-api-access-nczzr\") pod \"oauth-openshift-558db77b4-jn446\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.664490 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7d8w\" (UniqueName: \"kubernetes.io/projected/d0367f91-0d22-4fa6-8cfe-0b549698b42b-kube-api-access-l7d8w\") pod \"cluster-samples-operator-665b6dd947-m6kjl\" (UID: \"d0367f91-0d22-4fa6-8cfe-0b549698b42b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-m6kjl" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.664666 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/87db7503-6e3e-49b1-bc84-5230a444e8f6-proxy-tls\") pod \"machine-config-operator-74547568cd-nsqq5\" (UID: \"87db7503-6e3e-49b1-bc84-5230a444e8f6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nsqq5" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.665490 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-jn446\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.665577 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-jn446\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.665663 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/b6b25f1e-33d4-481d-8c73-24b95c61364f-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-ch964\" (UID: \"b6b25f1e-33d4-481d-8c73-24b95c61364f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ch964" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.665709 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09a7af3d-6bf2-4b7a-93d9-cecc73d993a1-etcd-client\") pod \"apiserver-76f77b778f-tphf5\" (UID: \"09a7af3d-6bf2-4b7a-93d9-cecc73d993a1\") " pod="openshift-apiserver/apiserver-76f77b778f-tphf5" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.665880 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.666339 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/14f07a0b-5e47-4636-9f17-cdf6f327b39d-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-cd2tv\" (UID: \"14f07a0b-5e47-4636-9f17-cdf6f327b39d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cd2tv" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.666394 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b48dffc-c51b-48e2-b22b-e809176c8830-config\") pod \"machine-api-operator-5694c8668f-852bd\" (UID: \"6b48dffc-c51b-48e2-b22b-e809176c8830\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-852bd" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.666462 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-jn446\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.666492 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6b48dffc-c51b-48e2-b22b-e809176c8830-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-852bd\" (UID: \"6b48dffc-c51b-48e2-b22b-e809176c8830\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-852bd" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.666522 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/d0367f91-0d22-4fa6-8cfe-0b549698b42b-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-m6kjl\" (UID: \"d0367f91-0d22-4fa6-8cfe-0b549698b42b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-m6kjl" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.666595 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-jn446\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.666668 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/45f3dc48-d71b-47f8-8567-90df740103b5-serving-cert\") pod \"route-controller-manager-6576b87f9c-nff6m\" (UID: \"45f3dc48-d71b-47f8-8567-90df740103b5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nff6m" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.666741 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09a7af3d-6bf2-4b7a-93d9-cecc73d993a1-encryption-config\") pod \"apiserver-76f77b778f-tphf5\" (UID: \"09a7af3d-6bf2-4b7a-93d9-cecc73d993a1\") " pod="openshift-apiserver/apiserver-76f77b778f-tphf5" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.666777 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxk8g\" (UniqueName: \"kubernetes.io/projected/c892e996-27b7-4b8d-bd92-7eac2df7702b-kube-api-access-vxk8g\") pod \"collect-profiles-29416665-lsfkp\" (UID: \"c892e996-27b7-4b8d-bd92-7eac2df7702b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416665-lsfkp" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.666833 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09a7af3d-6bf2-4b7a-93d9-cecc73d993a1-trusted-ca-bundle\") pod \"apiserver-76f77b778f-tphf5\" (UID: \"09a7af3d-6bf2-4b7a-93d9-cecc73d993a1\") " pod="openshift-apiserver/apiserver-76f77b778f-tphf5" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.667028 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09a7af3d-6bf2-4b7a-93d9-cecc73d993a1-etcd-serving-ca\") pod \"apiserver-76f77b778f-tphf5\" (UID: \"09a7af3d-6bf2-4b7a-93d9-cecc73d993a1\") " pod="openshift-apiserver/apiserver-76f77b778f-tphf5" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.667086 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/87db7503-6e3e-49b1-bc84-5230a444e8f6-images\") pod \"machine-config-operator-74547568cd-nsqq5\" (UID: \"87db7503-6e3e-49b1-bc84-5230a444e8f6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nsqq5" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.667146 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09a7af3d-6bf2-4b7a-93d9-cecc73d993a1-serving-cert\") pod \"apiserver-76f77b778f-tphf5\" (UID: \"09a7af3d-6bf2-4b7a-93d9-cecc73d993a1\") " pod="openshift-apiserver/apiserver-76f77b778f-tphf5" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.667175 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/09a7af3d-6bf2-4b7a-93d9-cecc73d993a1-node-pullsecrets\") pod \"apiserver-76f77b778f-tphf5\" (UID: \"09a7af3d-6bf2-4b7a-93d9-cecc73d993a1\") " pod="openshift-apiserver/apiserver-76f77b778f-tphf5" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.667209 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/14f07a0b-5e47-4636-9f17-cdf6f327b39d-audit-policies\") pod \"apiserver-7bbb656c7d-cd2tv\" (UID: \"14f07a0b-5e47-4636-9f17-cdf6f327b39d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cd2tv" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.667255 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-jn446\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:53:44 crc kubenswrapper[4809]: E1206 05:53:44.667532 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:45.167514564 +0000 UTC m=+150.056497696 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.668054 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-jn446\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.668081 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-jn446\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.668114 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0a9baebe-98e6-44be-ad16-882e47f09248-audit-policies\") pod \"oauth-openshift-558db77b4-jn446\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.668158 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dd4pd\" (UniqueName: \"kubernetes.io/projected/09a7af3d-6bf2-4b7a-93d9-cecc73d993a1-kube-api-access-dd4pd\") pod \"apiserver-76f77b778f-tphf5\" (UID: \"09a7af3d-6bf2-4b7a-93d9-cecc73d993a1\") " pod="openshift-apiserver/apiserver-76f77b778f-tphf5" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.668182 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c892e996-27b7-4b8d-bd92-7eac2df7702b-config-volume\") pod \"collect-profiles-29416665-lsfkp\" (UID: \"c892e996-27b7-4b8d-bd92-7eac2df7702b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416665-lsfkp" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.673024 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-s7q7v"] Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.678324 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qqv6x" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.713524 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pkq7h"] Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.720393 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jjf9r" Dec 06 05:53:44 crc kubenswrapper[4809]: W1206 05:53:44.730614 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2a1ed741_489c_455f_a344_404bbab66c22.slice/crio-c9c9298faabafb75977a5c6feec58035eadc9d3b13b143a27316776caab0c1d2 WatchSource:0}: Error finding container c9c9298faabafb75977a5c6feec58035eadc9d3b13b143a27316776caab0c1d2: Status 404 returned error can't find the container with id c9c9298faabafb75977a5c6feec58035eadc9d3b13b143a27316776caab0c1d2 Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.751972 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-g7pfc" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.757966 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dpbfw" Dec 06 05:53:44 crc kubenswrapper[4809]: W1206 05:53:44.768087 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4008f48_e17b_492f_8fa4_b09fb9d2ce70.slice/crio-545345580be0fae589076dca26ab6689a1d92c29693095bb2fd77ffec72ec008 WatchSource:0}: Error finding container 545345580be0fae589076dca26ab6689a1d92c29693095bb2fd77ffec72ec008: Status 404 returned error can't find the container with id 545345580be0fae589076dca26ab6689a1d92c29693095bb2fd77ffec72ec008 Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.774829 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.775216 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/c598f580-a438-4015-9a8a-f98b3b2b0b3d-csi-data-dir\") pod \"csi-hostpathplugin-p2lkt\" (UID: \"c598f580-a438-4015-9a8a-f98b3b2b0b3d\") " pod="hostpath-provisioner/csi-hostpathplugin-p2lkt" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.775279 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/45f3dc48-d71b-47f8-8567-90df740103b5-serving-cert\") pod \"route-controller-manager-6576b87f9c-nff6m\" (UID: \"45f3dc48-d71b-47f8-8567-90df740103b5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nff6m" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.775304 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/c598f580-a438-4015-9a8a-f98b3b2b0b3d-registration-dir\") pod \"csi-hostpathplugin-p2lkt\" (UID: \"c598f580-a438-4015-9a8a-f98b3b2b0b3d\") " pod="hostpath-provisioner/csi-hostpathplugin-p2lkt" Dec 06 05:53:44 crc kubenswrapper[4809]: E1206 05:53:44.775342 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:45.275307143 +0000 UTC m=+150.164290145 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.775402 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09a7af3d-6bf2-4b7a-93d9-cecc73d993a1-encryption-config\") pod \"apiserver-76f77b778f-tphf5\" (UID: \"09a7af3d-6bf2-4b7a-93d9-cecc73d993a1\") " pod="openshift-apiserver/apiserver-76f77b778f-tphf5" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.775456 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxk8g\" (UniqueName: \"kubernetes.io/projected/c892e996-27b7-4b8d-bd92-7eac2df7702b-kube-api-access-vxk8g\") pod \"collect-profiles-29416665-lsfkp\" (UID: \"c892e996-27b7-4b8d-bd92-7eac2df7702b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416665-lsfkp" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.775489 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mm7c2\" (UniqueName: \"kubernetes.io/projected/e1380457-61e9-44d2-8324-eb1bcb638056-kube-api-access-mm7c2\") pod \"ingress-canary-czhdd\" (UID: \"e1380457-61e9-44d2-8324-eb1bcb638056\") " pod="openshift-ingress-canary/ingress-canary-czhdd" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.775541 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09a7af3d-6bf2-4b7a-93d9-cecc73d993a1-trusted-ca-bundle\") pod \"apiserver-76f77b778f-tphf5\" (UID: \"09a7af3d-6bf2-4b7a-93d9-cecc73d993a1\") " pod="openshift-apiserver/apiserver-76f77b778f-tphf5" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.775967 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/c598f580-a438-4015-9a8a-f98b3b2b0b3d-plugins-dir\") pod \"csi-hostpathplugin-p2lkt\" (UID: \"c598f580-a438-4015-9a8a-f98b3b2b0b3d\") " pod="hostpath-provisioner/csi-hostpathplugin-p2lkt" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.776020 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09a7af3d-6bf2-4b7a-93d9-cecc73d993a1-etcd-serving-ca\") pod \"apiserver-76f77b778f-tphf5\" (UID: \"09a7af3d-6bf2-4b7a-93d9-cecc73d993a1\") " pod="openshift-apiserver/apiserver-76f77b778f-tphf5" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.776042 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/87db7503-6e3e-49b1-bc84-5230a444e8f6-images\") pod \"machine-config-operator-74547568cd-nsqq5\" (UID: \"87db7503-6e3e-49b1-bc84-5230a444e8f6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nsqq5" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.776069 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09a7af3d-6bf2-4b7a-93d9-cecc73d993a1-serving-cert\") pod \"apiserver-76f77b778f-tphf5\" (UID: \"09a7af3d-6bf2-4b7a-93d9-cecc73d993a1\") " pod="openshift-apiserver/apiserver-76f77b778f-tphf5" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.776107 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/14f07a0b-5e47-4636-9f17-cdf6f327b39d-audit-policies\") pod \"apiserver-7bbb656c7d-cd2tv\" (UID: \"14f07a0b-5e47-4636-9f17-cdf6f327b39d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cd2tv" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.776141 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-jn446\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.776164 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/09a7af3d-6bf2-4b7a-93d9-cecc73d993a1-node-pullsecrets\") pod \"apiserver-76f77b778f-tphf5\" (UID: \"09a7af3d-6bf2-4b7a-93d9-cecc73d993a1\") " pod="openshift-apiserver/apiserver-76f77b778f-tphf5" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.776191 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-jn446\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.776217 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-jn446\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.776263 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0a9baebe-98e6-44be-ad16-882e47f09248-audit-policies\") pod \"oauth-openshift-558db77b4-jn446\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.776315 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dd4pd\" (UniqueName: \"kubernetes.io/projected/09a7af3d-6bf2-4b7a-93d9-cecc73d993a1-kube-api-access-dd4pd\") pod \"apiserver-76f77b778f-tphf5\" (UID: \"09a7af3d-6bf2-4b7a-93d9-cecc73d993a1\") " pod="openshift-apiserver/apiserver-76f77b778f-tphf5" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.776343 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c892e996-27b7-4b8d-bd92-7eac2df7702b-config-volume\") pod \"collect-profiles-29416665-lsfkp\" (UID: \"c892e996-27b7-4b8d-bd92-7eac2df7702b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416665-lsfkp" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.776389 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/14f07a0b-5e47-4636-9f17-cdf6f327b39d-audit-dir\") pod \"apiserver-7bbb656c7d-cd2tv\" (UID: \"14f07a0b-5e47-4636-9f17-cdf6f327b39d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cd2tv" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.776467 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1f78456e-8b1a-472d-beb4-aabc4a911e01-auth-proxy-config\") pod \"machine-approver-56656f9798-znmpz\" (UID: \"1f78456e-8b1a-472d-beb4-aabc4a911e01\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-znmpz" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.776489 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0a9baebe-98e6-44be-ad16-882e47f09248-audit-dir\") pod \"oauth-openshift-558db77b4-jn446\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.776509 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-jn446\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.776837 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f6465618-cc08-4930-a290-0d1ffe135c98-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-wqnkb\" (UID: \"f6465618-cc08-4930-a290-0d1ffe135c98\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wqnkb" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.776872 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c892e996-27b7-4b8d-bd92-7eac2df7702b-secret-volume\") pod \"collect-profiles-29416665-lsfkp\" (UID: \"c892e996-27b7-4b8d-bd92-7eac2df7702b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416665-lsfkp" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.776899 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8063bdc5-b0d2-4a8b-91d2-12c3e1e5f938-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7sxnv\" (UID: \"8063bdc5-b0d2-4a8b-91d2-12c3e1e5f938\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7sxnv" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.776919 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fv8gt\" (UniqueName: \"kubernetes.io/projected/1f78456e-8b1a-472d-beb4-aabc4a911e01-kube-api-access-fv8gt\") pod \"machine-approver-56656f9798-znmpz\" (UID: \"1f78456e-8b1a-472d-beb4-aabc4a911e01\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-znmpz" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.777004 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/14f07a0b-5e47-4636-9f17-cdf6f327b39d-serving-cert\") pod \"apiserver-7bbb656c7d-cd2tv\" (UID: \"14f07a0b-5e47-4636-9f17-cdf6f327b39d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cd2tv" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.782501 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6465618-cc08-4930-a290-0d1ffe135c98-config\") pod \"kube-apiserver-operator-766d6c64bb-wqnkb\" (UID: \"f6465618-cc08-4930-a290-0d1ffe135c98\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wqnkb" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.782722 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8063bdc5-b0d2-4a8b-91d2-12c3e1e5f938-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7sxnv\" (UID: \"8063bdc5-b0d2-4a8b-91d2-12c3e1e5f938\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7sxnv" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.782750 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/14f07a0b-5e47-4636-9f17-cdf6f327b39d-encryption-config\") pod \"apiserver-7bbb656c7d-cd2tv\" (UID: \"14f07a0b-5e47-4636-9f17-cdf6f327b39d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cd2tv" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.782813 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-jn446\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.782846 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-jn446\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.782880 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09a7af3d-6bf2-4b7a-93d9-cecc73d993a1-config\") pod \"apiserver-76f77b778f-tphf5\" (UID: \"09a7af3d-6bf2-4b7a-93d9-cecc73d993a1\") " pod="openshift-apiserver/apiserver-76f77b778f-tphf5" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.782917 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/87db7503-6e3e-49b1-bc84-5230a444e8f6-auth-proxy-config\") pod \"machine-config-operator-74547568cd-nsqq5\" (UID: \"87db7503-6e3e-49b1-bc84-5230a444e8f6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nsqq5" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.783167 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45f3dc48-d71b-47f8-8567-90df740103b5-config\") pod \"route-controller-manager-6576b87f9c-nff6m\" (UID: \"45f3dc48-d71b-47f8-8567-90df740103b5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nff6m" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.783207 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpc9v\" (UniqueName: \"kubernetes.io/projected/c598f580-a438-4015-9a8a-f98b3b2b0b3d-kube-api-access-gpc9v\") pod \"csi-hostpathplugin-p2lkt\" (UID: \"c598f580-a438-4015-9a8a-f98b3b2b0b3d\") " pod="hostpath-provisioner/csi-hostpathplugin-p2lkt" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.783256 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e1380457-61e9-44d2-8324-eb1bcb638056-cert\") pod \"ingress-canary-czhdd\" (UID: \"e1380457-61e9-44d2-8324-eb1bcb638056\") " pod="openshift-ingress-canary/ingress-canary-czhdd" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.783284 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0a9a3552-d735-4692-ab80-2c1526680ab9-metrics-tls\") pod \"dns-default-s7px9\" (UID: \"0a9a3552-d735-4692-ab80-2c1526680ab9\") " pod="openshift-dns/dns-default-s7px9" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.783364 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8063bdc5-b0d2-4a8b-91d2-12c3e1e5f938-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7sxnv\" (UID: \"8063bdc5-b0d2-4a8b-91d2-12c3e1e5f938\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7sxnv" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.783392 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fhld\" (UniqueName: \"kubernetes.io/projected/45f3dc48-d71b-47f8-8567-90df740103b5-kube-api-access-7fhld\") pod \"route-controller-manager-6576b87f9c-nff6m\" (UID: \"45f3dc48-d71b-47f8-8567-90df740103b5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nff6m" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.783415 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-jn446\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.783439 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6ft7\" (UniqueName: \"kubernetes.io/projected/0a9a3552-d735-4692-ab80-2c1526680ab9-kube-api-access-s6ft7\") pod \"dns-default-s7px9\" (UID: \"0a9a3552-d735-4692-ab80-2c1526680ab9\") " pod="openshift-dns/dns-default-s7px9" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.783486 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cztt\" (UniqueName: \"kubernetes.io/projected/14f07a0b-5e47-4636-9f17-cdf6f327b39d-kube-api-access-7cztt\") pod \"apiserver-7bbb656c7d-cd2tv\" (UID: \"14f07a0b-5e47-4636-9f17-cdf6f327b39d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cd2tv" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.783509 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/09a7af3d-6bf2-4b7a-93d9-cecc73d993a1-audit-dir\") pod \"apiserver-76f77b778f-tphf5\" (UID: \"09a7af3d-6bf2-4b7a-93d9-cecc73d993a1\") " pod="openshift-apiserver/apiserver-76f77b778f-tphf5" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.783554 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktzqv\" (UniqueName: \"kubernetes.io/projected/6b48dffc-c51b-48e2-b22b-e809176c8830-kube-api-access-ktzqv\") pod \"machine-api-operator-5694c8668f-852bd\" (UID: \"6b48dffc-c51b-48e2-b22b-e809176c8830\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-852bd" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.783578 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5q7r\" (UniqueName: \"kubernetes.io/projected/ee96792c-e168-4c58-a477-cd88c14da612-kube-api-access-c5q7r\") pod \"marketplace-operator-79b997595-slt9h\" (UID: \"ee96792c-e168-4c58-a477-cd88c14da612\") " pod="openshift-marketplace/marketplace-operator-79b997595-slt9h" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.783641 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6b48dffc-c51b-48e2-b22b-e809176c8830-images\") pod \"machine-api-operator-5694c8668f-852bd\" (UID: \"6b48dffc-c51b-48e2-b22b-e809176c8830\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-852bd" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.783665 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/09a7af3d-6bf2-4b7a-93d9-cecc73d993a1-image-import-ca\") pod \"apiserver-76f77b778f-tphf5\" (UID: \"09a7af3d-6bf2-4b7a-93d9-cecc73d993a1\") " pod="openshift-apiserver/apiserver-76f77b778f-tphf5" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.783723 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/1f78456e-8b1a-472d-beb4-aabc4a911e01-machine-approver-tls\") pod \"machine-approver-56656f9798-znmpz\" (UID: \"1f78456e-8b1a-472d-beb4-aabc4a911e01\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-znmpz" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.783745 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/45f3dc48-d71b-47f8-8567-90df740103b5-client-ca\") pod \"route-controller-manager-6576b87f9c-nff6m\" (UID: \"45f3dc48-d71b-47f8-8567-90df740103b5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nff6m" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.783800 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/c598f580-a438-4015-9a8a-f98b3b2b0b3d-mountpoint-dir\") pod \"csi-hostpathplugin-p2lkt\" (UID: \"c598f580-a438-4015-9a8a-f98b3b2b0b3d\") " pod="hostpath-provisioner/csi-hostpathplugin-p2lkt" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.783882 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f78456e-8b1a-472d-beb4-aabc4a911e01-config\") pod \"machine-approver-56656f9798-znmpz\" (UID: \"1f78456e-8b1a-472d-beb4-aabc4a911e01\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-znmpz" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.783911 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f6465618-cc08-4930-a290-0d1ffe135c98-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-wqnkb\" (UID: \"f6465618-cc08-4930-a290-0d1ffe135c98\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wqnkb" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.783945 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0a9a3552-d735-4692-ab80-2c1526680ab9-config-volume\") pod \"dns-default-s7px9\" (UID: \"0a9a3552-d735-4692-ab80-2c1526680ab9\") " pod="openshift-dns/dns-default-s7px9" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.784055 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ee96792c-e168-4c58-a477-cd88c14da612-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-slt9h\" (UID: \"ee96792c-e168-4c58-a477-cd88c14da612\") " pod="openshift-marketplace/marketplace-operator-79b997595-slt9h" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.784087 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/14f07a0b-5e47-4636-9f17-cdf6f327b39d-etcd-client\") pod \"apiserver-7bbb656c7d-cd2tv\" (UID: \"14f07a0b-5e47-4636-9f17-cdf6f327b39d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cd2tv" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.784109 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/c598f580-a438-4015-9a8a-f98b3b2b0b3d-socket-dir\") pod \"csi-hostpathplugin-p2lkt\" (UID: \"c598f580-a438-4015-9a8a-f98b3b2b0b3d\") " pod="hostpath-provisioner/csi-hostpathplugin-p2lkt" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.784135 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grzz4\" (UniqueName: \"kubernetes.io/projected/b6b25f1e-33d4-481d-8c73-24b95c61364f-kube-api-access-grzz4\") pod \"package-server-manager-789f6589d5-ch964\" (UID: \"b6b25f1e-33d4-481d-8c73-24b95c61364f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ch964" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.784156 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4g75z\" (UniqueName: \"kubernetes.io/projected/87db7503-6e3e-49b1-bc84-5230a444e8f6-kube-api-access-4g75z\") pod \"machine-config-operator-74547568cd-nsqq5\" (UID: \"87db7503-6e3e-49b1-bc84-5230a444e8f6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nsqq5" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.784231 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/14f07a0b-5e47-4636-9f17-cdf6f327b39d-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-cd2tv\" (UID: \"14f07a0b-5e47-4636-9f17-cdf6f327b39d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cd2tv" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.784249 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ee96792c-e168-4c58-a477-cd88c14da612-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-slt9h\" (UID: \"ee96792c-e168-4c58-a477-cd88c14da612\") " pod="openshift-marketplace/marketplace-operator-79b997595-slt9h" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.784274 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/09a7af3d-6bf2-4b7a-93d9-cecc73d993a1-audit\") pod \"apiserver-76f77b778f-tphf5\" (UID: \"09a7af3d-6bf2-4b7a-93d9-cecc73d993a1\") " pod="openshift-apiserver/apiserver-76f77b778f-tphf5" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.784312 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nczzr\" (UniqueName: \"kubernetes.io/projected/0a9baebe-98e6-44be-ad16-882e47f09248-kube-api-access-nczzr\") pod \"oauth-openshift-558db77b4-jn446\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.784331 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7d8w\" (UniqueName: \"kubernetes.io/projected/d0367f91-0d22-4fa6-8cfe-0b549698b42b-kube-api-access-l7d8w\") pod \"cluster-samples-operator-665b6dd947-m6kjl\" (UID: \"d0367f91-0d22-4fa6-8cfe-0b549698b42b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-m6kjl" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.784392 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/87db7503-6e3e-49b1-bc84-5230a444e8f6-proxy-tls\") pod \"machine-config-operator-74547568cd-nsqq5\" (UID: \"87db7503-6e3e-49b1-bc84-5230a444e8f6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nsqq5" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.784463 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-jn446\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.784497 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-jn446\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.784531 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.784556 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/b6b25f1e-33d4-481d-8c73-24b95c61364f-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-ch964\" (UID: \"b6b25f1e-33d4-481d-8c73-24b95c61364f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ch964" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.784580 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09a7af3d-6bf2-4b7a-93d9-cecc73d993a1-etcd-client\") pod \"apiserver-76f77b778f-tphf5\" (UID: \"09a7af3d-6bf2-4b7a-93d9-cecc73d993a1\") " pod="openshift-apiserver/apiserver-76f77b778f-tphf5" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.784617 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/14f07a0b-5e47-4636-9f17-cdf6f327b39d-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-cd2tv\" (UID: \"14f07a0b-5e47-4636-9f17-cdf6f327b39d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cd2tv" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.784640 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b48dffc-c51b-48e2-b22b-e809176c8830-config\") pod \"machine-api-operator-5694c8668f-852bd\" (UID: \"6b48dffc-c51b-48e2-b22b-e809176c8830\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-852bd" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.784661 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-jn446\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.784699 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6b48dffc-c51b-48e2-b22b-e809176c8830-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-852bd\" (UID: \"6b48dffc-c51b-48e2-b22b-e809176c8830\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-852bd" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.784722 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/d0367f91-0d22-4fa6-8cfe-0b549698b42b-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-m6kjl\" (UID: \"d0367f91-0d22-4fa6-8cfe-0b549698b42b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-m6kjl" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.784777 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-jn446\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.788440 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-jn446\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.790350 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8063bdc5-b0d2-4a8b-91d2-12c3e1e5f938-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7sxnv\" (UID: \"8063bdc5-b0d2-4a8b-91d2-12c3e1e5f938\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7sxnv" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.797322 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09a7af3d-6bf2-4b7a-93d9-cecc73d993a1-serving-cert\") pod \"apiserver-76f77b778f-tphf5\" (UID: \"09a7af3d-6bf2-4b7a-93d9-cecc73d993a1\") " pod="openshift-apiserver/apiserver-76f77b778f-tphf5" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.799058 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/14f07a0b-5e47-4636-9f17-cdf6f327b39d-encryption-config\") pod \"apiserver-7bbb656c7d-cd2tv\" (UID: \"14f07a0b-5e47-4636-9f17-cdf6f327b39d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cd2tv" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.799647 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c892e996-27b7-4b8d-bd92-7eac2df7702b-secret-volume\") pod \"collect-profiles-29416665-lsfkp\" (UID: \"c892e996-27b7-4b8d-bd92-7eac2df7702b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416665-lsfkp" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.800130 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/14f07a0b-5e47-4636-9f17-cdf6f327b39d-audit-dir\") pod \"apiserver-7bbb656c7d-cd2tv\" (UID: \"14f07a0b-5e47-4636-9f17-cdf6f327b39d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cd2tv" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.800428 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c892e996-27b7-4b8d-bd92-7eac2df7702b-config-volume\") pod \"collect-profiles-29416665-lsfkp\" (UID: \"c892e996-27b7-4b8d-bd92-7eac2df7702b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416665-lsfkp" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.802181 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09a7af3d-6bf2-4b7a-93d9-cecc73d993a1-encryption-config\") pod \"apiserver-76f77b778f-tphf5\" (UID: \"09a7af3d-6bf2-4b7a-93d9-cecc73d993a1\") " pod="openshift-apiserver/apiserver-76f77b778f-tphf5" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.802646 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1f78456e-8b1a-472d-beb4-aabc4a911e01-auth-proxy-config\") pod \"machine-approver-56656f9798-znmpz\" (UID: \"1f78456e-8b1a-472d-beb4-aabc4a911e01\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-znmpz" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.802716 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0a9baebe-98e6-44be-ad16-882e47f09248-audit-dir\") pod \"oauth-openshift-558db77b4-jn446\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.803382 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-jn446\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.803456 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8063bdc5-b0d2-4a8b-91d2-12c3e1e5f938-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7sxnv\" (UID: \"8063bdc5-b0d2-4a8b-91d2-12c3e1e5f938\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7sxnv" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.804689 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6465618-cc08-4930-a290-0d1ffe135c98-config\") pod \"kube-apiserver-operator-766d6c64bb-wqnkb\" (UID: \"f6465618-cc08-4930-a290-0d1ffe135c98\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wqnkb" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.805979 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/14f07a0b-5e47-4636-9f17-cdf6f327b39d-serving-cert\") pod \"apiserver-7bbb656c7d-cd2tv\" (UID: \"14f07a0b-5e47-4636-9f17-cdf6f327b39d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cd2tv" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.806161 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/09a7af3d-6bf2-4b7a-93d9-cecc73d993a1-node-pullsecrets\") pod \"apiserver-76f77b778f-tphf5\" (UID: \"09a7af3d-6bf2-4b7a-93d9-cecc73d993a1\") " pod="openshift-apiserver/apiserver-76f77b778f-tphf5" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.806394 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09a7af3d-6bf2-4b7a-93d9-cecc73d993a1-trusted-ca-bundle\") pod \"apiserver-76f77b778f-tphf5\" (UID: \"09a7af3d-6bf2-4b7a-93d9-cecc73d993a1\") " pod="openshift-apiserver/apiserver-76f77b778f-tphf5" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.809347 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/87db7503-6e3e-49b1-bc84-5230a444e8f6-images\") pod \"machine-config-operator-74547568cd-nsqq5\" (UID: \"87db7503-6e3e-49b1-bc84-5230a444e8f6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nsqq5" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.809900 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/14f07a0b-5e47-4636-9f17-cdf6f327b39d-audit-policies\") pod \"apiserver-7bbb656c7d-cd2tv\" (UID: \"14f07a0b-5e47-4636-9f17-cdf6f327b39d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cd2tv" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.810125 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09a7af3d-6bf2-4b7a-93d9-cecc73d993a1-etcd-serving-ca\") pod \"apiserver-76f77b778f-tphf5\" (UID: \"09a7af3d-6bf2-4b7a-93d9-cecc73d993a1\") " pod="openshift-apiserver/apiserver-76f77b778f-tphf5" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.810710 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-jn446\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.809917 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0a9baebe-98e6-44be-ad16-882e47f09248-audit-policies\") pod \"oauth-openshift-558db77b4-jn446\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.810849 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/45f3dc48-d71b-47f8-8567-90df740103b5-serving-cert\") pod \"route-controller-manager-6576b87f9c-nff6m\" (UID: \"45f3dc48-d71b-47f8-8567-90df740103b5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nff6m" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.813728 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-mxv9r" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.813750 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/14f07a0b-5e47-4636-9f17-cdf6f327b39d-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-cd2tv\" (UID: \"14f07a0b-5e47-4636-9f17-cdf6f327b39d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cd2tv" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.818607 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/09a7af3d-6bf2-4b7a-93d9-cecc73d993a1-audit\") pod \"apiserver-76f77b778f-tphf5\" (UID: \"09a7af3d-6bf2-4b7a-93d9-cecc73d993a1\") " pod="openshift-apiserver/apiserver-76f77b778f-tphf5" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.819074 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/87db7503-6e3e-49b1-bc84-5230a444e8f6-auth-proxy-config\") pod \"machine-config-operator-74547568cd-nsqq5\" (UID: \"87db7503-6e3e-49b1-bc84-5230a444e8f6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nsqq5" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.833449 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/14f07a0b-5e47-4636-9f17-cdf6f327b39d-etcd-client\") pod \"apiserver-7bbb656c7d-cd2tv\" (UID: \"14f07a0b-5e47-4636-9f17-cdf6f327b39d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cd2tv" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.833457 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/09a7af3d-6bf2-4b7a-93d9-cecc73d993a1-audit-dir\") pod \"apiserver-76f77b778f-tphf5\" (UID: \"09a7af3d-6bf2-4b7a-93d9-cecc73d993a1\") " pod="openshift-apiserver/apiserver-76f77b778f-tphf5" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.833479 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09a7af3d-6bf2-4b7a-93d9-cecc73d993a1-etcd-client\") pod \"apiserver-76f77b778f-tphf5\" (UID: \"09a7af3d-6bf2-4b7a-93d9-cecc73d993a1\") " pod="openshift-apiserver/apiserver-76f77b778f-tphf5" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.834070 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/14f07a0b-5e47-4636-9f17-cdf6f327b39d-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-cd2tv\" (UID: \"14f07a0b-5e47-4636-9f17-cdf6f327b39d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cd2tv" Dec 06 05:53:44 crc kubenswrapper[4809]: E1206 05:53:44.834360 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:45.334341508 +0000 UTC m=+150.223324450 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.834368 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b48dffc-c51b-48e2-b22b-e809176c8830-config\") pod \"machine-api-operator-5694c8668f-852bd\" (UID: \"6b48dffc-c51b-48e2-b22b-e809176c8830\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-852bd" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.834737 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45f3dc48-d71b-47f8-8567-90df740103b5-config\") pod \"route-controller-manager-6576b87f9c-nff6m\" (UID: \"45f3dc48-d71b-47f8-8567-90df740103b5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nff6m" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.836035 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/09a7af3d-6bf2-4b7a-93d9-cecc73d993a1-image-import-ca\") pod \"apiserver-76f77b778f-tphf5\" (UID: \"09a7af3d-6bf2-4b7a-93d9-cecc73d993a1\") " pod="openshift-apiserver/apiserver-76f77b778f-tphf5" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.836427 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-jn446\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.836675 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09a7af3d-6bf2-4b7a-93d9-cecc73d993a1-config\") pod \"apiserver-76f77b778f-tphf5\" (UID: \"09a7af3d-6bf2-4b7a-93d9-cecc73d993a1\") " pod="openshift-apiserver/apiserver-76f77b778f-tphf5" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.836896 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-jn446\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.837065 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6b48dffc-c51b-48e2-b22b-e809176c8830-images\") pod \"machine-api-operator-5694c8668f-852bd\" (UID: \"6b48dffc-c51b-48e2-b22b-e809176c8830\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-852bd" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.837963 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ee96792c-e168-4c58-a477-cd88c14da612-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-slt9h\" (UID: \"ee96792c-e168-4c58-a477-cd88c14da612\") " pod="openshift-marketplace/marketplace-operator-79b997595-slt9h" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.838213 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f78456e-8b1a-472d-beb4-aabc4a911e01-config\") pod \"machine-approver-56656f9798-znmpz\" (UID: \"1f78456e-8b1a-472d-beb4-aabc4a911e01\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-znmpz" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.838435 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ee96792c-e168-4c58-a477-cd88c14da612-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-slt9h\" (UID: \"ee96792c-e168-4c58-a477-cd88c14da612\") " pod="openshift-marketplace/marketplace-operator-79b997595-slt9h" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.839232 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-jn446\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.839571 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-jn446\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.841715 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/45f3dc48-d71b-47f8-8567-90df740103b5-client-ca\") pod \"route-controller-manager-6576b87f9c-nff6m\" (UID: \"45f3dc48-d71b-47f8-8567-90df740103b5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nff6m" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.841782 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-jn446\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.846657 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-jn446\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.846698 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/1f78456e-8b1a-472d-beb4-aabc4a911e01-machine-approver-tls\") pod \"machine-approver-56656f9798-znmpz\" (UID: \"1f78456e-8b1a-472d-beb4-aabc4a911e01\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-znmpz" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.847204 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/87db7503-6e3e-49b1-bc84-5230a444e8f6-proxy-tls\") pod \"machine-config-operator-74547568cd-nsqq5\" (UID: \"87db7503-6e3e-49b1-bc84-5230a444e8f6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nsqq5" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.847698 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6b48dffc-c51b-48e2-b22b-e809176c8830-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-852bd\" (UID: \"6b48dffc-c51b-48e2-b22b-e809176c8830\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-852bd" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.847973 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-wcth2"] Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.850283 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/d0367f91-0d22-4fa6-8cfe-0b549698b42b-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-m6kjl\" (UID: \"d0367f91-0d22-4fa6-8cfe-0b549698b42b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-m6kjl" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.853456 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-jn446\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.856604 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/b6b25f1e-33d4-481d-8c73-24b95c61364f-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-ch964\" (UID: \"b6b25f1e-33d4-481d-8c73-24b95c61364f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ch964" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.859890 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fv8gt\" (UniqueName: \"kubernetes.io/projected/1f78456e-8b1a-472d-beb4-aabc4a911e01-kube-api-access-fv8gt\") pod \"machine-approver-56656f9798-znmpz\" (UID: \"1f78456e-8b1a-472d-beb4-aabc4a911e01\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-znmpz" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.861000 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f6465618-cc08-4930-a290-0d1ffe135c98-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-wqnkb\" (UID: \"f6465618-cc08-4930-a290-0d1ffe135c98\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wqnkb" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.861710 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-jn446\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.863704 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxk8g\" (UniqueName: \"kubernetes.io/projected/c892e996-27b7-4b8d-bd92-7eac2df7702b-kube-api-access-vxk8g\") pod \"collect-profiles-29416665-lsfkp\" (UID: \"c892e996-27b7-4b8d-bd92-7eac2df7702b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416665-lsfkp" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.885488 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:44 crc kubenswrapper[4809]: E1206 05:53:44.885642 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:45.385608991 +0000 UTC m=+150.274591933 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.885755 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/c598f580-a438-4015-9a8a-f98b3b2b0b3d-csi-data-dir\") pod \"csi-hostpathplugin-p2lkt\" (UID: \"c598f580-a438-4015-9a8a-f98b3b2b0b3d\") " pod="hostpath-provisioner/csi-hostpathplugin-p2lkt" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.885782 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/c598f580-a438-4015-9a8a-f98b3b2b0b3d-registration-dir\") pod \"csi-hostpathplugin-p2lkt\" (UID: \"c598f580-a438-4015-9a8a-f98b3b2b0b3d\") " pod="hostpath-provisioner/csi-hostpathplugin-p2lkt" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.885801 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mm7c2\" (UniqueName: \"kubernetes.io/projected/e1380457-61e9-44d2-8324-eb1bcb638056-kube-api-access-mm7c2\") pod \"ingress-canary-czhdd\" (UID: \"e1380457-61e9-44d2-8324-eb1bcb638056\") " pod="openshift-ingress-canary/ingress-canary-czhdd" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.885829 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/c598f580-a438-4015-9a8a-f98b3b2b0b3d-plugins-dir\") pod \"csi-hostpathplugin-p2lkt\" (UID: \"c598f580-a438-4015-9a8a-f98b3b2b0b3d\") " pod="hostpath-provisioner/csi-hostpathplugin-p2lkt" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.885884 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpc9v\" (UniqueName: \"kubernetes.io/projected/c598f580-a438-4015-9a8a-f98b3b2b0b3d-kube-api-access-gpc9v\") pod \"csi-hostpathplugin-p2lkt\" (UID: \"c598f580-a438-4015-9a8a-f98b3b2b0b3d\") " pod="hostpath-provisioner/csi-hostpathplugin-p2lkt" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.885900 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e1380457-61e9-44d2-8324-eb1bcb638056-cert\") pod \"ingress-canary-czhdd\" (UID: \"e1380457-61e9-44d2-8324-eb1bcb638056\") " pod="openshift-ingress-canary/ingress-canary-czhdd" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.885914 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0a9a3552-d735-4692-ab80-2c1526680ab9-metrics-tls\") pod \"dns-default-s7px9\" (UID: \"0a9a3552-d735-4692-ab80-2c1526680ab9\") " pod="openshift-dns/dns-default-s7px9" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.885980 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6ft7\" (UniqueName: \"kubernetes.io/projected/0a9a3552-d735-4692-ab80-2c1526680ab9-kube-api-access-s6ft7\") pod \"dns-default-s7px9\" (UID: \"0a9a3552-d735-4692-ab80-2c1526680ab9\") " pod="openshift-dns/dns-default-s7px9" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.886021 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/c598f580-a438-4015-9a8a-f98b3b2b0b3d-mountpoint-dir\") pod \"csi-hostpathplugin-p2lkt\" (UID: \"c598f580-a438-4015-9a8a-f98b3b2b0b3d\") " pod="hostpath-provisioner/csi-hostpathplugin-p2lkt" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.886044 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0a9a3552-d735-4692-ab80-2c1526680ab9-config-volume\") pod \"dns-default-s7px9\" (UID: \"0a9a3552-d735-4692-ab80-2c1526680ab9\") " pod="openshift-dns/dns-default-s7px9" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.886060 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/c598f580-a438-4015-9a8a-f98b3b2b0b3d-socket-dir\") pod \"csi-hostpathplugin-p2lkt\" (UID: \"c598f580-a438-4015-9a8a-f98b3b2b0b3d\") " pod="hostpath-provisioner/csi-hostpathplugin-p2lkt" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.886546 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/c598f580-a438-4015-9a8a-f98b3b2b0b3d-mountpoint-dir\") pod \"csi-hostpathplugin-p2lkt\" (UID: \"c598f580-a438-4015-9a8a-f98b3b2b0b3d\") " pod="hostpath-provisioner/csi-hostpathplugin-p2lkt" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.887042 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/c598f580-a438-4015-9a8a-f98b3b2b0b3d-csi-data-dir\") pod \"csi-hostpathplugin-p2lkt\" (UID: \"c598f580-a438-4015-9a8a-f98b3b2b0b3d\") " pod="hostpath-provisioner/csi-hostpathplugin-p2lkt" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.887204 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/c598f580-a438-4015-9a8a-f98b3b2b0b3d-socket-dir\") pod \"csi-hostpathplugin-p2lkt\" (UID: \"c598f580-a438-4015-9a8a-f98b3b2b0b3d\") " pod="hostpath-provisioner/csi-hostpathplugin-p2lkt" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.887258 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/c598f580-a438-4015-9a8a-f98b3b2b0b3d-plugins-dir\") pod \"csi-hostpathplugin-p2lkt\" (UID: \"c598f580-a438-4015-9a8a-f98b3b2b0b3d\") " pod="hostpath-provisioner/csi-hostpathplugin-p2lkt" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.887362 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/c598f580-a438-4015-9a8a-f98b3b2b0b3d-registration-dir\") pod \"csi-hostpathplugin-p2lkt\" (UID: \"c598f580-a438-4015-9a8a-f98b3b2b0b3d\") " pod="hostpath-provisioner/csi-hostpathplugin-p2lkt" Dec 06 05:53:44 crc kubenswrapper[4809]: E1206 05:53:44.887664 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:45.387651436 +0000 UTC m=+150.276634378 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.887741 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0a9a3552-d735-4692-ab80-2c1526680ab9-config-volume\") pod \"dns-default-s7px9\" (UID: \"0a9a3552-d735-4692-ab80-2c1526680ab9\") " pod="openshift-dns/dns-default-s7px9" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.895924 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-8ssrl"] Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.886106 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.899358 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e1380457-61e9-44d2-8324-eb1bcb638056-cert\") pod \"ingress-canary-czhdd\" (UID: \"e1380457-61e9-44d2-8324-eb1bcb638056\") " pod="openshift-ingress-canary/ingress-canary-czhdd" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.900799 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f6465618-cc08-4930-a290-0d1ffe135c98-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-wqnkb\" (UID: \"f6465618-cc08-4930-a290-0d1ffe135c98\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wqnkb" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.904485 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0a9a3552-d735-4692-ab80-2c1526680ab9-metrics-tls\") pod \"dns-default-s7px9\" (UID: \"0a9a3552-d735-4692-ab80-2c1526680ab9\") " pod="openshift-dns/dns-default-s7px9" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.910990 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dd4pd\" (UniqueName: \"kubernetes.io/projected/09a7af3d-6bf2-4b7a-93d9-cecc73d993a1-kube-api-access-dd4pd\") pod \"apiserver-76f77b778f-tphf5\" (UID: \"09a7af3d-6bf2-4b7a-93d9-cecc73d993a1\") " pod="openshift-apiserver/apiserver-76f77b778f-tphf5" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.923863 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grzz4\" (UniqueName: \"kubernetes.io/projected/b6b25f1e-33d4-481d-8c73-24b95c61364f-kube-api-access-grzz4\") pod \"package-server-manager-789f6589d5-ch964\" (UID: \"b6b25f1e-33d4-481d-8c73-24b95c61364f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ch964" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.937792 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-hlvjb"] Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.950575 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wqnkb" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.952959 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4g75z\" (UniqueName: \"kubernetes.io/projected/87db7503-6e3e-49b1-bc84-5230a444e8f6-kube-api-access-4g75z\") pod \"machine-config-operator-74547568cd-nsqq5\" (UID: \"87db7503-6e3e-49b1-bc84-5230a444e8f6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nsqq5" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.972855 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ch964" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.978443 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nczzr\" (UniqueName: \"kubernetes.io/projected/0a9baebe-98e6-44be-ad16-882e47f09248-kube-api-access-nczzr\") pod \"oauth-openshift-558db77b4-jn446\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.985062 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8063bdc5-b0d2-4a8b-91d2-12c3e1e5f938-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7sxnv\" (UID: \"8063bdc5-b0d2-4a8b-91d2-12c3e1e5f938\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7sxnv" Dec 06 05:53:44 crc kubenswrapper[4809]: I1206 05:53:44.998703 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:44 crc kubenswrapper[4809]: E1206 05:53:44.999103 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:45.499086095 +0000 UTC m=+150.388069037 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.002198 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fhld\" (UniqueName: \"kubernetes.io/projected/45f3dc48-d71b-47f8-8567-90df740103b5-kube-api-access-7fhld\") pod \"route-controller-manager-6576b87f9c-nff6m\" (UID: \"45f3dc48-d71b-47f8-8567-90df740103b5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nff6m" Dec 06 05:53:45 crc kubenswrapper[4809]: W1206 05:53:45.002650 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd233521e_75db_4da3_9cd3_0480a99eba90.slice/crio-9503b0e93fab33ab3955b95693a7b3983cd6117016e7ab7c799013be89cd0e66 WatchSource:0}: Error finding container 9503b0e93fab33ab3955b95693a7b3983cd6117016e7ab7c799013be89cd0e66: Status 404 returned error can't find the container with id 9503b0e93fab33ab3955b95693a7b3983cd6117016e7ab7c799013be89cd0e66 Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.011428 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416665-lsfkp" Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.025323 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7d8w\" (UniqueName: \"kubernetes.io/projected/d0367f91-0d22-4fa6-8cfe-0b549698b42b-kube-api-access-l7d8w\") pod \"cluster-samples-operator-665b6dd947-m6kjl\" (UID: \"d0367f91-0d22-4fa6-8cfe-0b549698b42b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-m6kjl" Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.035328 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zj6j9"] Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.046404 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cztt\" (UniqueName: \"kubernetes.io/projected/14f07a0b-5e47-4636-9f17-cdf6f327b39d-kube-api-access-7cztt\") pod \"apiserver-7bbb656c7d-cd2tv\" (UID: \"14f07a0b-5e47-4636-9f17-cdf6f327b39d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cd2tv" Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.062779 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.073579 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-znmpz" Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.075396 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktzqv\" (UniqueName: \"kubernetes.io/projected/6b48dffc-c51b-48e2-b22b-e809176c8830-kube-api-access-ktzqv\") pod \"machine-api-operator-5694c8668f-852bd\" (UID: \"6b48dffc-c51b-48e2-b22b-e809176c8830\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-852bd" Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.081214 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nsqq5" Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.088014 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-m6kjl" Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.088897 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5q7r\" (UniqueName: \"kubernetes.io/projected/ee96792c-e168-4c58-a477-cd88c14da612-kube-api-access-c5q7r\") pod \"marketplace-operator-79b997595-slt9h\" (UID: \"ee96792c-e168-4c58-a477-cd88c14da612\") " pod="openshift-marketplace/marketplace-operator-79b997595-slt9h" Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.093012 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-852bd" Dec 06 05:53:45 crc kubenswrapper[4809]: W1206 05:53:45.094735 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd844d778_6b3d_400a_8f71_db1c2e643878.slice/crio-2e012115be293fc92c30b25aabca7a8c8ed6a9649ce4777791435717bf97d0f5 WatchSource:0}: Error finding container 2e012115be293fc92c30b25aabca7a8c8ed6a9649ce4777791435717bf97d0f5: Status 404 returned error can't find the container with id 2e012115be293fc92c30b25aabca7a8c8ed6a9649ce4777791435717bf97d0f5 Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.102469 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7sxnv" Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.102678 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:45 crc kubenswrapper[4809]: E1206 05:53:45.103118 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:45.603100601 +0000 UTC m=+150.492083543 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.114925 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-224mv"] Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.137976 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6ft7\" (UniqueName: \"kubernetes.io/projected/0a9a3552-d735-4692-ab80-2c1526680ab9-kube-api-access-s6ft7\") pod \"dns-default-s7px9\" (UID: \"0a9a3552-d735-4692-ab80-2c1526680ab9\") " pod="openshift-dns/dns-default-s7px9" Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.138014 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-927z2" event={"ID":"d7d02f2c-c052-4320-bac4-394fb40c57d2","Type":"ContainerStarted","Data":"7029482a25e7a08aa7def28578b456f88a0c19e61e39bd2a90371769ec75f0da"} Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.139398 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-4pwcw" event={"ID":"8e03364a-1840-4c91-8b2b-c7e2071fc73b","Type":"ContainerStarted","Data":"4007a4fcd8930f1f9d5f287878aac2fe01b09a892931d190ed6ac6d779049f92"} Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.139435 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-4pwcw" event={"ID":"8e03364a-1840-4c91-8b2b-c7e2071fc73b","Type":"ContainerStarted","Data":"16846d1b1f5b3efa865f1863bb744eaa55819731e81fe580e571a1e290114936"} Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.140851 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-s7q7v" event={"ID":"f4008f48-e17b-492f-8fa4-b09fb9d2ce70","Type":"ContainerStarted","Data":"545345580be0fae589076dca26ab6689a1d92c29693095bb2fd77ffec72ec008"} Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.141978 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pkq7h" event={"ID":"00bf9f05-7bc3-40e1-a2e9-1af1bf93f014","Type":"ContainerStarted","Data":"4e72c30f30b8927035f589fe945656c0694109575b3f22eb38fbe7f01e1c63d2"} Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.149317 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kjlr8" event={"ID":"d30c4960-b4da-41cf-9d1c-771e0bc7b262","Type":"ContainerStarted","Data":"6853141a272b1ed42b8e71b68ed45e59aca6f387b8dc5c94fa99b1eaaadfbecf"} Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.149368 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kjlr8" event={"ID":"d30c4960-b4da-41cf-9d1c-771e0bc7b262","Type":"ContainerStarted","Data":"2f45028eba7977ab9137069ce2c3992238a6bae630daef4e3fbe8e108a544dcc"} Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.153846 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-s7px9" Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.155526 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mm7c2\" (UniqueName: \"kubernetes.io/projected/e1380457-61e9-44d2-8324-eb1bcb638056-kube-api-access-mm7c2\") pod \"ingress-canary-czhdd\" (UID: \"e1380457-61e9-44d2-8324-eb1bcb638056\") " pod="openshift-ingress-canary/ingress-canary-czhdd" Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.184541 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r7684" event={"ID":"cf124a04-e72c-4c67-bd76-52648b4d20c9","Type":"ContainerStarted","Data":"bf324abf796155d1bf410fef1d9ac5cc9c842b88ae02b866470e32a53838a66c"} Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.184593 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r7684" event={"ID":"cf124a04-e72c-4c67-bd76-52648b4d20c9","Type":"ContainerStarted","Data":"cd2b6959582e85000cb9d7715d3869c8fa94499e78f418a632ebf21fd345f94f"} Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.186623 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpc9v\" (UniqueName: \"kubernetes.io/projected/c598f580-a438-4015-9a8a-f98b3b2b0b3d-kube-api-access-gpc9v\") pod \"csi-hostpathplugin-p2lkt\" (UID: \"c598f580-a438-4015-9a8a-f98b3b2b0b3d\") " pod="hostpath-provisioner/csi-hostpathplugin-p2lkt" Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.189633 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-qqv6x"] Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.195799 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"b37b9ef392ed1f99be43ecdc6f1fca76ea0c1cafd08779d07c52f61894573b1c"} Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.204828 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-tphf5" Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.205800 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.206909 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8ssrl" event={"ID":"d233521e-75db-4da3-9cd3-0480a99eba90","Type":"ContainerStarted","Data":"9503b0e93fab33ab3955b95693a7b3983cd6117016e7ab7c799013be89cd0e66"} Dec 06 05:53:45 crc kubenswrapper[4809]: E1206 05:53:45.207038 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:45.707007234 +0000 UTC m=+150.595990186 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.209582 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-7qpjs" event={"ID":"2a1ed741-489c-455f-a344-404bbab66c22","Type":"ContainerStarted","Data":"c9c9298faabafb75977a5c6feec58035eadc9d3b13b143a27316776caab0c1d2"} Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.212107 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zj6j9" event={"ID":"d844d778-6b3d-400a-8f71-db1c2e643878","Type":"ContainerStarted","Data":"2e012115be293fc92c30b25aabca7a8c8ed6a9649ce4777791435717bf97d0f5"} Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.224178 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-c8sn6" event={"ID":"26b87745-a15b-4f2d-b98e-b24afc5d512b","Type":"ContainerStarted","Data":"9c93dbc5a99fdf76953ee72f24dded9480b43ddc36b8396ccebb3c89224bee61"} Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.228307 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nff6m" Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.234120 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-ggvcl" event={"ID":"49fef68f-df3c-4bb7-80cf-f0f651797762","Type":"ContainerStarted","Data":"f8ea221c8a98918db5b341707f6a71678c5e760cbe478ec79c6b59ecba3680fb"} Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.244503 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cd2tv" Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.245120 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wcth2" event={"ID":"cdf2d3c6-659f-4dc5-a21c-d29cc86ef239","Type":"ContainerStarted","Data":"48ba3fdb8e59eeb73decb55f40bde294d48ddf42811b34e1c4becd1dce736522"} Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.247124 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-drj7v"] Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.274354 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-hlvjb" event={"ID":"f3a30985-f268-44de-93b9-bce0f9aebc6d","Type":"ContainerStarted","Data":"90e2d9431cc4728a7cf5496425c7b19cd713013475926e4234fed6966c0e0320"} Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.274474 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jjf9r"] Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.278497 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"fb948897024821de719e6b0490250172a102d977c058fc062a5ccfa5f32436f8"} Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.278812 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.280153 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9rgdg" event={"ID":"e2885220-fafe-48da-9ed7-98942886ab1f","Type":"ContainerStarted","Data":"ffd046ea24bb4cdbf0c07cadd285d9bf4d8f134eecc36eef3552ee8e078d4353"} Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.280182 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9rgdg" event={"ID":"e2885220-fafe-48da-9ed7-98942886ab1f","Type":"ContainerStarted","Data":"26e8dc55f563cb88ca71eeb0cffb270945a93248d0da5997f3c6e2162488cc7f"} Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.280945 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-n6z57" event={"ID":"e5356f1d-5afd-47bd-89af-72abfa8f901f","Type":"ContainerStarted","Data":"e801928518299496c9341ab8c60a6af2f1a5846e1120bc50e3ba320518bbb8ac"} Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.286353 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sdg4b" event={"ID":"1718efb3-c63c-4aac-bc59-2cc85a59ba61","Type":"ContainerStarted","Data":"b71317de92f13286a4078a28bb7bdb688b12da8214a75ac5c7d4bd5a483a2383"} Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.300698 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-slt9h" Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.307149 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:45 crc kubenswrapper[4809]: E1206 05:53:45.308864 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:45.808826659 +0000 UTC m=+150.697809591 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:45 crc kubenswrapper[4809]: W1206 05:53:45.404207 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod63b0cea8_1716_4c17_9f87_f10e776cf47f.slice/crio-80f5882f31d086a5f8a9d793224753dbb70b0a58226d38730c463369a9ec2918 WatchSource:0}: Error finding container 80f5882f31d086a5f8a9d793224753dbb70b0a58226d38730c463369a9ec2918: Status 404 returned error can't find the container with id 80f5882f31d086a5f8a9d793224753dbb70b0a58226d38730c463369a9ec2918 Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.423559 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:45 crc kubenswrapper[4809]: E1206 05:53:45.425466 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:45.925431789 +0000 UTC m=+150.814414741 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.436405 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.435208 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-p2lkt" Dec 06 05:53:45 crc kubenswrapper[4809]: E1206 05:53:45.438789 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:45.938772025 +0000 UTC m=+150.827754967 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.446485 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-czhdd" Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.475216 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dpbfw"] Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.538316 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:45 crc kubenswrapper[4809]: E1206 05:53:45.538736 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:46.038716629 +0000 UTC m=+150.927699581 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:45 crc kubenswrapper[4809]: W1206 05:53:45.633003 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podde0bb137_3c1d_4632_a7ef_6922f027a3c1.slice/crio-1b04c5f14d80538c24f37de25eaa4ae37b8ffc832379ccfd5acfe3cc00ba56f5 WatchSource:0}: Error finding container 1b04c5f14d80538c24f37de25eaa4ae37b8ffc832379ccfd5acfe3cc00ba56f5: Status 404 returned error can't find the container with id 1b04c5f14d80538c24f37de25eaa4ae37b8ffc832379ccfd5acfe3cc00ba56f5 Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.640323 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:45 crc kubenswrapper[4809]: E1206 05:53:45.641286 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:46.141254784 +0000 UTC m=+151.030237726 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.683162 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gt9kb"] Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.716604 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wqnkb"] Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.743323 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:45 crc kubenswrapper[4809]: E1206 05:53:45.743687 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:46.243672317 +0000 UTC m=+151.132655259 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.777300 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-mxv9r"] Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.838873 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kr48t"] Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.845416 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:45 crc kubenswrapper[4809]: E1206 05:53:45.845798 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:46.34577553 +0000 UTC m=+151.234758472 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:45 crc kubenswrapper[4809]: W1206 05:53:45.934299 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcea05156_8322_404a_8b33_867aeaaa905e.slice/crio-14b5c5cc53e56e3c38878ebfad454a441cf8389d6ef1355d4f2b75a3d7337801 WatchSource:0}: Error finding container 14b5c5cc53e56e3c38878ebfad454a441cf8389d6ef1355d4f2b75a3d7337801: Status 404 returned error can't find the container with id 14b5c5cc53e56e3c38878ebfad454a441cf8389d6ef1355d4f2b75a3d7337801 Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.946138 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-852bd"] Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.949011 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:45 crc kubenswrapper[4809]: E1206 05:53:45.949313 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:46.449279261 +0000 UTC m=+151.338262213 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:45 crc kubenswrapper[4809]: I1206 05:53:45.949543 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:45 crc kubenswrapper[4809]: E1206 05:53:45.950179 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:46.450167296 +0000 UTC m=+151.339150238 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:46 crc kubenswrapper[4809]: I1206 05:53:46.018477 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r7684" podStartSLOduration=131.018457904 podStartE2EDuration="2m11.018457904s" podCreationTimestamp="2025-12-06 05:51:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:53:45.979640782 +0000 UTC m=+150.868623724" watchObservedRunningTime="2025-12-06 05:53:46.018457904 +0000 UTC m=+150.907440846" Dec 06 05:53:46 crc kubenswrapper[4809]: I1206 05:53:46.044844 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416665-lsfkp"] Dec 06 05:53:46 crc kubenswrapper[4809]: I1206 05:53:46.051498 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:46 crc kubenswrapper[4809]: E1206 05:53:46.052154 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:46.552134296 +0000 UTC m=+151.441117248 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:46 crc kubenswrapper[4809]: I1206 05:53:46.152746 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:46 crc kubenswrapper[4809]: E1206 05:53:46.153041 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:46.653028846 +0000 UTC m=+151.542011788 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:46 crc kubenswrapper[4809]: I1206 05:53:46.254829 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:46 crc kubenswrapper[4809]: E1206 05:53:46.254989 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:46.754959625 +0000 UTC m=+151.643942577 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:46 crc kubenswrapper[4809]: I1206 05:53:46.261879 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:46 crc kubenswrapper[4809]: E1206 05:53:46.262699 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:46.762682106 +0000 UTC m=+151.651665048 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:46 crc kubenswrapper[4809]: I1206 05:53:46.322507 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ch964"] Dec 06 05:53:46 crc kubenswrapper[4809]: I1206 05:53:46.337074 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416665-lsfkp" event={"ID":"c892e996-27b7-4b8d-bd92-7eac2df7702b","Type":"ContainerStarted","Data":"b73033163a7fa503fbc6be3c2b6de4efe81ecc843d0c92400f5406b9a9e7e344"} Dec 06 05:53:46 crc kubenswrapper[4809]: I1206 05:53:46.339027 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7sxnv"] Dec 06 05:53:46 crc kubenswrapper[4809]: I1206 05:53:46.340412 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gt9kb" event={"ID":"c8d79e6f-40b0-4ecc-8501-454ba5cfe691","Type":"ContainerStarted","Data":"b6d876f8ea60d6107f003b486b72891466f1de0ba14ff3c107cf422e83464c3d"} Dec 06 05:53:46 crc kubenswrapper[4809]: I1206 05:53:46.343193 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-mxv9r" event={"ID":"cea05156-8322-404a-8b33-867aeaaa905e","Type":"ContainerStarted","Data":"14b5c5cc53e56e3c38878ebfad454a441cf8389d6ef1355d4f2b75a3d7337801"} Dec 06 05:53:46 crc kubenswrapper[4809]: I1206 05:53:46.344050 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-znmpz" event={"ID":"1f78456e-8b1a-472d-beb4-aabc4a911e01","Type":"ContainerStarted","Data":"79f2aef217c088df649cb0f24f424a1f7bb584dbb5beb430eeaa30d4a1b12aec"} Dec 06 05:53:46 crc kubenswrapper[4809]: I1206 05:53:46.345148 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dpbfw" event={"ID":"de0bb137-3c1d-4632-a7ef-6922f027a3c1","Type":"ContainerStarted","Data":"1b04c5f14d80538c24f37de25eaa4ae37b8ffc832379ccfd5acfe3cc00ba56f5"} Dec 06 05:53:46 crc kubenswrapper[4809]: I1206 05:53:46.346557 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-ggvcl" event={"ID":"49fef68f-df3c-4bb7-80cf-f0f651797762","Type":"ContainerStarted","Data":"b4f1ca19e4f9be63b5be9830ba9302245bb78083537d87f172a89ffc807173d9"} Dec 06 05:53:46 crc kubenswrapper[4809]: I1206 05:53:46.348656 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-852bd" event={"ID":"6b48dffc-c51b-48e2-b22b-e809176c8830","Type":"ContainerStarted","Data":"ecd6eaa6a4bb8c775684319e091980d86470ec7434d3f82040d90a300c31b8a3"} Dec 06 05:53:46 crc kubenswrapper[4809]: I1206 05:53:46.362853 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:46 crc kubenswrapper[4809]: E1206 05:53:46.362996 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:46.86297255 +0000 UTC m=+151.751955492 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:46 crc kubenswrapper[4809]: I1206 05:53:46.363219 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kr48t" event={"ID":"d0b4b35c-3cb0-405d-941b-1b3946b228a0","Type":"ContainerStarted","Data":"44c350a907285742e9610c5074a1a0e5e9c2576127ae3a1f44c56a5352090d24"} Dec 06 05:53:46 crc kubenswrapper[4809]: I1206 05:53:46.363388 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:46 crc kubenswrapper[4809]: E1206 05:53:46.363791 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:46.863780191 +0000 UTC m=+151.752763133 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:46 crc kubenswrapper[4809]: I1206 05:53:46.381225 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qqv6x" event={"ID":"8a225b5c-4ad5-4f08-a0cb-ae38a8fa83d4","Type":"ContainerStarted","Data":"ac1bdaf96f3fd32fc66766d555d8649f20309b9ec1e39a80a832bdbf6b387db3"} Dec 06 05:53:46 crc kubenswrapper[4809]: I1206 05:53:46.396254 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-c8sn6" event={"ID":"26b87745-a15b-4f2d-b98e-b24afc5d512b","Type":"ContainerStarted","Data":"367819a8620909c508bfb5d62e88221a9ef09151d681581a06569726b57eb244"} Dec 06 05:53:46 crc kubenswrapper[4809]: I1206 05:53:46.410652 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-g7pfc" event={"ID":"b119ee35-1a33-4a34-9fca-62ad465a20f7","Type":"ContainerStarted","Data":"24dc91625ed062aa5d82a79a88d7f2aafb3581af6523f9bbdd6d348290f1fada"} Dec 06 05:53:46 crc kubenswrapper[4809]: I1206 05:53:46.438901 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jjf9r" event={"ID":"e5e4e451-e3ee-434e-8e24-0c795925a48e","Type":"ContainerStarted","Data":"3756cadd9aca53f51dd8faf914898d74ba820f8c9b497675a1b02ac985700ade"} Dec 06 05:53:46 crc kubenswrapper[4809]: I1206 05:53:46.447469 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-224mv" event={"ID":"8e290142-b2fc-492e-8763-ff41224579fc","Type":"ContainerStarted","Data":"0697e70a8c439d433dea01d4281772c49dd248da8b6ff50026cbf75cfaaa313c"} Dec 06 05:53:46 crc kubenswrapper[4809]: I1206 05:53:46.448745 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-drj7v" event={"ID":"63b0cea8-1716-4c17-9f87-f10e776cf47f","Type":"ContainerStarted","Data":"80f5882f31d086a5f8a9d793224753dbb70b0a58226d38730c463369a9ec2918"} Dec 06 05:53:46 crc kubenswrapper[4809]: I1206 05:53:46.453074 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wqnkb" event={"ID":"f6465618-cc08-4930-a290-0d1ffe135c98","Type":"ContainerStarted","Data":"f124ff8d54cfc045ab52832d085ec362e7b7f97be5d8e9d5c89cc785f5497233"} Dec 06 05:53:46 crc kubenswrapper[4809]: I1206 05:53:46.464807 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:46 crc kubenswrapper[4809]: E1206 05:53:46.465061 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:46.965044583 +0000 UTC m=+151.854027515 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:46 crc kubenswrapper[4809]: I1206 05:53:46.465164 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:46 crc kubenswrapper[4809]: E1206 05:53:46.465439 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:46.965431503 +0000 UTC m=+151.854414445 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:46 crc kubenswrapper[4809]: I1206 05:53:46.504245 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jn446"] Dec 06 05:53:46 crc kubenswrapper[4809]: W1206 05:53:46.561326 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0a9baebe_98e6_44be_ad16_882e47f09248.slice/crio-dad7c50235935b49033383b51383d232bb7c41a9c13c3c777faba482031af985 WatchSource:0}: Error finding container dad7c50235935b49033383b51383d232bb7c41a9c13c3c777faba482031af985: Status 404 returned error can't find the container with id dad7c50235935b49033383b51383d232bb7c41a9c13c3c777faba482031af985 Dec 06 05:53:46 crc kubenswrapper[4809]: I1206 05:53:46.566097 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:46 crc kubenswrapper[4809]: E1206 05:53:46.566350 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:47.066310343 +0000 UTC m=+151.955293295 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:46 crc kubenswrapper[4809]: I1206 05:53:46.566525 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:46 crc kubenswrapper[4809]: E1206 05:53:46.569184 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:47.069174811 +0000 UTC m=+151.958157753 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:46 crc kubenswrapper[4809]: I1206 05:53:46.577559 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-p2lkt"] Dec 06 05:53:46 crc kubenswrapper[4809]: I1206 05:53:46.668672 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-4pwcw" podStartSLOduration=132.668649263 podStartE2EDuration="2m12.668649263s" podCreationTimestamp="2025-12-06 05:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:53:46.664493189 +0000 UTC m=+151.553476141" watchObservedRunningTime="2025-12-06 05:53:46.668649263 +0000 UTC m=+151.557632195" Dec 06 05:53:46 crc kubenswrapper[4809]: I1206 05:53:46.671669 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:46 crc kubenswrapper[4809]: E1206 05:53:46.672081 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:47.172062037 +0000 UTC m=+152.061044979 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:46 crc kubenswrapper[4809]: I1206 05:53:46.712538 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-slt9h"] Dec 06 05:53:46 crc kubenswrapper[4809]: I1206 05:53:46.728369 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-nsqq5"] Dec 06 05:53:46 crc kubenswrapper[4809]: I1206 05:53:46.747382 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-m6kjl"] Dec 06 05:53:46 crc kubenswrapper[4809]: W1206 05:53:46.758407 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod87db7503_6e3e_49b1_bc84_5230a444e8f6.slice/crio-69b593749277a06f3ec17d32824518987a79f72eeced67cf512254304ec00d10 WatchSource:0}: Error finding container 69b593749277a06f3ec17d32824518987a79f72eeced67cf512254304ec00d10: Status 404 returned error can't find the container with id 69b593749277a06f3ec17d32824518987a79f72eeced67cf512254304ec00d10 Dec 06 05:53:46 crc kubenswrapper[4809]: I1206 05:53:46.773605 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:46 crc kubenswrapper[4809]: E1206 05:53:46.774134 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:47.274117409 +0000 UTC m=+152.163100351 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:46 crc kubenswrapper[4809]: I1206 05:53:46.813637 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-s7px9"] Dec 06 05:53:46 crc kubenswrapper[4809]: I1206 05:53:46.816527 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-nff6m"] Dec 06 05:53:46 crc kubenswrapper[4809]: I1206 05:53:46.818485 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-czhdd"] Dec 06 05:53:46 crc kubenswrapper[4809]: I1206 05:53:46.823682 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9rgdg" podStartSLOduration=131.823643323 podStartE2EDuration="2m11.823643323s" podCreationTimestamp="2025-12-06 05:51:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:53:46.814226296 +0000 UTC m=+151.703209238" watchObservedRunningTime="2025-12-06 05:53:46.823643323 +0000 UTC m=+151.712626265" Dec 06 05:53:46 crc kubenswrapper[4809]: I1206 05:53:46.880951 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:46 crc kubenswrapper[4809]: E1206 05:53:46.881530 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:47.381515397 +0000 UTC m=+152.270498339 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:46 crc kubenswrapper[4809]: I1206 05:53:46.894507 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-c8sn6" podStartSLOduration=131.894489842 podStartE2EDuration="2m11.894489842s" podCreationTimestamp="2025-12-06 05:51:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:53:46.87687085 +0000 UTC m=+151.765853792" watchObservedRunningTime="2025-12-06 05:53:46.894489842 +0000 UTC m=+151.783472784" Dec 06 05:53:46 crc kubenswrapper[4809]: I1206 05:53:46.894760 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-tphf5"] Dec 06 05:53:46 crc kubenswrapper[4809]: I1206 05:53:46.911551 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-cd2tv"] Dec 06 05:53:46 crc kubenswrapper[4809]: I1206 05:53:46.985604 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:46 crc kubenswrapper[4809]: E1206 05:53:46.986038 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:47.486023336 +0000 UTC m=+152.375006278 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:47 crc kubenswrapper[4809]: I1206 05:53:47.086573 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:47 crc kubenswrapper[4809]: E1206 05:53:47.086745 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:47.586720631 +0000 UTC m=+152.475703573 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:47 crc kubenswrapper[4809]: I1206 05:53:47.087060 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:47 crc kubenswrapper[4809]: E1206 05:53:47.087430 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:47.58742171 +0000 UTC m=+152.476404652 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:47 crc kubenswrapper[4809]: I1206 05:53:47.188442 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:47 crc kubenswrapper[4809]: E1206 05:53:47.188820 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:47.688801274 +0000 UTC m=+152.577784216 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:47 crc kubenswrapper[4809]: I1206 05:53:47.290300 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:47 crc kubenswrapper[4809]: E1206 05:53:47.290980 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:47.790956929 +0000 UTC m=+152.679939871 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:47 crc kubenswrapper[4809]: I1206 05:53:47.391763 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:47 crc kubenswrapper[4809]: E1206 05:53:47.392255 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:47.89223555 +0000 UTC m=+152.781218492 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:47 crc kubenswrapper[4809]: I1206 05:53:47.458512 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-jn446" event={"ID":"0a9baebe-98e6-44be-ad16-882e47f09248","Type":"ContainerStarted","Data":"dad7c50235935b49033383b51383d232bb7c41a9c13c3c777faba482031af985"} Dec 06 05:53:47 crc kubenswrapper[4809]: I1206 05:53:47.462278 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-7qpjs" event={"ID":"2a1ed741-489c-455f-a344-404bbab66c22","Type":"ContainerStarted","Data":"d6349ee6d3b2d24851d9fb4c381c6461d8aade431f1bc4d59880c6989f34e131"} Dec 06 05:53:47 crc kubenswrapper[4809]: I1206 05:53:47.463235 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nff6m" event={"ID":"45f3dc48-d71b-47f8-8567-90df740103b5","Type":"ContainerStarted","Data":"9b88cf79838ed43cf246dba0624b73c0b37eea8cf22f6b2c637b1214303990d4"} Dec 06 05:53:47 crc kubenswrapper[4809]: I1206 05:53:47.464676 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cd2tv" event={"ID":"14f07a0b-5e47-4636-9f17-cdf6f327b39d","Type":"ContainerStarted","Data":"f4bb92bfdee7b229669173974967a01924943c6d56c1df860d9e4e090b07cccf"} Dec 06 05:53:47 crc kubenswrapper[4809]: I1206 05:53:47.465464 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ch964" event={"ID":"b6b25f1e-33d4-481d-8c73-24b95c61364f","Type":"ContainerStarted","Data":"e5e01f182cbd7ecb003411cfdd6100340d7c2e675f39b4d1cc97fe5525526791"} Dec 06 05:53:47 crc kubenswrapper[4809]: I1206 05:53:47.466178 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7sxnv" event={"ID":"8063bdc5-b0d2-4a8b-91d2-12c3e1e5f938","Type":"ContainerStarted","Data":"6f3b62750216e446a9d12a82c1ddb90885f7794797453499ca5dd82fb1314c38"} Dec 06 05:53:47 crc kubenswrapper[4809]: I1206 05:53:47.470529 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-slt9h" event={"ID":"ee96792c-e168-4c58-a477-cd88c14da612","Type":"ContainerStarted","Data":"65d6d119611bca3ae30cc5f8c6cd1480d76ea44c6ad1509ceea897b0e9bbd8b0"} Dec 06 05:53:47 crc kubenswrapper[4809]: I1206 05:53:47.471492 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-s7px9" event={"ID":"0a9a3552-d735-4692-ab80-2c1526680ab9","Type":"ContainerStarted","Data":"0c868a5e3ef92d6ad5911aa07f2cb8715a1ee443fad859b86903cee6c986f65f"} Dec 06 05:53:47 crc kubenswrapper[4809]: I1206 05:53:47.472238 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-czhdd" event={"ID":"e1380457-61e9-44d2-8324-eb1bcb638056","Type":"ContainerStarted","Data":"9e775b94c954bb7b93f2badfbcf263b5d663d3d36b189d90e45171f84eb462d6"} Dec 06 05:53:47 crc kubenswrapper[4809]: I1206 05:53:47.473338 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-tphf5" event={"ID":"09a7af3d-6bf2-4b7a-93d9-cecc73d993a1","Type":"ContainerStarted","Data":"a518b88bbb1cb1879213c6cd1586539a3a2e01709b6081e518b1f5652cf438fb"} Dec 06 05:53:47 crc kubenswrapper[4809]: I1206 05:53:47.474121 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-p2lkt" event={"ID":"c598f580-a438-4015-9a8a-f98b3b2b0b3d","Type":"ContainerStarted","Data":"149b1b908cf8c90d3d7924e5015e9b89526d8c8eeea23ed4575b076f46e51dd7"} Dec 06 05:53:47 crc kubenswrapper[4809]: I1206 05:53:47.475596 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nsqq5" event={"ID":"87db7503-6e3e-49b1-bc84-5230a444e8f6","Type":"ContainerStarted","Data":"69b593749277a06f3ec17d32824518987a79f72eeced67cf512254304ec00d10"} Dec 06 05:53:47 crc kubenswrapper[4809]: I1206 05:53:47.493758 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:47 crc kubenswrapper[4809]: E1206 05:53:47.494304 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:47.994280002 +0000 UTC m=+152.883262994 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:47 crc kubenswrapper[4809]: I1206 05:53:47.595557 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:47 crc kubenswrapper[4809]: E1206 05:53:47.596217 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:48.09619356 +0000 UTC m=+152.985176492 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:47 crc kubenswrapper[4809]: I1206 05:53:47.697033 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:47 crc kubenswrapper[4809]: E1206 05:53:47.697544 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:48.197518252 +0000 UTC m=+153.086501224 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:47 crc kubenswrapper[4809]: I1206 05:53:47.798129 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:47 crc kubenswrapper[4809]: E1206 05:53:47.798481 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:48.298464194 +0000 UTC m=+153.187447136 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:47 crc kubenswrapper[4809]: I1206 05:53:47.899709 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:47 crc kubenswrapper[4809]: E1206 05:53:47.900079 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:48.400063804 +0000 UTC m=+153.289046756 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:48 crc kubenswrapper[4809]: I1206 05:53:48.000627 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:48 crc kubenswrapper[4809]: E1206 05:53:48.000854 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:48.5008257 +0000 UTC m=+153.389808682 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:48 crc kubenswrapper[4809]: I1206 05:53:48.001039 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:48 crc kubenswrapper[4809]: E1206 05:53:48.001421 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:48.501405776 +0000 UTC m=+153.390388718 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:48 crc kubenswrapper[4809]: I1206 05:53:48.102345 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:48 crc kubenswrapper[4809]: E1206 05:53:48.102464 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:48.60244494 +0000 UTC m=+153.491427892 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:48 crc kubenswrapper[4809]: E1206 05:53:48.103164 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:48.60315356 +0000 UTC m=+153.492136512 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:48 crc kubenswrapper[4809]: I1206 05:53:48.102817 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:48 crc kubenswrapper[4809]: I1206 05:53:48.204228 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:48 crc kubenswrapper[4809]: E1206 05:53:48.204409 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:48.704379989 +0000 UTC m=+153.593362941 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:48 crc kubenswrapper[4809]: I1206 05:53:48.204558 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:48 crc kubenswrapper[4809]: E1206 05:53:48.204870 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:48.704857882 +0000 UTC m=+153.593840824 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:48 crc kubenswrapper[4809]: I1206 05:53:48.305919 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:48 crc kubenswrapper[4809]: E1206 05:53:48.306072 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:48.806043981 +0000 UTC m=+153.695026923 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:48 crc kubenswrapper[4809]: I1206 05:53:48.306218 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:48 crc kubenswrapper[4809]: E1206 05:53:48.306763 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:48.80675424 +0000 UTC m=+153.695737172 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:48 crc kubenswrapper[4809]: I1206 05:53:48.407373 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:48 crc kubenswrapper[4809]: E1206 05:53:48.407570 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:48.907542537 +0000 UTC m=+153.796525479 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:48 crc kubenswrapper[4809]: I1206 05:53:48.407644 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:48 crc kubenswrapper[4809]: E1206 05:53:48.407986 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:48.907977699 +0000 UTC m=+153.796960641 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:48 crc kubenswrapper[4809]: I1206 05:53:48.481033 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sdg4b" event={"ID":"1718efb3-c63c-4aac-bc59-2cc85a59ba61","Type":"ContainerStarted","Data":"ab27d680fae997ad7e54926d783adc2299f4d86e35205cd51e29c3e37031977a"} Dec 06 05:53:48 crc kubenswrapper[4809]: I1206 05:53:48.482680 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pkq7h" event={"ID":"00bf9f05-7bc3-40e1-a2e9-1af1bf93f014","Type":"ContainerStarted","Data":"acb169d8cd10a0487d4f986882748c2249d3ce63f3ea0d6206b33ef6effae202"} Dec 06 05:53:48 crc kubenswrapper[4809]: I1206 05:53:48.508596 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:48 crc kubenswrapper[4809]: E1206 05:53:48.508762 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:49.008738486 +0000 UTC m=+153.897721428 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:48 crc kubenswrapper[4809]: I1206 05:53:48.508854 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:48 crc kubenswrapper[4809]: E1206 05:53:48.509179 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:49.009171598 +0000 UTC m=+153.898154540 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:48 crc kubenswrapper[4809]: I1206 05:53:48.609885 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:48 crc kubenswrapper[4809]: E1206 05:53:48.610069 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:49.110040618 +0000 UTC m=+153.999023570 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:48 crc kubenswrapper[4809]: I1206 05:53:48.610235 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:48 crc kubenswrapper[4809]: E1206 05:53:48.610530 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:49.110519521 +0000 UTC m=+153.999502533 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:48 crc kubenswrapper[4809]: I1206 05:53:48.711869 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:48 crc kubenswrapper[4809]: E1206 05:53:48.712018 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:49.211992367 +0000 UTC m=+154.100975309 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:48 crc kubenswrapper[4809]: I1206 05:53:48.712197 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:48 crc kubenswrapper[4809]: E1206 05:53:48.712462 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:49.21245257 +0000 UTC m=+154.101435512 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:48 crc kubenswrapper[4809]: I1206 05:53:48.813514 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:48 crc kubenswrapper[4809]: E1206 05:53:48.813620 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:49.313603797 +0000 UTC m=+154.202586739 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:48 crc kubenswrapper[4809]: I1206 05:53:48.813888 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:48 crc kubenswrapper[4809]: E1206 05:53:48.814164 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:49.314156423 +0000 UTC m=+154.203139365 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:48 crc kubenswrapper[4809]: I1206 05:53:48.914519 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:48 crc kubenswrapper[4809]: E1206 05:53:48.914702 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:49.414677032 +0000 UTC m=+154.303659974 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:48 crc kubenswrapper[4809]: I1206 05:53:48.914869 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:48 crc kubenswrapper[4809]: E1206 05:53:48.915225 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:49.415217127 +0000 UTC m=+154.304200069 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:49 crc kubenswrapper[4809]: I1206 05:53:49.016374 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:49 crc kubenswrapper[4809]: E1206 05:53:49.016606 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:49.51655966 +0000 UTC m=+154.405542602 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:49 crc kubenswrapper[4809]: I1206 05:53:49.016678 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:49 crc kubenswrapper[4809]: E1206 05:53:49.016991 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:49.516979551 +0000 UTC m=+154.405962483 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:49 crc kubenswrapper[4809]: I1206 05:53:49.117883 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:49 crc kubenswrapper[4809]: E1206 05:53:49.118161 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:49.618132719 +0000 UTC m=+154.507115671 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:49 crc kubenswrapper[4809]: I1206 05:53:49.219149 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:49 crc kubenswrapper[4809]: E1206 05:53:49.219619 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:49.719602215 +0000 UTC m=+154.608585157 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:49 crc kubenswrapper[4809]: I1206 05:53:49.320222 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:49 crc kubenswrapper[4809]: E1206 05:53:49.320392 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:49.820359061 +0000 UTC m=+154.709342023 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:49 crc kubenswrapper[4809]: I1206 05:53:49.320718 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:49 crc kubenswrapper[4809]: E1206 05:53:49.321106 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:49.821093272 +0000 UTC m=+154.710076214 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:49 crc kubenswrapper[4809]: I1206 05:53:49.421778 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:49 crc kubenswrapper[4809]: E1206 05:53:49.421988 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:49.921964432 +0000 UTC m=+154.810947374 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:49 crc kubenswrapper[4809]: I1206 05:53:49.422077 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:49 crc kubenswrapper[4809]: E1206 05:53:49.422412 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:49.922404673 +0000 UTC m=+154.811387615 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:49 crc kubenswrapper[4809]: I1206 05:53:49.489508 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-927z2" event={"ID":"d7d02f2c-c052-4320-bac4-394fb40c57d2","Type":"ContainerStarted","Data":"43e57375b0cfd2f17135fea4927d9a00ceed269e050bb08f3c8bb2f756165d9c"} Dec 06 05:53:49 crc kubenswrapper[4809]: I1206 05:53:49.490556 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8ssrl" event={"ID":"d233521e-75db-4da3-9cd3-0480a99eba90","Type":"ContainerStarted","Data":"9a8278124c0f3f8cfc0a096e5a34525d932811a85623dfb1fbb584ded57553c0"} Dec 06 05:53:49 crc kubenswrapper[4809]: I1206 05:53:49.492444 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-n6z57" event={"ID":"e5356f1d-5afd-47bd-89af-72abfa8f901f","Type":"ContainerStarted","Data":"e061b12da080bb0dc2c8e03d029012d9f630feb21d1998eb7a5ed7f369824e29"} Dec 06 05:53:49 crc kubenswrapper[4809]: I1206 05:53:49.494075 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-hlvjb" event={"ID":"f3a30985-f268-44de-93b9-bce0f9aebc6d","Type":"ContainerStarted","Data":"14009f9cf4e890d82d10c1569da8ae672118c24e2c0f3c5f4fb375548f3e2114"} Dec 06 05:53:49 crc kubenswrapper[4809]: I1206 05:53:49.495487 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-g7pfc" event={"ID":"b119ee35-1a33-4a34-9fca-62ad465a20f7","Type":"ContainerStarted","Data":"9f8fd0564bcfdc7f026a97bd8a793cf3cd24d5146ee03f7574f88212fffc8fdd"} Dec 06 05:53:49 crc kubenswrapper[4809]: I1206 05:53:49.495791 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-ggvcl" Dec 06 05:53:49 crc kubenswrapper[4809]: I1206 05:53:49.508427 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-ggvcl" Dec 06 05:53:49 crc kubenswrapper[4809]: I1206 05:53:49.512645 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-ggvcl" podStartSLOduration=135.512628021 podStartE2EDuration="2m15.512628021s" podCreationTimestamp="2025-12-06 05:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:53:49.510527714 +0000 UTC m=+154.399510676" watchObservedRunningTime="2025-12-06 05:53:49.512628021 +0000 UTC m=+154.401610953" Dec 06 05:53:49 crc kubenswrapper[4809]: I1206 05:53:49.523564 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:49 crc kubenswrapper[4809]: E1206 05:53:49.523728 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:50.023692424 +0000 UTC m=+154.912675366 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:49 crc kubenswrapper[4809]: I1206 05:53:49.523903 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:49 crc kubenswrapper[4809]: E1206 05:53:49.524294 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:50.024283631 +0000 UTC m=+154.913266643 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:49 crc kubenswrapper[4809]: I1206 05:53:49.626591 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:49 crc kubenswrapper[4809]: E1206 05:53:49.627620 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:50.127605008 +0000 UTC m=+155.016587950 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:49 crc kubenswrapper[4809]: I1206 05:53:49.728097 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:49 crc kubenswrapper[4809]: E1206 05:53:49.728666 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:50.228653622 +0000 UTC m=+155.117636564 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:49 crc kubenswrapper[4809]: I1206 05:53:49.794995 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 06 05:53:49 crc kubenswrapper[4809]: I1206 05:53:49.795713 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 05:53:49 crc kubenswrapper[4809]: I1206 05:53:49.807837 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 06 05:53:49 crc kubenswrapper[4809]: I1206 05:53:49.807900 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 06 05:53:49 crc kubenswrapper[4809]: I1206 05:53:49.817693 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 06 05:53:49 crc kubenswrapper[4809]: I1206 05:53:49.829619 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:49 crc kubenswrapper[4809]: I1206 05:53:49.829895 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bebb2fb5-644c-49bf-9899-3bc6116cb90e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"bebb2fb5-644c-49bf-9899-3bc6116cb90e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 05:53:49 crc kubenswrapper[4809]: I1206 05:53:49.829986 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bebb2fb5-644c-49bf-9899-3bc6116cb90e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"bebb2fb5-644c-49bf-9899-3bc6116cb90e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 05:53:49 crc kubenswrapper[4809]: E1206 05:53:49.830087 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:50.330070626 +0000 UTC m=+155.219053568 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:49 crc kubenswrapper[4809]: I1206 05:53:49.931043 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bebb2fb5-644c-49bf-9899-3bc6116cb90e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"bebb2fb5-644c-49bf-9899-3bc6116cb90e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 05:53:49 crc kubenswrapper[4809]: I1206 05:53:49.931086 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:49 crc kubenswrapper[4809]: I1206 05:53:49.931134 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bebb2fb5-644c-49bf-9899-3bc6116cb90e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"bebb2fb5-644c-49bf-9899-3bc6116cb90e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 05:53:49 crc kubenswrapper[4809]: I1206 05:53:49.931191 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bebb2fb5-644c-49bf-9899-3bc6116cb90e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"bebb2fb5-644c-49bf-9899-3bc6116cb90e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 05:53:49 crc kubenswrapper[4809]: E1206 05:53:49.931515 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:50.431498652 +0000 UTC m=+155.320481594 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:49 crc kubenswrapper[4809]: I1206 05:53:49.951742 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bebb2fb5-644c-49bf-9899-3bc6116cb90e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"bebb2fb5-644c-49bf-9899-3bc6116cb90e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 05:53:50 crc kubenswrapper[4809]: I1206 05:53:50.032911 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:50 crc kubenswrapper[4809]: E1206 05:53:50.033458 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:50.5334429 +0000 UTC m=+155.422425842 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:50 crc kubenswrapper[4809]: I1206 05:53:50.121848 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 05:53:50 crc kubenswrapper[4809]: I1206 05:53:50.134675 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:50 crc kubenswrapper[4809]: E1206 05:53:50.135034 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:50.63502068 +0000 UTC m=+155.524003622 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:50 crc kubenswrapper[4809]: I1206 05:53:50.237591 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:50 crc kubenswrapper[4809]: E1206 05:53:50.238155 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:50.738137801 +0000 UTC m=+155.627120743 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:50 crc kubenswrapper[4809]: I1206 05:53:50.310015 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 06 05:53:50 crc kubenswrapper[4809]: I1206 05:53:50.339564 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:50 crc kubenswrapper[4809]: E1206 05:53:50.339900 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:50.839888014 +0000 UTC m=+155.728870956 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:50 crc kubenswrapper[4809]: I1206 05:53:50.441099 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:50 crc kubenswrapper[4809]: E1206 05:53:50.441299 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:50.941274159 +0000 UTC m=+155.830257101 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:50 crc kubenswrapper[4809]: I1206 05:53:50.441354 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:50 crc kubenswrapper[4809]: E1206 05:53:50.441827 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:50.941819983 +0000 UTC m=+155.830802925 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:50 crc kubenswrapper[4809]: I1206 05:53:50.505129 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gt9kb" event={"ID":"c8d79e6f-40b0-4ecc-8501-454ba5cfe691","Type":"ContainerStarted","Data":"e7e6c95d3a1b815e358fad7f731249543ccf36339223dcd7b7245d67137ee618"} Dec 06 05:53:50 crc kubenswrapper[4809]: I1206 05:53:50.506784 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-mxv9r" event={"ID":"cea05156-8322-404a-8b33-867aeaaa905e","Type":"ContainerStarted","Data":"da9a7d368ac93021d6497f0ae05cbbbd157b496cfab9f2a614e58ecf4552040b"} Dec 06 05:53:50 crc kubenswrapper[4809]: I1206 05:53:50.511262 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kjlr8" event={"ID":"d30c4960-b4da-41cf-9d1c-771e0bc7b262","Type":"ContainerStarted","Data":"165ded9196a204f0a709456597317f28c2ce4facb90aac111955f5ab89772cdd"} Dec 06 05:53:50 crc kubenswrapper[4809]: I1206 05:53:50.512961 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ch964" event={"ID":"b6b25f1e-33d4-481d-8c73-24b95c61364f","Type":"ContainerStarted","Data":"503d77c8164c540e207da0f06d884bd034c6950fbb7110e80ccff4727804aa0c"} Dec 06 05:53:50 crc kubenswrapper[4809]: I1206 05:53:50.514124 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wcth2" event={"ID":"cdf2d3c6-659f-4dc5-a21c-d29cc86ef239","Type":"ContainerStarted","Data":"ecbdbfa2a5e8726e61ee8db569d96c96b2a39954dea58878835e57bd1228a71e"} Dec 06 05:53:50 crc kubenswrapper[4809]: I1206 05:53:50.515855 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7sxnv" event={"ID":"8063bdc5-b0d2-4a8b-91d2-12c3e1e5f938","Type":"ContainerStarted","Data":"b4c43d2168b2ff87b6142421fea4eb54ee24b4b07732025b92cc513a8c60c2a8"} Dec 06 05:53:50 crc kubenswrapper[4809]: I1206 05:53:50.517405 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dpbfw" event={"ID":"de0bb137-3c1d-4632-a7ef-6922f027a3c1","Type":"ContainerStarted","Data":"d256f6f2fc74b6e6b632f9bf24ec1581afc0ce3bde1ce161f788b88653368028"} Dec 06 05:53:50 crc kubenswrapper[4809]: I1206 05:53:50.518879 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kr48t" event={"ID":"d0b4b35c-3cb0-405d-941b-1b3946b228a0","Type":"ContainerStarted","Data":"796fd7919c08df384d807dd730de9e77f4cb3aafccda261571e186bb2522b731"} Dec 06 05:53:50 crc kubenswrapper[4809]: I1206 05:53:50.520276 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cd2tv" event={"ID":"14f07a0b-5e47-4636-9f17-cdf6f327b39d","Type":"ContainerStarted","Data":"330fd4776cb247b0a87400f2913f6e2dad2e41676979b8aa1044b4304641b556"} Dec 06 05:53:50 crc kubenswrapper[4809]: I1206 05:53:50.521720 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-czhdd" event={"ID":"e1380457-61e9-44d2-8324-eb1bcb638056","Type":"ContainerStarted","Data":"fd61c64cd8198fea6b4f3a15909044c74cebedf30fca8ed25c98b74db7f7a939"} Dec 06 05:53:50 crc kubenswrapper[4809]: I1206 05:53:50.522922 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wqnkb" event={"ID":"f6465618-cc08-4930-a290-0d1ffe135c98","Type":"ContainerStarted","Data":"a97319668a68c43e098ae11994f52df34078a56905d5b17a0889e5eeea6112e7"} Dec 06 05:53:50 crc kubenswrapper[4809]: I1206 05:53:50.523819 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"bebb2fb5-644c-49bf-9899-3bc6116cb90e","Type":"ContainerStarted","Data":"299ee7eedf129c41cdefc4a0d9ba23e43f38f2983a4ec1f5fbe1f8bf35c04e70"} Dec 06 05:53:50 crc kubenswrapper[4809]: I1206 05:53:50.524884 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416665-lsfkp" event={"ID":"c892e996-27b7-4b8d-bd92-7eac2df7702b","Type":"ContainerStarted","Data":"7ba2c355548a958343d1e97f6c244b4c1cb70b0083c1fb59afb79e37ce2ee82c"} Dec 06 05:53:50 crc kubenswrapper[4809]: I1206 05:53:50.526252 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-224mv" event={"ID":"8e290142-b2fc-492e-8763-ff41224579fc","Type":"ContainerStarted","Data":"d116376b53a80838b1f87b436f3bf9245716e9d0a8f7b95014f9a7aeffd5e784"} Dec 06 05:53:50 crc kubenswrapper[4809]: I1206 05:53:50.527366 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jjf9r" event={"ID":"e5e4e451-e3ee-434e-8e24-0c795925a48e","Type":"ContainerStarted","Data":"17e00e064844bbaa3aeb76dd9b61ea649de4b3f1cdb9d6b7b0238d8f5b30b9a1"} Dec 06 05:53:50 crc kubenswrapper[4809]: I1206 05:53:50.528406 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-m6kjl" event={"ID":"d0367f91-0d22-4fa6-8cfe-0b549698b42b","Type":"ContainerStarted","Data":"8f08f103f0fac3fcaa0a1acdd1163b12be721f2f450fcdf9bc79e945971c249d"} Dec 06 05:53:50 crc kubenswrapper[4809]: I1206 05:53:50.530004 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-s7px9" event={"ID":"0a9a3552-d735-4692-ab80-2c1526680ab9","Type":"ContainerStarted","Data":"a699e0267ce39a53293745768fbe7b7a469bd1b34fec17f62f6962aab8e300b8"} Dec 06 05:53:50 crc kubenswrapper[4809]: I1206 05:53:50.531784 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-znmpz" event={"ID":"1f78456e-8b1a-472d-beb4-aabc4a911e01","Type":"ContainerStarted","Data":"f458773302a6ffb58b7725f34b2c7a45624e239d3b5e841666a3da06e611c848"} Dec 06 05:53:50 crc kubenswrapper[4809]: I1206 05:53:50.544108 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zj6j9" event={"ID":"d844d778-6b3d-400a-8f71-db1c2e643878","Type":"ContainerStarted","Data":"cf1bd9db059b511a3f5a7eaeef54931b0329feea71e2756a67618edbbcc62ec2"} Dec 06 05:53:50 crc kubenswrapper[4809]: I1206 05:53:50.546796 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:50 crc kubenswrapper[4809]: E1206 05:53:50.546986 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:51.04695549 +0000 UTC m=+155.935938442 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:50 crc kubenswrapper[4809]: I1206 05:53:50.547153 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:50 crc kubenswrapper[4809]: I1206 05:53:50.547286 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-slt9h" event={"ID":"ee96792c-e168-4c58-a477-cd88c14da612","Type":"ContainerStarted","Data":"349bce7f833411f0b73b5cfe568b8093ff256296c433415448588ec0b9d28efc"} Dec 06 05:53:50 crc kubenswrapper[4809]: E1206 05:53:50.547718 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:51.047699731 +0000 UTC m=+155.936682743 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:50 crc kubenswrapper[4809]: I1206 05:53:50.550056 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-drj7v" event={"ID":"63b0cea8-1716-4c17-9f87-f10e776cf47f","Type":"ContainerStarted","Data":"8f3b9a9c1b13b219e3f1d9f085e79316d913e38c3eedb5f2b28595ab139cd59c"} Dec 06 05:53:50 crc kubenswrapper[4809]: I1206 05:53:50.551817 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qqv6x" event={"ID":"8a225b5c-4ad5-4f08-a0cb-ae38a8fa83d4","Type":"ContainerStarted","Data":"2af066a5c9a9002f590516c228f86045539671a2d2f7308f16b61703c27f6c95"} Dec 06 05:53:50 crc kubenswrapper[4809]: I1206 05:53:50.553923 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-852bd" event={"ID":"6b48dffc-c51b-48e2-b22b-e809176c8830","Type":"ContainerStarted","Data":"247566479e4ab85b2228e12f3f120b4e523aaf5bab2d4dbda63b2b58de5ecca4"} Dec 06 05:53:50 crc kubenswrapper[4809]: I1206 05:53:50.556091 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-s7q7v" event={"ID":"f4008f48-e17b-492f-8fa4-b09fb9d2ce70","Type":"ContainerStarted","Data":"1f26bc50a8f7a21bf0e873dfce22072ba88e2541095cbff1a4c4cffaec497319"} Dec 06 05:53:50 crc kubenswrapper[4809]: I1206 05:53:50.599203 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-7qpjs" podStartSLOduration=136.599181308 podStartE2EDuration="2m16.599181308s" podCreationTimestamp="2025-12-06 05:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:53:50.58277694 +0000 UTC m=+155.471759902" watchObservedRunningTime="2025-12-06 05:53:50.599181308 +0000 UTC m=+155.488164250" Dec 06 05:53:50 crc kubenswrapper[4809]: I1206 05:53:50.600322 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sdg4b" podStartSLOduration=136.600312009 podStartE2EDuration="2m16.600312009s" podCreationTimestamp="2025-12-06 05:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:53:50.598261674 +0000 UTC m=+155.487244626" watchObservedRunningTime="2025-12-06 05:53:50.600312009 +0000 UTC m=+155.489294951" Dec 06 05:53:50 crc kubenswrapper[4809]: I1206 05:53:50.648251 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:50 crc kubenswrapper[4809]: E1206 05:53:50.648467 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:51.148422616 +0000 UTC m=+156.037405568 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:50 crc kubenswrapper[4809]: I1206 05:53:50.648729 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:50 crc kubenswrapper[4809]: E1206 05:53:50.650056 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:51.15003763 +0000 UTC m=+156.039020672 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:50 crc kubenswrapper[4809]: I1206 05:53:50.751385 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:50 crc kubenswrapper[4809]: E1206 05:53:50.751789 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:51.251767723 +0000 UTC m=+156.140750665 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:50 crc kubenswrapper[4809]: I1206 05:53:50.852662 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:50 crc kubenswrapper[4809]: E1206 05:53:50.853136 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:51.353116606 +0000 UTC m=+156.242099628 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:50 crc kubenswrapper[4809]: I1206 05:53:50.954318 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:50 crc kubenswrapper[4809]: E1206 05:53:50.954460 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:51.454426808 +0000 UTC m=+156.343409750 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:50 crc kubenswrapper[4809]: I1206 05:53:50.954721 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:50 crc kubenswrapper[4809]: E1206 05:53:50.955077 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:51.455064146 +0000 UTC m=+156.344047088 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.055904 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:51 crc kubenswrapper[4809]: E1206 05:53:51.056333 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:51.556311505 +0000 UTC m=+156.445294457 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.088438 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.157684 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:51 crc kubenswrapper[4809]: E1206 05:53:51.158176 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:51.658159592 +0000 UTC m=+156.547142534 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.258870 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:51 crc kubenswrapper[4809]: E1206 05:53:51.259070 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:51.759043482 +0000 UTC m=+156.648026434 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.259345 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:51 crc kubenswrapper[4809]: E1206 05:53:51.259602 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:51.759590717 +0000 UTC m=+156.648573659 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.363691 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:51 crc kubenswrapper[4809]: E1206 05:53:51.364251 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:51.86423337 +0000 UTC m=+156.753216312 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.465634 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:51 crc kubenswrapper[4809]: E1206 05:53:51.465975 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:51.965959373 +0000 UTC m=+156.854942315 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.469276 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-24qxt"] Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.470157 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-24qxt" Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.471883 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.482351 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-24qxt"] Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.567124 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.567285 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/893020b2-fa30-4371-86f9-a585517839ce-utilities\") pod \"community-operators-24qxt\" (UID: \"893020b2-fa30-4371-86f9-a585517839ce\") " pod="openshift-marketplace/community-operators-24qxt" Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.567342 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/893020b2-fa30-4371-86f9-a585517839ce-catalog-content\") pod \"community-operators-24qxt\" (UID: \"893020b2-fa30-4371-86f9-a585517839ce\") " pod="openshift-marketplace/community-operators-24qxt" Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.567395 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ld72t\" (UniqueName: \"kubernetes.io/projected/893020b2-fa30-4371-86f9-a585517839ce-kube-api-access-ld72t\") pod \"community-operators-24qxt\" (UID: \"893020b2-fa30-4371-86f9-a585517839ce\") " pod="openshift-marketplace/community-operators-24qxt" Dec 06 05:53:51 crc kubenswrapper[4809]: E1206 05:53:51.567483 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:52.06746881 +0000 UTC m=+156.956451752 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.584075 4809 generic.go:334] "Generic (PLEG): container finished" podID="09a7af3d-6bf2-4b7a-93d9-cecc73d993a1" containerID="0cb2714e8d310b4bf6971dbfc352d4adaf719f77598d015efa3d272df90c8b27" exitCode=0 Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.584163 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-tphf5" event={"ID":"09a7af3d-6bf2-4b7a-93d9-cecc73d993a1","Type":"ContainerDied","Data":"0cb2714e8d310b4bf6971dbfc352d4adaf719f77598d015efa3d272df90c8b27"} Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.588964 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-p2lkt" event={"ID":"c598f580-a438-4015-9a8a-f98b3b2b0b3d","Type":"ContainerStarted","Data":"2352c784a1e72e31f3c9a8dd8187182dca44b4c3566b5f8b4714d59527f6ded5"} Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.603324 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"bebb2fb5-644c-49bf-9899-3bc6116cb90e","Type":"ContainerStarted","Data":"d3548183cf720c9be50500538d3af4646621c6053a02f42f3a750d773f63ed1b"} Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.625108 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nsqq5" event={"ID":"87db7503-6e3e-49b1-bc84-5230a444e8f6","Type":"ContainerStarted","Data":"3ac7d815b2af7d674f7c9809a8414412b2b9cf149476965e71f47471ce5654ca"} Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.629302 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-jn446" event={"ID":"0a9baebe-98e6-44be-ad16-882e47f09248","Type":"ContainerStarted","Data":"65e9f6726d60aa3c96833be77b3799ee27896dbd41ffe685acd2f6a01d1fc947"} Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.629926 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.631486 4809 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-jn446 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.38:6443/healthz\": dial tcp 10.217.0.38:6443: connect: connection refused" start-of-body= Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.631522 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-jn446" podUID="0a9baebe-98e6-44be-ad16-882e47f09248" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.38:6443/healthz\": dial tcp 10.217.0.38:6443: connect: connection refused" Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.670754 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/893020b2-fa30-4371-86f9-a585517839ce-utilities\") pod \"community-operators-24qxt\" (UID: \"893020b2-fa30-4371-86f9-a585517839ce\") " pod="openshift-marketplace/community-operators-24qxt" Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.670830 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.670860 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/893020b2-fa30-4371-86f9-a585517839ce-catalog-content\") pod \"community-operators-24qxt\" (UID: \"893020b2-fa30-4371-86f9-a585517839ce\") " pod="openshift-marketplace/community-operators-24qxt" Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.670941 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ld72t\" (UniqueName: \"kubernetes.io/projected/893020b2-fa30-4371-86f9-a585517839ce-kube-api-access-ld72t\") pod \"community-operators-24qxt\" (UID: \"893020b2-fa30-4371-86f9-a585517839ce\") " pod="openshift-marketplace/community-operators-24qxt" Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.671860 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/893020b2-fa30-4371-86f9-a585517839ce-utilities\") pod \"community-operators-24qxt\" (UID: \"893020b2-fa30-4371-86f9-a585517839ce\") " pod="openshift-marketplace/community-operators-24qxt" Dec 06 05:53:51 crc kubenswrapper[4809]: E1206 05:53:51.672275 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:52.172259427 +0000 UTC m=+157.061242369 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.672732 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/893020b2-fa30-4371-86f9-a585517839ce-catalog-content\") pod \"community-operators-24qxt\" (UID: \"893020b2-fa30-4371-86f9-a585517839ce\") " pod="openshift-marketplace/community-operators-24qxt" Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.674582 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-jn446" podStartSLOduration=137.67455787 podStartE2EDuration="2m17.67455787s" podCreationTimestamp="2025-12-06 05:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:53:51.659694294 +0000 UTC m=+156.548677236" watchObservedRunningTime="2025-12-06 05:53:51.67455787 +0000 UTC m=+156.563540812" Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.674730 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zk7tm"] Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.679745 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zk7tm" Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.696208 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.698708 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zk7tm"] Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.705302 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8ssrl" event={"ID":"d233521e-75db-4da3-9cd3-0480a99eba90","Type":"ContainerStarted","Data":"f03bac0a594758334f8ee19abef723f5dabc8fd7a10461aa30acdaad4b2d1330"} Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.732253 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ld72t\" (UniqueName: \"kubernetes.io/projected/893020b2-fa30-4371-86f9-a585517839ce-kube-api-access-ld72t\") pod \"community-operators-24qxt\" (UID: \"893020b2-fa30-4371-86f9-a585517839ce\") " pod="openshift-marketplace/community-operators-24qxt" Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.749986 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nff6m" event={"ID":"45f3dc48-d71b-47f8-8567-90df740103b5","Type":"ContainerStarted","Data":"4507cc1b69f02c2fd4dcfe1dbeee2ae4957f6de7013d22737bc865859b71ae48"} Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.750950 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nff6m" Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.753243 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-m6kjl" event={"ID":"d0367f91-0d22-4fa6-8cfe-0b549698b42b","Type":"ContainerStarted","Data":"34091297a1030d940d9c0427ac15cdef1df1a95334103a007fef35fdd9ad6263"} Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.756577 4809 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-nff6m container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" start-of-body= Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.756618 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nff6m" podUID="45f3dc48-d71b-47f8-8567-90df740103b5" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.771918 4809 generic.go:334] "Generic (PLEG): container finished" podID="14f07a0b-5e47-4636-9f17-cdf6f327b39d" containerID="330fd4776cb247b0a87400f2913f6e2dad2e41676979b8aa1044b4304641b556" exitCode=0 Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.772018 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cd2tv" event={"ID":"14f07a0b-5e47-4636-9f17-cdf6f327b39d","Type":"ContainerDied","Data":"330fd4776cb247b0a87400f2913f6e2dad2e41676979b8aa1044b4304641b556"} Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.773258 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.773435 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwbq2\" (UniqueName: \"kubernetes.io/projected/5e67c2b8-8bd1-466c-98d8-e93d411162ba-kube-api-access-qwbq2\") pod \"certified-operators-zk7tm\" (UID: \"5e67c2b8-8bd1-466c-98d8-e93d411162ba\") " pod="openshift-marketplace/certified-operators-zk7tm" Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.773522 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e67c2b8-8bd1-466c-98d8-e93d411162ba-catalog-content\") pod \"certified-operators-zk7tm\" (UID: \"5e67c2b8-8bd1-466c-98d8-e93d411162ba\") " pod="openshift-marketplace/certified-operators-zk7tm" Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.773576 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e67c2b8-8bd1-466c-98d8-e93d411162ba-utilities\") pod \"certified-operators-zk7tm\" (UID: \"5e67c2b8-8bd1-466c-98d8-e93d411162ba\") " pod="openshift-marketplace/certified-operators-zk7tm" Dec 06 05:53:51 crc kubenswrapper[4809]: E1206 05:53:51.774123 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:52.274105314 +0000 UTC m=+157.163088266 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.788291 4809 generic.go:334] "Generic (PLEG): container finished" podID="8a225b5c-4ad5-4f08-a0cb-ae38a8fa83d4" containerID="2af066a5c9a9002f590516c228f86045539671a2d2f7308f16b61703c27f6c95" exitCode=0 Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.789303 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qqv6x" event={"ID":"8a225b5c-4ad5-4f08-a0cb-ae38a8fa83d4","Type":"ContainerDied","Data":"2af066a5c9a9002f590516c228f86045539671a2d2f7308f16b61703c27f6c95"} Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.793055 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jjf9r" Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.793081 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pkq7h" Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.793091 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-slt9h" Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.793100 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-224mv" Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.793841 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gt9kb" Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.794063 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-hlvjb" Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.798729 4809 patch_prober.go:28] interesting pod/downloads-7954f5f757-224mv container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.798769 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-224mv" podUID="8e290142-b2fc-492e-8763-ff41224579fc" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.37:8080/\": dial tcp 10.217.0.37:8080: connect: connection refused" Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.798834 4809 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-slt9h container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" start-of-body= Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.798847 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-slt9h" podUID="ee96792c-e168-4c58-a477-cd88c14da612" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.805481 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-24qxt" Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.819458 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nff6m" podStartSLOduration=136.819441434 podStartE2EDuration="2m16.819441434s" podCreationTimestamp="2025-12-06 05:51:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:53:51.785584698 +0000 UTC m=+156.674567650" watchObservedRunningTime="2025-12-06 05:53:51.819441434 +0000 UTC m=+156.708424376" Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.833363 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pkq7h" Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.833408 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gt9kb" Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.864460 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-n6z57" podStartSLOduration=136.864439985 podStartE2EDuration="2m16.864439985s" podCreationTimestamp="2025-12-06 05:51:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:53:51.857496825 +0000 UTC m=+156.746479767" watchObservedRunningTime="2025-12-06 05:53:51.864439985 +0000 UTC m=+156.753422927" Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.869742 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9zm48"] Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.870905 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9zm48" Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.879781 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.880210 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwbq2\" (UniqueName: \"kubernetes.io/projected/5e67c2b8-8bd1-466c-98d8-e93d411162ba-kube-api-access-qwbq2\") pod \"certified-operators-zk7tm\" (UID: \"5e67c2b8-8bd1-466c-98d8-e93d411162ba\") " pod="openshift-marketplace/certified-operators-zk7tm" Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.880913 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e67c2b8-8bd1-466c-98d8-e93d411162ba-catalog-content\") pod \"certified-operators-zk7tm\" (UID: \"5e67c2b8-8bd1-466c-98d8-e93d411162ba\") " pod="openshift-marketplace/certified-operators-zk7tm" Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.881504 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e67c2b8-8bd1-466c-98d8-e93d411162ba-utilities\") pod \"certified-operators-zk7tm\" (UID: \"5e67c2b8-8bd1-466c-98d8-e93d411162ba\") " pod="openshift-marketplace/certified-operators-zk7tm" Dec 06 05:53:51 crc kubenswrapper[4809]: E1206 05:53:51.904527 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:52.404510232 +0000 UTC m=+157.293493174 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.908956 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e67c2b8-8bd1-466c-98d8-e93d411162ba-catalog-content\") pod \"certified-operators-zk7tm\" (UID: \"5e67c2b8-8bd1-466c-98d8-e93d411162ba\") " pod="openshift-marketplace/certified-operators-zk7tm" Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.914866 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gt9kb" podStartSLOduration=136.914840984 podStartE2EDuration="2m16.914840984s" podCreationTimestamp="2025-12-06 05:51:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:53:51.903044442 +0000 UTC m=+156.792027384" watchObservedRunningTime="2025-12-06 05:53:51.914840984 +0000 UTC m=+156.803823926" Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.915826 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9zm48"] Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.917312 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e67c2b8-8bd1-466c-98d8-e93d411162ba-utilities\") pod \"certified-operators-zk7tm\" (UID: \"5e67c2b8-8bd1-466c-98d8-e93d411162ba\") " pod="openshift-marketplace/certified-operators-zk7tm" Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.968367 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwbq2\" (UniqueName: \"kubernetes.io/projected/5e67c2b8-8bd1-466c-98d8-e93d411162ba-kube-api-access-qwbq2\") pod \"certified-operators-zk7tm\" (UID: \"5e67c2b8-8bd1-466c-98d8-e93d411162ba\") " pod="openshift-marketplace/certified-operators-zk7tm" Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.982818 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.983224 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c822a806-352f-4b4f-8549-051ff20c4ca2-utilities\") pod \"community-operators-9zm48\" (UID: \"c822a806-352f-4b4f-8549-051ff20c4ca2\") " pod="openshift-marketplace/community-operators-9zm48" Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.983265 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7f6b\" (UniqueName: \"kubernetes.io/projected/c822a806-352f-4b4f-8549-051ff20c4ca2-kube-api-access-p7f6b\") pod \"community-operators-9zm48\" (UID: \"c822a806-352f-4b4f-8549-051ff20c4ca2\") " pod="openshift-marketplace/community-operators-9zm48" Dec 06 05:53:51 crc kubenswrapper[4809]: I1206 05:53:51.983300 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c822a806-352f-4b4f-8549-051ff20c4ca2-catalog-content\") pod \"community-operators-9zm48\" (UID: \"c822a806-352f-4b4f-8549-051ff20c4ca2\") " pod="openshift-marketplace/community-operators-9zm48" Dec 06 05:53:51 crc kubenswrapper[4809]: E1206 05:53:51.983447 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:52.483432971 +0000 UTC m=+157.372415913 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.017953 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kr48t" podStartSLOduration=138.017926355 podStartE2EDuration="2m18.017926355s" podCreationTimestamp="2025-12-06 05:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:53:52.017203824 +0000 UTC m=+156.906186766" watchObservedRunningTime="2025-12-06 05:53:52.017926355 +0000 UTC m=+156.906909287" Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.026796 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zj6j9" podStartSLOduration=138.026778736 podStartE2EDuration="2m18.026778736s" podCreationTimestamp="2025-12-06 05:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:53:51.97100509 +0000 UTC m=+156.859988042" watchObservedRunningTime="2025-12-06 05:53:52.026778736 +0000 UTC m=+156.915761678" Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.049215 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-slt9h" podStartSLOduration=137.04919664 podStartE2EDuration="2m17.04919664s" podCreationTimestamp="2025-12-06 05:51:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:53:52.047442442 +0000 UTC m=+156.936425384" watchObservedRunningTime="2025-12-06 05:53:52.04919664 +0000 UTC m=+156.938179582" Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.088101 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.088169 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c822a806-352f-4b4f-8549-051ff20c4ca2-utilities\") pod \"community-operators-9zm48\" (UID: \"c822a806-352f-4b4f-8549-051ff20c4ca2\") " pod="openshift-marketplace/community-operators-9zm48" Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.088194 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7f6b\" (UniqueName: \"kubernetes.io/projected/c822a806-352f-4b4f-8549-051ff20c4ca2-kube-api-access-p7f6b\") pod \"community-operators-9zm48\" (UID: \"c822a806-352f-4b4f-8549-051ff20c4ca2\") " pod="openshift-marketplace/community-operators-9zm48" Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.088223 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c822a806-352f-4b4f-8549-051ff20c4ca2-catalog-content\") pod \"community-operators-9zm48\" (UID: \"c822a806-352f-4b4f-8549-051ff20c4ca2\") " pod="openshift-marketplace/community-operators-9zm48" Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.088779 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c822a806-352f-4b4f-8549-051ff20c4ca2-catalog-content\") pod \"community-operators-9zm48\" (UID: \"c822a806-352f-4b4f-8549-051ff20c4ca2\") " pod="openshift-marketplace/community-operators-9zm48" Dec 06 05:53:52 crc kubenswrapper[4809]: E1206 05:53:52.089104 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:52.589086622 +0000 UTC m=+157.478069564 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.089544 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c822a806-352f-4b4f-8549-051ff20c4ca2-utilities\") pod \"community-operators-9zm48\" (UID: \"c822a806-352f-4b4f-8549-051ff20c4ca2\") " pod="openshift-marketplace/community-operators-9zm48" Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.098747 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-hlvjb" podStartSLOduration=138.098730385 podStartE2EDuration="2m18.098730385s" podCreationTimestamp="2025-12-06 05:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:53:52.094411107 +0000 UTC m=+156.983394049" watchObservedRunningTime="2025-12-06 05:53:52.098730385 +0000 UTC m=+156.987713347" Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.101898 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dvwjq"] Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.102865 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dvwjq"] Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.103134 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dvwjq" Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.116329 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7f6b\" (UniqueName: \"kubernetes.io/projected/c822a806-352f-4b4f-8549-051ff20c4ca2-kube-api-access-p7f6b\") pod \"community-operators-9zm48\" (UID: \"c822a806-352f-4b4f-8549-051ff20c4ca2\") " pod="openshift-marketplace/community-operators-9zm48" Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.162515 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zk7tm" Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.179348 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-224mv" podStartSLOduration=138.17932895 podStartE2EDuration="2m18.17932895s" podCreationTimestamp="2025-12-06 05:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:53:52.154434109 +0000 UTC m=+157.043417051" watchObservedRunningTime="2025-12-06 05:53:52.17932895 +0000 UTC m=+157.068311892" Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.189484 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.189773 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9b9l\" (UniqueName: \"kubernetes.io/projected/a83cbcc6-62ac-4814-9fce-eb7ea465bcbf-kube-api-access-j9b9l\") pod \"certified-operators-dvwjq\" (UID: \"a83cbcc6-62ac-4814-9fce-eb7ea465bcbf\") " pod="openshift-marketplace/certified-operators-dvwjq" Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.189829 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a83cbcc6-62ac-4814-9fce-eb7ea465bcbf-utilities\") pod \"certified-operators-dvwjq\" (UID: \"a83cbcc6-62ac-4814-9fce-eb7ea465bcbf\") " pod="openshift-marketplace/certified-operators-dvwjq" Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.189868 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a83cbcc6-62ac-4814-9fce-eb7ea465bcbf-catalog-content\") pod \"certified-operators-dvwjq\" (UID: \"a83cbcc6-62ac-4814-9fce-eb7ea465bcbf\") " pod="openshift-marketplace/certified-operators-dvwjq" Dec 06 05:53:52 crc kubenswrapper[4809]: E1206 05:53:52.192836 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:52.692768018 +0000 UTC m=+157.581750970 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.238587 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7sxnv" podStartSLOduration=137.23856524 podStartE2EDuration="2m17.23856524s" podCreationTimestamp="2025-12-06 05:51:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:53:52.235362773 +0000 UTC m=+157.124345735" watchObservedRunningTime="2025-12-06 05:53:52.23856524 +0000 UTC m=+157.127548192" Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.239187 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kjlr8" podStartSLOduration=137.239180818 podStartE2EDuration="2m17.239180818s" podCreationTimestamp="2025-12-06 05:51:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:53:52.184302687 +0000 UTC m=+157.073285619" watchObservedRunningTime="2025-12-06 05:53:52.239180818 +0000 UTC m=+157.128163760" Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.275653 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9zm48" Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.280297 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wqnkb" podStartSLOduration=137.280273402 podStartE2EDuration="2m17.280273402s" podCreationTimestamp="2025-12-06 05:51:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:53:52.279010127 +0000 UTC m=+157.167993069" watchObservedRunningTime="2025-12-06 05:53:52.280273402 +0000 UTC m=+157.169256344" Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.281882 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-24qxt"] Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.299794 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.299842 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9b9l\" (UniqueName: \"kubernetes.io/projected/a83cbcc6-62ac-4814-9fce-eb7ea465bcbf-kube-api-access-j9b9l\") pod \"certified-operators-dvwjq\" (UID: \"a83cbcc6-62ac-4814-9fce-eb7ea465bcbf\") " pod="openshift-marketplace/certified-operators-dvwjq" Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.299889 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a83cbcc6-62ac-4814-9fce-eb7ea465bcbf-utilities\") pod \"certified-operators-dvwjq\" (UID: \"a83cbcc6-62ac-4814-9fce-eb7ea465bcbf\") " pod="openshift-marketplace/certified-operators-dvwjq" Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.299923 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a83cbcc6-62ac-4814-9fce-eb7ea465bcbf-catalog-content\") pod \"certified-operators-dvwjq\" (UID: \"a83cbcc6-62ac-4814-9fce-eb7ea465bcbf\") " pod="openshift-marketplace/certified-operators-dvwjq" Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.300403 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a83cbcc6-62ac-4814-9fce-eb7ea465bcbf-catalog-content\") pod \"certified-operators-dvwjq\" (UID: \"a83cbcc6-62ac-4814-9fce-eb7ea465bcbf\") " pod="openshift-marketplace/certified-operators-dvwjq" Dec 06 05:53:52 crc kubenswrapper[4809]: E1206 05:53:52.300690 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:52.80067552 +0000 UTC m=+157.689658462 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.301374 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a83cbcc6-62ac-4814-9fce-eb7ea465bcbf-utilities\") pod \"certified-operators-dvwjq\" (UID: \"a83cbcc6-62ac-4814-9fce-eb7ea465bcbf\") " pod="openshift-marketplace/certified-operators-dvwjq" Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.318460 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jjf9r" podStartSLOduration=137.318441156 podStartE2EDuration="2m17.318441156s" podCreationTimestamp="2025-12-06 05:51:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:53:52.316111813 +0000 UTC m=+157.205094755" watchObservedRunningTime="2025-12-06 05:53:52.318441156 +0000 UTC m=+157.207424098" Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.327316 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9b9l\" (UniqueName: \"kubernetes.io/projected/a83cbcc6-62ac-4814-9fce-eb7ea465bcbf-kube-api-access-j9b9l\") pod \"certified-operators-dvwjq\" (UID: \"a83cbcc6-62ac-4814-9fce-eb7ea465bcbf\") " pod="openshift-marketplace/certified-operators-dvwjq" Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.351244 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-mxv9r" podStartSLOduration=137.351226133 podStartE2EDuration="2m17.351226133s" podCreationTimestamp="2025-12-06 05:51:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:53:52.350364639 +0000 UTC m=+157.239347581" watchObservedRunningTime="2025-12-06 05:53:52.351226133 +0000 UTC m=+157.240209075" Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.398425 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pkq7h" podStartSLOduration=137.398408554 podStartE2EDuration="2m17.398408554s" podCreationTimestamp="2025-12-06 05:51:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:53:52.388358249 +0000 UTC m=+157.277341191" watchObservedRunningTime="2025-12-06 05:53:52.398408554 +0000 UTC m=+157.287391496" Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.401384 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:52 crc kubenswrapper[4809]: E1206 05:53:52.401609 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:52.901590931 +0000 UTC m=+157.790573863 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.401730 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:52 crc kubenswrapper[4809]: E1206 05:53:52.402112 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:52.902100325 +0000 UTC m=+157.791083267 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.449552 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-g7pfc" podStartSLOduration=11.449529803 podStartE2EDuration="11.449529803s" podCreationTimestamp="2025-12-06 05:53:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:53:52.441308518 +0000 UTC m=+157.330291460" watchObservedRunningTime="2025-12-06 05:53:52.449529803 +0000 UTC m=+157.338512745" Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.494636 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dvwjq" Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.505498 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:52 crc kubenswrapper[4809]: E1206 05:53:52.505872 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:53.005855263 +0000 UTC m=+157.894838205 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.509621 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dpbfw" podStartSLOduration=137.509608956 podStartE2EDuration="2m17.509608956s" podCreationTimestamp="2025-12-06 05:51:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:53:52.494296147 +0000 UTC m=+157.383279089" watchObservedRunningTime="2025-12-06 05:53:52.509608956 +0000 UTC m=+157.398591898" Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.546387 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29416665-lsfkp" podStartSLOduration=138.546372382 podStartE2EDuration="2m18.546372382s" podCreationTimestamp="2025-12-06 05:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:53:52.510481251 +0000 UTC m=+157.399464203" watchObservedRunningTime="2025-12-06 05:53:52.546372382 +0000 UTC m=+157.435355324" Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.549643 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-hlvjb" Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.573694 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-drj7v" podStartSLOduration=137.573674279 podStartE2EDuration="2m17.573674279s" podCreationTimestamp="2025-12-06 05:51:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:53:52.549761115 +0000 UTC m=+157.438744057" watchObservedRunningTime="2025-12-06 05:53:52.573674279 +0000 UTC m=+157.462657221" Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.574726 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-czhdd" podStartSLOduration=10.574721137000001 podStartE2EDuration="10.574721137s" podCreationTimestamp="2025-12-06 05:53:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:53:52.570604775 +0000 UTC m=+157.459587717" watchObservedRunningTime="2025-12-06 05:53:52.574721137 +0000 UTC m=+157.463704089" Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.610828 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:52 crc kubenswrapper[4809]: E1206 05:53:52.611425 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:53.111413132 +0000 UTC m=+158.000396074 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.619951 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-n6z57" Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.637169 4809 patch_prober.go:28] interesting pod/router-default-5444994796-n6z57 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 05:53:52 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Dec 06 05:53:52 crc kubenswrapper[4809]: [+]process-running ok Dec 06 05:53:52 crc kubenswrapper[4809]: healthz check failed Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.637233 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n6z57" podUID="e5356f1d-5afd-47bd-89af-72abfa8f901f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.726704 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:52 crc kubenswrapper[4809]: E1206 05:53:52.727444 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:53.227428405 +0000 UTC m=+158.116411347 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.790036 4809 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-jjf9r container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.790106 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jjf9r" podUID="e5e4e451-e3ee-434e-8e24-0c795925a48e" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.34:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.832695 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:52 crc kubenswrapper[4809]: E1206 05:53:52.833115 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:53.333103367 +0000 UTC m=+158.222086309 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.858150 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-s7px9" event={"ID":"0a9a3552-d735-4692-ab80-2c1526680ab9","Type":"ContainerStarted","Data":"5a3c756e2288b744f0e13b089d3857cbf2d08244d9a167bf1acf2ff7063250d1"} Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.891498 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nsqq5" event={"ID":"87db7503-6e3e-49b1-bc84-5230a444e8f6","Type":"ContainerStarted","Data":"d8963e183b501f2eb924543b82012201263ffa6e0fb52a615d07067b3b4c8b02"} Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.933701 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:52 crc kubenswrapper[4809]: E1206 05:53:52.934329 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:53.434299955 +0000 UTC m=+158.323282897 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.935373 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ch964" event={"ID":"b6b25f1e-33d4-481d-8c73-24b95c61364f","Type":"ContainerStarted","Data":"330c20f6df3e7bfaaaff722d2f7a2e357959ecf53a67cb350ddcb2a4fe611e77"} Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.936429 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ch964" Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.958698 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nsqq5" podStartSLOduration=137.958669343 podStartE2EDuration="2m17.958669343s" podCreationTimestamp="2025-12-06 05:51:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:53:52.9392021 +0000 UTC m=+157.828185042" watchObservedRunningTime="2025-12-06 05:53:52.958669343 +0000 UTC m=+157.847652285" Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.981306 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ch964" podStartSLOduration=137.981288861 podStartE2EDuration="2m17.981288861s" podCreationTimestamp="2025-12-06 05:51:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:53:52.980038007 +0000 UTC m=+157.869020949" watchObservedRunningTime="2025-12-06 05:53:52.981288861 +0000 UTC m=+157.870271803" Dec 06 05:53:52 crc kubenswrapper[4809]: I1206 05:53:52.989246 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-m6kjl" event={"ID":"d0367f91-0d22-4fa6-8cfe-0b549698b42b","Type":"ContainerStarted","Data":"0a511e6fa2dc602b974a4102a9240632fb27967c97ff22e60d5ba63ab1409846"} Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.038069 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:53 crc kubenswrapper[4809]: E1206 05:53:53.039096 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:53.539081802 +0000 UTC m=+158.428064804 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.047164 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zk7tm"] Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.047913 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-m6kjl" podStartSLOduration=139.047895243 podStartE2EDuration="2m19.047895243s" podCreationTimestamp="2025-12-06 05:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:53:53.043275167 +0000 UTC m=+157.932258129" watchObservedRunningTime="2025-12-06 05:53:53.047895243 +0000 UTC m=+157.936878185" Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.072513 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cd2tv" event={"ID":"14f07a0b-5e47-4636-9f17-cdf6f327b39d","Type":"ContainerStarted","Data":"6d086d3b1ddcff4233dad38cfe83ac8d869ad2865e97ef107191ba04a49ebf20"} Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.112402 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cd2tv" podStartSLOduration=138.112380147 podStartE2EDuration="2m18.112380147s" podCreationTimestamp="2025-12-06 05:51:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:53:53.098501567 +0000 UTC m=+157.987484509" watchObservedRunningTime="2025-12-06 05:53:53.112380147 +0000 UTC m=+158.001363089" Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.131332 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-s7q7v" event={"ID":"f4008f48-e17b-492f-8fa4-b09fb9d2ce70","Type":"ContainerStarted","Data":"3a1e8a9d4184bbdcf446e20df8edff2c50ad980548d89beae0bbe88ba9d1aebc"} Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.142514 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:53 crc kubenswrapper[4809]: E1206 05:53:53.143693 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:53.643674804 +0000 UTC m=+158.532657746 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.159194 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wcth2" event={"ID":"cdf2d3c6-659f-4dc5-a21c-d29cc86ef239","Type":"ContainerStarted","Data":"da04d5a7af09107069e77dc83b0c68d26ad3ba89583943551e7463bfd5350d23"} Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.167963 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-s7q7v" podStartSLOduration=138.167948068 podStartE2EDuration="2m18.167948068s" podCreationTimestamp="2025-12-06 05:51:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:53:53.165824 +0000 UTC m=+158.054806942" watchObservedRunningTime="2025-12-06 05:53:53.167948068 +0000 UTC m=+158.056931010" Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.183975 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9zm48"] Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.186011 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-24qxt" event={"ID":"893020b2-fa30-4371-86f9-a585517839ce","Type":"ContainerStarted","Data":"2ad2d97bb0902f7353810c067603e67f7dc4a0c5339e0a9f536eb8729e07d753"} Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.186049 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-24qxt" event={"ID":"893020b2-fa30-4371-86f9-a585517839ce","Type":"ContainerStarted","Data":"7dbcb9c7799bcc1e1ab2ed512b6573baac17ee3f42d68c69e661de5457fd6989"} Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.194325 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.213210 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-927z2" event={"ID":"d7d02f2c-c052-4320-bac4-394fb40c57d2","Type":"ContainerStarted","Data":"66f46bce70573bde8a8d9673a2a8a16e5dcad72d538a5e101707ba714f671f28"} Dec 06 05:53:53 crc kubenswrapper[4809]: W1206 05:53:53.218224 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc822a806_352f_4b4f_8549_051ff20c4ca2.slice/crio-2f5eb5c5202712e416c3fec6986590204363bf55169b18391542dc759ee946f8 WatchSource:0}: Error finding container 2f5eb5c5202712e416c3fec6986590204363bf55169b18391542dc759ee946f8: Status 404 returned error can't find the container with id 2f5eb5c5202712e416c3fec6986590204363bf55169b18391542dc759ee946f8 Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.220300 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wcth2" podStartSLOduration=138.220279979 podStartE2EDuration="2m18.220279979s" podCreationTimestamp="2025-12-06 05:51:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:53:53.203462959 +0000 UTC m=+158.092445901" watchObservedRunningTime="2025-12-06 05:53:53.220279979 +0000 UTC m=+158.109262921" Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.225234 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qqv6x" event={"ID":"8a225b5c-4ad5-4f08-a0cb-ae38a8fa83d4","Type":"ContainerStarted","Data":"e51fca1443665c2a16e03e8f0a50922fa24e5a1de41ce2dcd1b0461b4e23cea0"} Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.225801 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qqv6x" Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.244309 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:53 crc kubenswrapper[4809]: E1206 05:53:53.246185 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:53.746173978 +0000 UTC m=+158.635156920 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.246300 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-znmpz" event={"ID":"1f78456e-8b1a-472d-beb4-aabc4a911e01","Type":"ContainerStarted","Data":"06cf201916e92b85ffda0bc6a2043b7fa491c88ee877a32b2247f0306b335d5e"} Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.253369 4809 generic.go:334] "Generic (PLEG): container finished" podID="bebb2fb5-644c-49bf-9899-3bc6116cb90e" containerID="d3548183cf720c9be50500538d3af4646621c6053a02f42f3a750d773f63ed1b" exitCode=0 Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.253454 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"bebb2fb5-644c-49bf-9899-3bc6116cb90e","Type":"ContainerDied","Data":"d3548183cf720c9be50500538d3af4646621c6053a02f42f3a750d773f63ed1b"} Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.259596 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-852bd" event={"ID":"6b48dffc-c51b-48e2-b22b-e809176c8830","Type":"ContainerStarted","Data":"3acab5b430fa92b8b01b1eef20369803e4f22db3d1e6e4c474afe3387ade6c60"} Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.261088 4809 patch_prober.go:28] interesting pod/downloads-7954f5f757-224mv container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.261132 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-224mv" podUID="8e290142-b2fc-492e-8763-ff41224579fc" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.37:8080/\": dial tcp 10.217.0.37:8080: connect: connection refused" Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.262480 4809 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-slt9h container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" start-of-body= Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.262504 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-slt9h" podUID="ee96792c-e168-4c58-a477-cd88c14da612" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.271645 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-927z2" podStartSLOduration=138.271628644 podStartE2EDuration="2m18.271628644s" podCreationTimestamp="2025-12-06 05:51:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:53:53.271325986 +0000 UTC m=+158.160308928" watchObservedRunningTime="2025-12-06 05:53:53.271628644 +0000 UTC m=+158.160611586" Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.286662 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jjf9r" Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.323137 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qqv6x" podStartSLOduration=139.323109593 podStartE2EDuration="2m19.323109593s" podCreationTimestamp="2025-12-06 05:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:53:53.321455128 +0000 UTC m=+158.210438080" watchObservedRunningTime="2025-12-06 05:53:53.323109593 +0000 UTC m=+158.212092535" Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.349619 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:53 crc kubenswrapper[4809]: E1206 05:53:53.351272 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:53.851255333 +0000 UTC m=+158.740238265 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.453678 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:53 crc kubenswrapper[4809]: E1206 05:53:53.454395 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:53.954379365 +0000 UTC m=+158.843362307 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.479230 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8ssrl" podStartSLOduration=138.479209733 podStartE2EDuration="2m18.479209733s" podCreationTimestamp="2025-12-06 05:51:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:53:53.465595792 +0000 UTC m=+158.354578744" watchObservedRunningTime="2025-12-06 05:53:53.479209733 +0000 UTC m=+158.368192675" Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.512571 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wj2h7"] Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.523891 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wj2h7" Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.528593 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.544422 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wj2h7"] Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.551400 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-znmpz" podStartSLOduration=139.551374828 podStartE2EDuration="2m19.551374828s" podCreationTimestamp="2025-12-06 05:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:53:53.51048845 +0000 UTC m=+158.399471392" watchObservedRunningTime="2025-12-06 05:53:53.551374828 +0000 UTC m=+158.440357770" Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.554700 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:53 crc kubenswrapper[4809]: E1206 05:53:53.555237 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:54.055218724 +0000 UTC m=+158.944201666 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.555439 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-852bd" podStartSLOduration=138.555424509 podStartE2EDuration="2m18.555424509s" podCreationTimestamp="2025-12-06 05:51:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:53:53.540540201 +0000 UTC m=+158.429523153" watchObservedRunningTime="2025-12-06 05:53:53.555424509 +0000 UTC m=+158.444407451" Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.589184 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dvwjq"] Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.624171 4809 patch_prober.go:28] interesting pod/router-default-5444994796-n6z57 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 05:53:53 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Dec 06 05:53:53 crc kubenswrapper[4809]: [+]process-running ok Dec 06 05:53:53 crc kubenswrapper[4809]: healthz check failed Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.624239 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n6z57" podUID="e5356f1d-5afd-47bd-89af-72abfa8f901f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.656628 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.656769 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c4b2f56-0b03-41ae-af23-6192b8bd3d34-catalog-content\") pod \"redhat-marketplace-wj2h7\" (UID: \"2c4b2f56-0b03-41ae-af23-6192b8bd3d34\") " pod="openshift-marketplace/redhat-marketplace-wj2h7" Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.656849 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gglkq\" (UniqueName: \"kubernetes.io/projected/2c4b2f56-0b03-41ae-af23-6192b8bd3d34-kube-api-access-gglkq\") pod \"redhat-marketplace-wj2h7\" (UID: \"2c4b2f56-0b03-41ae-af23-6192b8bd3d34\") " pod="openshift-marketplace/redhat-marketplace-wj2h7" Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.656981 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c4b2f56-0b03-41ae-af23-6192b8bd3d34-utilities\") pod \"redhat-marketplace-wj2h7\" (UID: \"2c4b2f56-0b03-41ae-af23-6192b8bd3d34\") " pod="openshift-marketplace/redhat-marketplace-wj2h7" Dec 06 05:53:53 crc kubenswrapper[4809]: E1206 05:53:53.657452 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:54.15742222 +0000 UTC m=+159.046405162 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.758374 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.758727 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c4b2f56-0b03-41ae-af23-6192b8bd3d34-catalog-content\") pod \"redhat-marketplace-wj2h7\" (UID: \"2c4b2f56-0b03-41ae-af23-6192b8bd3d34\") " pod="openshift-marketplace/redhat-marketplace-wj2h7" Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.758776 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gglkq\" (UniqueName: \"kubernetes.io/projected/2c4b2f56-0b03-41ae-af23-6192b8bd3d34-kube-api-access-gglkq\") pod \"redhat-marketplace-wj2h7\" (UID: \"2c4b2f56-0b03-41ae-af23-6192b8bd3d34\") " pod="openshift-marketplace/redhat-marketplace-wj2h7" Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.758842 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c4b2f56-0b03-41ae-af23-6192b8bd3d34-utilities\") pod \"redhat-marketplace-wj2h7\" (UID: \"2c4b2f56-0b03-41ae-af23-6192b8bd3d34\") " pod="openshift-marketplace/redhat-marketplace-wj2h7" Dec 06 05:53:53 crc kubenswrapper[4809]: E1206 05:53:53.759217 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:54.259190863 +0000 UTC m=+159.148173805 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.759386 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c4b2f56-0b03-41ae-af23-6192b8bd3d34-utilities\") pod \"redhat-marketplace-wj2h7\" (UID: \"2c4b2f56-0b03-41ae-af23-6192b8bd3d34\") " pod="openshift-marketplace/redhat-marketplace-wj2h7" Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.759512 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c4b2f56-0b03-41ae-af23-6192b8bd3d34-catalog-content\") pod \"redhat-marketplace-wj2h7\" (UID: \"2c4b2f56-0b03-41ae-af23-6192b8bd3d34\") " pod="openshift-marketplace/redhat-marketplace-wj2h7" Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.788251 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gglkq\" (UniqueName: \"kubernetes.io/projected/2c4b2f56-0b03-41ae-af23-6192b8bd3d34-kube-api-access-gglkq\") pod \"redhat-marketplace-wj2h7\" (UID: \"2c4b2f56-0b03-41ae-af23-6192b8bd3d34\") " pod="openshift-marketplace/redhat-marketplace-wj2h7" Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.860175 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:53 crc kubenswrapper[4809]: E1206 05:53:53.860596 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:54.360575678 +0000 UTC m=+159.249558670 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.866360 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ls5nx"] Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.866678 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wj2h7" Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.867448 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ls5nx" Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.881781 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ls5nx"] Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.960906 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:53 crc kubenswrapper[4809]: E1206 05:53:53.961057 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:54.461032736 +0000 UTC m=+159.350015678 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.961521 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b475ab56-d949-44e5-a480-f59d8a1d737e-utilities\") pod \"redhat-marketplace-ls5nx\" (UID: \"b475ab56-d949-44e5-a480-f59d8a1d737e\") " pod="openshift-marketplace/redhat-marketplace-ls5nx" Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.961601 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.961649 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffd5h\" (UniqueName: \"kubernetes.io/projected/b475ab56-d949-44e5-a480-f59d8a1d737e-kube-api-access-ffd5h\") pod \"redhat-marketplace-ls5nx\" (UID: \"b475ab56-d949-44e5-a480-f59d8a1d737e\") " pod="openshift-marketplace/redhat-marketplace-ls5nx" Dec 06 05:53:53 crc kubenswrapper[4809]: I1206 05:53:53.961867 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b475ab56-d949-44e5-a480-f59d8a1d737e-catalog-content\") pod \"redhat-marketplace-ls5nx\" (UID: \"b475ab56-d949-44e5-a480-f59d8a1d737e\") " pod="openshift-marketplace/redhat-marketplace-ls5nx" Dec 06 05:53:53 crc kubenswrapper[4809]: E1206 05:53:53.961966 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:54.461950451 +0000 UTC m=+159.350933403 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.039918 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-4pwcw" Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.040001 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-4pwcw" Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.043849 4809 patch_prober.go:28] interesting pod/console-f9d7485db-4pwcw container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.6:8443/health\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.043914 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-4pwcw" podUID="8e03364a-1840-4c91-8b2b-c7e2071fc73b" containerName="console" probeResult="failure" output="Get \"https://10.217.0.6:8443/health\": dial tcp 10.217.0.6:8443: connect: connection refused" Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.063361 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:54 crc kubenswrapper[4809]: E1206 05:53:54.063536 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:54.56351052 +0000 UTC m=+159.452493462 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.063673 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b475ab56-d949-44e5-a480-f59d8a1d737e-catalog-content\") pod \"redhat-marketplace-ls5nx\" (UID: \"b475ab56-d949-44e5-a480-f59d8a1d737e\") " pod="openshift-marketplace/redhat-marketplace-ls5nx" Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.063738 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b475ab56-d949-44e5-a480-f59d8a1d737e-utilities\") pod \"redhat-marketplace-ls5nx\" (UID: \"b475ab56-d949-44e5-a480-f59d8a1d737e\") " pod="openshift-marketplace/redhat-marketplace-ls5nx" Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.063829 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.063881 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffd5h\" (UniqueName: \"kubernetes.io/projected/b475ab56-d949-44e5-a480-f59d8a1d737e-kube-api-access-ffd5h\") pod \"redhat-marketplace-ls5nx\" (UID: \"b475ab56-d949-44e5-a480-f59d8a1d737e\") " pod="openshift-marketplace/redhat-marketplace-ls5nx" Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.064166 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b475ab56-d949-44e5-a480-f59d8a1d737e-catalog-content\") pod \"redhat-marketplace-ls5nx\" (UID: \"b475ab56-d949-44e5-a480-f59d8a1d737e\") " pod="openshift-marketplace/redhat-marketplace-ls5nx" Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.064187 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b475ab56-d949-44e5-a480-f59d8a1d737e-utilities\") pod \"redhat-marketplace-ls5nx\" (UID: \"b475ab56-d949-44e5-a480-f59d8a1d737e\") " pod="openshift-marketplace/redhat-marketplace-ls5nx" Dec 06 05:53:54 crc kubenswrapper[4809]: E1206 05:53:54.064370 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:54.564362223 +0000 UTC m=+159.453345165 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.076064 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wj2h7"] Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.084367 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffd5h\" (UniqueName: \"kubernetes.io/projected/b475ab56-d949-44e5-a480-f59d8a1d737e-kube-api-access-ffd5h\") pod \"redhat-marketplace-ls5nx\" (UID: \"b475ab56-d949-44e5-a480-f59d8a1d737e\") " pod="openshift-marketplace/redhat-marketplace-ls5nx" Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.166164 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:54 crc kubenswrapper[4809]: E1206 05:53:54.166527 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:54.666490577 +0000 UTC m=+159.555473509 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.220961 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ls5nx" Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.223397 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.231074 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nff6m" Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.277311 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:54 crc kubenswrapper[4809]: E1206 05:53:54.277692 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:54.777680689 +0000 UTC m=+159.666663631 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.290784 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dvwjq" event={"ID":"a83cbcc6-62ac-4814-9fce-eb7ea465bcbf","Type":"ContainerStarted","Data":"c93dd1919eb659fb476b1a42a1897b415fbd0e41fcbbf08b63d2bc95644f037f"} Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.292862 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9zm48" event={"ID":"c822a806-352f-4b4f-8549-051ff20c4ca2","Type":"ContainerStarted","Data":"2f5eb5c5202712e416c3fec6986590204363bf55169b18391542dc759ee946f8"} Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.304002 4809 generic.go:334] "Generic (PLEG): container finished" podID="893020b2-fa30-4371-86f9-a585517839ce" containerID="2ad2d97bb0902f7353810c067603e67f7dc4a0c5339e0a9f536eb8729e07d753" exitCode=0 Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.304063 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-24qxt" event={"ID":"893020b2-fa30-4371-86f9-a585517839ce","Type":"ContainerDied","Data":"2ad2d97bb0902f7353810c067603e67f7dc4a0c5339e0a9f536eb8729e07d753"} Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.307599 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zk7tm" event={"ID":"5e67c2b8-8bd1-466c-98d8-e93d411162ba","Type":"ContainerStarted","Data":"c672463a6b1697c333512809b2c102eaf5fc994ca7c2b0bd7c2e9ffd6436fac0"} Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.307643 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zk7tm" event={"ID":"5e67c2b8-8bd1-466c-98d8-e93d411162ba","Type":"ContainerStarted","Data":"1ae5116ddd876d822729ec2c212a93c39ac1adac9d1ae8936821138f3c6fa2a1"} Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.310601 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wj2h7" event={"ID":"2c4b2f56-0b03-41ae-af23-6192b8bd3d34","Type":"ContainerStarted","Data":"e8b5b9437be705b12df2db7d1f3d605c98cd02d84ae03d3cbdd5a044275dbd78"} Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.354039 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-s7px9" podStartSLOduration=12.354016287 podStartE2EDuration="12.354016287s" podCreationTimestamp="2025-12-06 05:53:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:53:54.338273947 +0000 UTC m=+159.227256919" watchObservedRunningTime="2025-12-06 05:53:54.354016287 +0000 UTC m=+159.242999229" Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.387295 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:54 crc kubenswrapper[4809]: E1206 05:53:54.387783 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:54.887765341 +0000 UTC m=+159.776748283 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.483754 4809 patch_prober.go:28] interesting pod/downloads-7954f5f757-224mv container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.37:8080/\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.483810 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-224mv" podUID="8e290142-b2fc-492e-8763-ff41224579fc" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.37:8080/\": dial tcp 10.217.0.37:8080: connect: connection refused" Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.483967 4809 patch_prober.go:28] interesting pod/downloads-7954f5f757-224mv container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.484023 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-224mv" podUID="8e290142-b2fc-492e-8763-ff41224579fc" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.37:8080/\": dial tcp 10.217.0.37:8080: connect: connection refused" Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.489412 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:54 crc kubenswrapper[4809]: E1206 05:53:54.499259 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:54.999241751 +0000 UTC m=+159.888224693 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.555170 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ls5nx"] Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.592527 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:54 crc kubenswrapper[4809]: E1206 05:53:54.592901 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:55.092880553 +0000 UTC m=+159.981863495 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.620771 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-n6z57" Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.629115 4809 patch_prober.go:28] interesting pod/router-default-5444994796-n6z57 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 05:53:54 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Dec 06 05:53:54 crc kubenswrapper[4809]: [+]process-running ok Dec 06 05:53:54 crc kubenswrapper[4809]: healthz check failed Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.629186 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n6z57" podUID="e5356f1d-5afd-47bd-89af-72abfa8f901f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.671343 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lpb28"] Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.672404 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lpb28" Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.679524 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.705669 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:54 crc kubenswrapper[4809]: E1206 05:53:54.705990 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:55.205978647 +0000 UTC m=+160.094961589 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.735962 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lpb28"] Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.757999 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.806980 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:54 crc kubenswrapper[4809]: E1206 05:53:54.807191 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:55.307164555 +0000 UTC m=+160.196147487 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.807347 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40a90e94-16ee-42ba-beff-5103476432c1-catalog-content\") pod \"redhat-operators-lpb28\" (UID: \"40a90e94-16ee-42ba-beff-5103476432c1\") " pod="openshift-marketplace/redhat-operators-lpb28" Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.807393 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.807423 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40a90e94-16ee-42ba-beff-5103476432c1-utilities\") pod \"redhat-operators-lpb28\" (UID: \"40a90e94-16ee-42ba-beff-5103476432c1\") " pod="openshift-marketplace/redhat-operators-lpb28" Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.807447 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dssmc\" (UniqueName: \"kubernetes.io/projected/40a90e94-16ee-42ba-beff-5103476432c1-kube-api-access-dssmc\") pod \"redhat-operators-lpb28\" (UID: \"40a90e94-16ee-42ba-beff-5103476432c1\") " pod="openshift-marketplace/redhat-operators-lpb28" Dec 06 05:53:54 crc kubenswrapper[4809]: E1206 05:53:54.807886 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:55.307870454 +0000 UTC m=+160.196853476 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.908584 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bebb2fb5-644c-49bf-9899-3bc6116cb90e-kubelet-dir\") pod \"bebb2fb5-644c-49bf-9899-3bc6116cb90e\" (UID: \"bebb2fb5-644c-49bf-9899-3bc6116cb90e\") " Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.908704 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bebb2fb5-644c-49bf-9899-3bc6116cb90e-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "bebb2fb5-644c-49bf-9899-3bc6116cb90e" (UID: "bebb2fb5-644c-49bf-9899-3bc6116cb90e"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.909016 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:54 crc kubenswrapper[4809]: E1206 05:53:54.909131 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:55.409113955 +0000 UTC m=+160.298096897 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.909218 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bebb2fb5-644c-49bf-9899-3bc6116cb90e-kube-api-access\") pod \"bebb2fb5-644c-49bf-9899-3bc6116cb90e\" (UID: \"bebb2fb5-644c-49bf-9899-3bc6116cb90e\") " Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.910555 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40a90e94-16ee-42ba-beff-5103476432c1-catalog-content\") pod \"redhat-operators-lpb28\" (UID: \"40a90e94-16ee-42ba-beff-5103476432c1\") " pod="openshift-marketplace/redhat-operators-lpb28" Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.910617 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.910671 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40a90e94-16ee-42ba-beff-5103476432c1-utilities\") pod \"redhat-operators-lpb28\" (UID: \"40a90e94-16ee-42ba-beff-5103476432c1\") " pod="openshift-marketplace/redhat-operators-lpb28" Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.910701 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dssmc\" (UniqueName: \"kubernetes.io/projected/40a90e94-16ee-42ba-beff-5103476432c1-kube-api-access-dssmc\") pod \"redhat-operators-lpb28\" (UID: \"40a90e94-16ee-42ba-beff-5103476432c1\") " pod="openshift-marketplace/redhat-operators-lpb28" Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.910839 4809 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bebb2fb5-644c-49bf-9899-3bc6116cb90e-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 06 05:53:54 crc kubenswrapper[4809]: E1206 05:53:54.911091 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:55.411067648 +0000 UTC m=+160.300050590 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.911573 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40a90e94-16ee-42ba-beff-5103476432c1-utilities\") pod \"redhat-operators-lpb28\" (UID: \"40a90e94-16ee-42ba-beff-5103476432c1\") " pod="openshift-marketplace/redhat-operators-lpb28" Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.926625 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bebb2fb5-644c-49bf-9899-3bc6116cb90e-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "bebb2fb5-644c-49bf-9899-3bc6116cb90e" (UID: "bebb2fb5-644c-49bf-9899-3bc6116cb90e"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:53:54 crc kubenswrapper[4809]: I1206 05:53:54.949581 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dssmc\" (UniqueName: \"kubernetes.io/projected/40a90e94-16ee-42ba-beff-5103476432c1-kube-api-access-dssmc\") pod \"redhat-operators-lpb28\" (UID: \"40a90e94-16ee-42ba-beff-5103476432c1\") " pod="openshift-marketplace/redhat-operators-lpb28" Dec 06 05:53:55 crc kubenswrapper[4809]: I1206 05:53:55.011646 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:55 crc kubenswrapper[4809]: E1206 05:53:55.012029 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:55.511979579 +0000 UTC m=+160.400962581 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:55 crc kubenswrapper[4809]: I1206 05:53:55.012520 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:55 crc kubenswrapper[4809]: I1206 05:53:55.012656 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bebb2fb5-644c-49bf-9899-3bc6116cb90e-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 05:53:55 crc kubenswrapper[4809]: E1206 05:53:55.012892 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:55.512875764 +0000 UTC m=+160.401858786 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:55 crc kubenswrapper[4809]: I1206 05:53:55.074337 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hfgk4"] Dec 06 05:53:55 crc kubenswrapper[4809]: E1206 05:53:55.074701 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bebb2fb5-644c-49bf-9899-3bc6116cb90e" containerName="pruner" Dec 06 05:53:55 crc kubenswrapper[4809]: I1206 05:53:55.074736 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="bebb2fb5-644c-49bf-9899-3bc6116cb90e" containerName="pruner" Dec 06 05:53:55 crc kubenswrapper[4809]: I1206 05:53:55.074909 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="bebb2fb5-644c-49bf-9899-3bc6116cb90e" containerName="pruner" Dec 06 05:53:55 crc kubenswrapper[4809]: I1206 05:53:55.076186 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hfgk4" Dec 06 05:53:55 crc kubenswrapper[4809]: I1206 05:53:55.082327 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hfgk4"] Dec 06 05:53:55 crc kubenswrapper[4809]: I1206 05:53:55.113545 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:55 crc kubenswrapper[4809]: E1206 05:53:55.113867 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:55.613853076 +0000 UTC m=+160.502836018 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:55 crc kubenswrapper[4809]: I1206 05:53:55.214773 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e356b822-4e99-42f8-99ba-7aea197f3d95-utilities\") pod \"redhat-operators-hfgk4\" (UID: \"e356b822-4e99-42f8-99ba-7aea197f3d95\") " pod="openshift-marketplace/redhat-operators-hfgk4" Dec 06 05:53:55 crc kubenswrapper[4809]: I1206 05:53:55.214858 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e356b822-4e99-42f8-99ba-7aea197f3d95-catalog-content\") pod \"redhat-operators-hfgk4\" (UID: \"e356b822-4e99-42f8-99ba-7aea197f3d95\") " pod="openshift-marketplace/redhat-operators-hfgk4" Dec 06 05:53:55 crc kubenswrapper[4809]: I1206 05:53:55.214897 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:55 crc kubenswrapper[4809]: I1206 05:53:55.214999 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5kmb\" (UniqueName: \"kubernetes.io/projected/e356b822-4e99-42f8-99ba-7aea197f3d95-kube-api-access-z5kmb\") pod \"redhat-operators-hfgk4\" (UID: \"e356b822-4e99-42f8-99ba-7aea197f3d95\") " pod="openshift-marketplace/redhat-operators-hfgk4" Dec 06 05:53:55 crc kubenswrapper[4809]: E1206 05:53:55.215302 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:55.715288341 +0000 UTC m=+160.604271363 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:55 crc kubenswrapper[4809]: I1206 05:53:55.245146 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cd2tv" Dec 06 05:53:55 crc kubenswrapper[4809]: I1206 05:53:55.245202 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cd2tv" Dec 06 05:53:55 crc kubenswrapper[4809]: I1206 05:53:55.303119 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-slt9h" Dec 06 05:53:55 crc kubenswrapper[4809]: I1206 05:53:55.315307 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"bebb2fb5-644c-49bf-9899-3bc6116cb90e","Type":"ContainerDied","Data":"299ee7eedf129c41cdefc4a0d9ba23e43f38f2983a4ec1f5fbe1f8bf35c04e70"} Dec 06 05:53:55 crc kubenswrapper[4809]: I1206 05:53:55.315340 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="299ee7eedf129c41cdefc4a0d9ba23e43f38f2983a4ec1f5fbe1f8bf35c04e70" Dec 06 05:53:55 crc kubenswrapper[4809]: I1206 05:53:55.315384 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 05:53:55 crc kubenswrapper[4809]: I1206 05:53:55.319509 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:55 crc kubenswrapper[4809]: E1206 05:53:55.319616 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:55.819596075 +0000 UTC m=+160.708579017 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:55 crc kubenswrapper[4809]: I1206 05:53:55.319916 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e356b822-4e99-42f8-99ba-7aea197f3d95-utilities\") pod \"redhat-operators-hfgk4\" (UID: \"e356b822-4e99-42f8-99ba-7aea197f3d95\") " pod="openshift-marketplace/redhat-operators-hfgk4" Dec 06 05:53:55 crc kubenswrapper[4809]: I1206 05:53:55.320046 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e356b822-4e99-42f8-99ba-7aea197f3d95-catalog-content\") pod \"redhat-operators-hfgk4\" (UID: \"e356b822-4e99-42f8-99ba-7aea197f3d95\") " pod="openshift-marketplace/redhat-operators-hfgk4" Dec 06 05:53:55 crc kubenswrapper[4809]: I1206 05:53:55.320081 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:55 crc kubenswrapper[4809]: I1206 05:53:55.320165 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5kmb\" (UniqueName: \"kubernetes.io/projected/e356b822-4e99-42f8-99ba-7aea197f3d95-kube-api-access-z5kmb\") pod \"redhat-operators-hfgk4\" (UID: \"e356b822-4e99-42f8-99ba-7aea197f3d95\") " pod="openshift-marketplace/redhat-operators-hfgk4" Dec 06 05:53:55 crc kubenswrapper[4809]: E1206 05:53:55.320520 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:55.820504309 +0000 UTC m=+160.709487271 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:55 crc kubenswrapper[4809]: I1206 05:53:55.320515 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e356b822-4e99-42f8-99ba-7aea197f3d95-utilities\") pod \"redhat-operators-hfgk4\" (UID: \"e356b822-4e99-42f8-99ba-7aea197f3d95\") " pod="openshift-marketplace/redhat-operators-hfgk4" Dec 06 05:53:55 crc kubenswrapper[4809]: I1206 05:53:55.320575 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e356b822-4e99-42f8-99ba-7aea197f3d95-catalog-content\") pod \"redhat-operators-hfgk4\" (UID: \"e356b822-4e99-42f8-99ba-7aea197f3d95\") " pod="openshift-marketplace/redhat-operators-hfgk4" Dec 06 05:53:55 crc kubenswrapper[4809]: I1206 05:53:55.321087 4809 generic.go:334] "Generic (PLEG): container finished" podID="5e67c2b8-8bd1-466c-98d8-e93d411162ba" containerID="c672463a6b1697c333512809b2c102eaf5fc994ca7c2b0bd7c2e9ffd6436fac0" exitCode=0 Dec 06 05:53:55 crc kubenswrapper[4809]: I1206 05:53:55.321145 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zk7tm" event={"ID":"5e67c2b8-8bd1-466c-98d8-e93d411162ba","Type":"ContainerDied","Data":"c672463a6b1697c333512809b2c102eaf5fc994ca7c2b0bd7c2e9ffd6436fac0"} Dec 06 05:53:55 crc kubenswrapper[4809]: I1206 05:53:55.322337 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ls5nx" event={"ID":"b475ab56-d949-44e5-a480-f59d8a1d737e","Type":"ContainerStarted","Data":"f413b3d3dd61ba7fc1ac276f14b507e33ab610d693fceb2d105f6d66bcfbb406"} Dec 06 05:53:55 crc kubenswrapper[4809]: I1206 05:53:55.322904 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-s7px9" Dec 06 05:53:55 crc kubenswrapper[4809]: I1206 05:53:55.357717 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5kmb\" (UniqueName: \"kubernetes.io/projected/e356b822-4e99-42f8-99ba-7aea197f3d95-kube-api-access-z5kmb\") pod \"redhat-operators-hfgk4\" (UID: \"e356b822-4e99-42f8-99ba-7aea197f3d95\") " pod="openshift-marketplace/redhat-operators-hfgk4" Dec 06 05:53:55 crc kubenswrapper[4809]: I1206 05:53:55.394279 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hfgk4" Dec 06 05:53:55 crc kubenswrapper[4809]: I1206 05:53:55.421484 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:55 crc kubenswrapper[4809]: E1206 05:53:55.422828 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:55.922800679 +0000 UTC m=+160.811783621 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:55 crc kubenswrapper[4809]: I1206 05:53:55.524821 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:55 crc kubenswrapper[4809]: E1206 05:53:55.525550 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:56.025534069 +0000 UTC m=+160.914517011 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:55 crc kubenswrapper[4809]: I1206 05:53:55.599339 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hfgk4"] Dec 06 05:53:55 crc kubenswrapper[4809]: W1206 05:53:55.604170 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode356b822_4e99_42f8_99ba_7aea197f3d95.slice/crio-3ad5e7413596dd464ef04ea4692a871bf193a5d69854708021f197793efb26c7 WatchSource:0}: Error finding container 3ad5e7413596dd464ef04ea4692a871bf193a5d69854708021f197793efb26c7: Status 404 returned error can't find the container with id 3ad5e7413596dd464ef04ea4692a871bf193a5d69854708021f197793efb26c7 Dec 06 05:53:55 crc kubenswrapper[4809]: I1206 05:53:55.620147 4809 patch_prober.go:28] interesting pod/router-default-5444994796-n6z57 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 05:53:55 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Dec 06 05:53:55 crc kubenswrapper[4809]: [+]process-running ok Dec 06 05:53:55 crc kubenswrapper[4809]: healthz check failed Dec 06 05:53:55 crc kubenswrapper[4809]: I1206 05:53:55.620201 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n6z57" podUID="e5356f1d-5afd-47bd-89af-72abfa8f901f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 05:53:55 crc kubenswrapper[4809]: I1206 05:53:55.628587 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:55 crc kubenswrapper[4809]: E1206 05:53:55.628988 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:56.128969229 +0000 UTC m=+161.017952171 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:55 crc kubenswrapper[4809]: I1206 05:53:55.629024 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:55 crc kubenswrapper[4809]: E1206 05:53:55.629299 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:56.129277788 +0000 UTC m=+161.018260730 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:55 crc kubenswrapper[4809]: I1206 05:53:55.730577 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:55 crc kubenswrapper[4809]: E1206 05:53:55.730722 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:56.230700072 +0000 UTC m=+161.119683034 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:55 crc kubenswrapper[4809]: I1206 05:53:55.731209 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:55 crc kubenswrapper[4809]: E1206 05:53:55.731561 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:56.231550136 +0000 UTC m=+161.120533078 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:55 crc kubenswrapper[4809]: I1206 05:53:55.831897 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:55 crc kubenswrapper[4809]: E1206 05:53:55.832041 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:56.332018435 +0000 UTC m=+161.221001387 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:55 crc kubenswrapper[4809]: I1206 05:53:55.832267 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:55 crc kubenswrapper[4809]: E1206 05:53:55.832549 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:56.332538818 +0000 UTC m=+161.221521760 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:55 crc kubenswrapper[4809]: I1206 05:53:55.923179 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40a90e94-16ee-42ba-beff-5103476432c1-catalog-content\") pod \"redhat-operators-lpb28\" (UID: \"40a90e94-16ee-42ba-beff-5103476432c1\") " pod="openshift-marketplace/redhat-operators-lpb28" Dec 06 05:53:55 crc kubenswrapper[4809]: I1206 05:53:55.933532 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:55 crc kubenswrapper[4809]: E1206 05:53:55.933746 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:56.433714746 +0000 UTC m=+161.322697698 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:55 crc kubenswrapper[4809]: I1206 05:53:55.933904 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:55 crc kubenswrapper[4809]: E1206 05:53:55.934268 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:56.434251932 +0000 UTC m=+161.323234964 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:55 crc kubenswrapper[4809]: I1206 05:53:55.956226 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lpb28" Dec 06 05:53:56 crc kubenswrapper[4809]: I1206 05:53:56.035615 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:56 crc kubenswrapper[4809]: E1206 05:53:56.035797 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:56.535774669 +0000 UTC m=+161.424757611 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:56 crc kubenswrapper[4809]: I1206 05:53:56.035913 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:56 crc kubenswrapper[4809]: E1206 05:53:56.036199 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:56.536191421 +0000 UTC m=+161.425174363 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:56 crc kubenswrapper[4809]: I1206 05:53:56.136925 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:56 crc kubenswrapper[4809]: E1206 05:53:56.137294 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:56.637278656 +0000 UTC m=+161.526261598 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:56 crc kubenswrapper[4809]: I1206 05:53:56.164649 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lpb28"] Dec 06 05:53:56 crc kubenswrapper[4809]: I1206 05:53:56.238390 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:56 crc kubenswrapper[4809]: E1206 05:53:56.238744 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:56.738728461 +0000 UTC m=+161.627711403 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:56 crc kubenswrapper[4809]: I1206 05:53:56.322721 4809 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-qqv6x container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.30:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 05:53:56 crc kubenswrapper[4809]: I1206 05:53:56.323071 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qqv6x" podUID="8a225b5c-4ad5-4f08-a0cb-ae38a8fa83d4" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.30:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 05:53:56 crc kubenswrapper[4809]: I1206 05:53:56.328263 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hfgk4" event={"ID":"e356b822-4e99-42f8-99ba-7aea197f3d95","Type":"ContainerStarted","Data":"3ad5e7413596dd464ef04ea4692a871bf193a5d69854708021f197793efb26c7"} Dec 06 05:53:56 crc kubenswrapper[4809]: I1206 05:53:56.329654 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lpb28" event={"ID":"40a90e94-16ee-42ba-beff-5103476432c1","Type":"ContainerStarted","Data":"3417db87bbbc2d4bd8805f87c74161a886bb760bf9cf47491c3c7ffa124afa32"} Dec 06 05:53:56 crc kubenswrapper[4809]: I1206 05:53:56.331395 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9zm48" event={"ID":"c822a806-352f-4b4f-8549-051ff20c4ca2","Type":"ContainerStarted","Data":"3ba2cbb88fc0c5d32e76c8b1b832077c55033f367ed4a3e955e7122340be0961"} Dec 06 05:53:56 crc kubenswrapper[4809]: I1206 05:53:56.340106 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:56 crc kubenswrapper[4809]: E1206 05:53:56.340396 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:56.840382833 +0000 UTC m=+161.729365765 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:56 crc kubenswrapper[4809]: I1206 05:53:56.441885 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:56 crc kubenswrapper[4809]: E1206 05:53:56.442453 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:56.942433045 +0000 UTC m=+161.831416017 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:56 crc kubenswrapper[4809]: I1206 05:53:56.543173 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:56 crc kubenswrapper[4809]: E1206 05:53:56.543354 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:57.043328285 +0000 UTC m=+161.932311217 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:56 crc kubenswrapper[4809]: I1206 05:53:56.543762 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:56 crc kubenswrapper[4809]: E1206 05:53:56.544077 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:57.044069745 +0000 UTC m=+161.933052687 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:56 crc kubenswrapper[4809]: I1206 05:53:56.620548 4809 patch_prober.go:28] interesting pod/router-default-5444994796-n6z57 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 05:53:56 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Dec 06 05:53:56 crc kubenswrapper[4809]: [+]process-running ok Dec 06 05:53:56 crc kubenswrapper[4809]: healthz check failed Dec 06 05:53:56 crc kubenswrapper[4809]: I1206 05:53:56.620621 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n6z57" podUID="e5356f1d-5afd-47bd-89af-72abfa8f901f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 05:53:56 crc kubenswrapper[4809]: I1206 05:53:56.644355 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:56 crc kubenswrapper[4809]: E1206 05:53:56.644520 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:57.144498843 +0000 UTC m=+162.033481785 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:56 crc kubenswrapper[4809]: I1206 05:53:56.644621 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:56 crc kubenswrapper[4809]: E1206 05:53:56.644963 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:57.144953265 +0000 UTC m=+162.033936217 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:56 crc kubenswrapper[4809]: I1206 05:53:56.746260 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:56 crc kubenswrapper[4809]: E1206 05:53:56.746437 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:57.246415792 +0000 UTC m=+162.135398744 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:56 crc kubenswrapper[4809]: I1206 05:53:56.746819 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:56 crc kubenswrapper[4809]: E1206 05:53:56.747323 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:57.247299146 +0000 UTC m=+162.136282118 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:56 crc kubenswrapper[4809]: I1206 05:53:56.847499 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:56 crc kubenswrapper[4809]: E1206 05:53:56.847733 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:57.347691332 +0000 UTC m=+162.236674314 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:56 crc kubenswrapper[4809]: I1206 05:53:56.848135 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:56 crc kubenswrapper[4809]: E1206 05:53:56.848518 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:57.348499405 +0000 UTC m=+162.237482377 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:56 crc kubenswrapper[4809]: I1206 05:53:56.949466 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:56 crc kubenswrapper[4809]: E1206 05:53:56.949759 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:57.449743744 +0000 UTC m=+162.338726686 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:57 crc kubenswrapper[4809]: I1206 05:53:57.051113 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:57 crc kubenswrapper[4809]: I1206 05:53:57.051194 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a4f75fd3-03e8-41e7-8926-f6ff21a5b681-metrics-certs\") pod \"network-metrics-daemon-9k8zx\" (UID: \"a4f75fd3-03e8-41e7-8926-f6ff21a5b681\") " pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:53:57 crc kubenswrapper[4809]: E1206 05:53:57.052444 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:57.552425513 +0000 UTC m=+162.441408475 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:57 crc kubenswrapper[4809]: I1206 05:53:57.062213 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a4f75fd3-03e8-41e7-8926-f6ff21a5b681-metrics-certs\") pod \"network-metrics-daemon-9k8zx\" (UID: \"a4f75fd3-03e8-41e7-8926-f6ff21a5b681\") " pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:53:57 crc kubenswrapper[4809]: I1206 05:53:57.151910 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:57 crc kubenswrapper[4809]: E1206 05:53:57.152079 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:57.65205927 +0000 UTC m=+162.541042212 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:57 crc kubenswrapper[4809]: I1206 05:53:57.152477 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:57 crc kubenswrapper[4809]: E1206 05:53:57.152849 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:57.652838341 +0000 UTC m=+162.541821283 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:57 crc kubenswrapper[4809]: I1206 05:53:57.253370 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:57 crc kubenswrapper[4809]: E1206 05:53:57.253792 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:57.753772112 +0000 UTC m=+162.642755054 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:57 crc kubenswrapper[4809]: I1206 05:53:57.295040 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cd2tv" Dec 06 05:53:57 crc kubenswrapper[4809]: I1206 05:53:57.306482 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cd2tv" Dec 06 05:53:57 crc kubenswrapper[4809]: I1206 05:53:57.308451 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qqv6x" Dec 06 05:53:57 crc kubenswrapper[4809]: I1206 05:53:57.314169 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9k8zx" Dec 06 05:53:57 crc kubenswrapper[4809]: I1206 05:53:57.354722 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:57 crc kubenswrapper[4809]: E1206 05:53:57.356193 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:57.856172723 +0000 UTC m=+162.745155735 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:57 crc kubenswrapper[4809]: I1206 05:53:57.358106 4809 generic.go:334] "Generic (PLEG): container finished" podID="c822a806-352f-4b4f-8549-051ff20c4ca2" containerID="3ba2cbb88fc0c5d32e76c8b1b832077c55033f367ed4a3e955e7122340be0961" exitCode=0 Dec 06 05:53:57 crc kubenswrapper[4809]: I1206 05:53:57.359100 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9zm48" event={"ID":"c822a806-352f-4b4f-8549-051ff20c4ca2","Type":"ContainerDied","Data":"3ba2cbb88fc0c5d32e76c8b1b832077c55033f367ed4a3e955e7122340be0961"} Dec 06 05:53:57 crc kubenswrapper[4809]: I1206 05:53:57.459033 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:57 crc kubenswrapper[4809]: E1206 05:53:57.459366 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:57.959340206 +0000 UTC m=+162.848323138 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:57 crc kubenswrapper[4809]: I1206 05:53:57.561219 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:57 crc kubenswrapper[4809]: E1206 05:53:57.561629 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:58.061613435 +0000 UTC m=+162.950596377 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:57 crc kubenswrapper[4809]: I1206 05:53:57.622981 4809 patch_prober.go:28] interesting pod/router-default-5444994796-n6z57 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 05:53:57 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Dec 06 05:53:57 crc kubenswrapper[4809]: [+]process-running ok Dec 06 05:53:57 crc kubenswrapper[4809]: healthz check failed Dec 06 05:53:57 crc kubenswrapper[4809]: I1206 05:53:57.623056 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n6z57" podUID="e5356f1d-5afd-47bd-89af-72abfa8f901f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 05:53:57 crc kubenswrapper[4809]: I1206 05:53:57.657435 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-9k8zx"] Dec 06 05:53:57 crc kubenswrapper[4809]: I1206 05:53:57.662455 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:57 crc kubenswrapper[4809]: E1206 05:53:57.662632 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:58.162610568 +0000 UTC m=+163.051593510 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:57 crc kubenswrapper[4809]: I1206 05:53:57.662760 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:57 crc kubenswrapper[4809]: E1206 05:53:57.663231 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:58.163212695 +0000 UTC m=+163.052195697 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:57 crc kubenswrapper[4809]: I1206 05:53:57.665197 4809 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 06 05:53:57 crc kubenswrapper[4809]: I1206 05:53:57.763586 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:57 crc kubenswrapper[4809]: E1206 05:53:57.764104 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:58.264078934 +0000 UTC m=+163.153061876 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:57 crc kubenswrapper[4809]: I1206 05:53:57.764318 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:57 crc kubenswrapper[4809]: E1206 05:53:57.764658 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:58.264643229 +0000 UTC m=+163.153626171 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:57 crc kubenswrapper[4809]: W1206 05:53:57.819558 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda4f75fd3_03e8_41e7_8926_f6ff21a5b681.slice/crio-dd9c9981731d86558aa612125b8a4abcc720058b21734a6007d998d4273e9f5e WatchSource:0}: Error finding container dd9c9981731d86558aa612125b8a4abcc720058b21734a6007d998d4273e9f5e: Status 404 returned error can't find the container with id dd9c9981731d86558aa612125b8a4abcc720058b21734a6007d998d4273e9f5e Dec 06 05:53:57 crc kubenswrapper[4809]: I1206 05:53:57.865735 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:57 crc kubenswrapper[4809]: E1206 05:53:57.866325 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:53:58.366301811 +0000 UTC m=+163.255284763 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:57 crc kubenswrapper[4809]: I1206 05:53:57.965209 4809 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-06T05:53:57.665217409Z","Handler":null,"Name":""} Dec 06 05:53:57 crc kubenswrapper[4809]: I1206 05:53:57.966891 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:57 crc kubenswrapper[4809]: E1206 05:53:57.967192 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:53:58.467180881 +0000 UTC m=+163.356163823 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kwzrc" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:53:57 crc kubenswrapper[4809]: I1206 05:53:57.967833 4809 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 06 05:53:57 crc kubenswrapper[4809]: I1206 05:53:57.967853 4809 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 06 05:53:58 crc kubenswrapper[4809]: I1206 05:53:58.002188 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 06 05:53:58 crc kubenswrapper[4809]: I1206 05:53:58.003027 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 05:53:58 crc kubenswrapper[4809]: I1206 05:53:58.005165 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 06 05:53:58 crc kubenswrapper[4809]: I1206 05:53:58.006363 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 06 05:53:58 crc kubenswrapper[4809]: I1206 05:53:58.045433 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 06 05:53:58 crc kubenswrapper[4809]: I1206 05:53:58.067697 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:53:58 crc kubenswrapper[4809]: I1206 05:53:58.073747 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 06 05:53:58 crc kubenswrapper[4809]: I1206 05:53:58.170363 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3b196d34-f754-42a0-a82b-8a139e316aa2-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"3b196d34-f754-42a0-a82b-8a139e316aa2\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 05:53:58 crc kubenswrapper[4809]: I1206 05:53:58.170472 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:58 crc kubenswrapper[4809]: I1206 05:53:58.170525 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3b196d34-f754-42a0-a82b-8a139e316aa2-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"3b196d34-f754-42a0-a82b-8a139e316aa2\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 05:53:58 crc kubenswrapper[4809]: I1206 05:53:58.271651 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3b196d34-f754-42a0-a82b-8a139e316aa2-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"3b196d34-f754-42a0-a82b-8a139e316aa2\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 05:53:58 crc kubenswrapper[4809]: I1206 05:53:58.271697 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3b196d34-f754-42a0-a82b-8a139e316aa2-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"3b196d34-f754-42a0-a82b-8a139e316aa2\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 05:53:58 crc kubenswrapper[4809]: I1206 05:53:58.271975 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3b196d34-f754-42a0-a82b-8a139e316aa2-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"3b196d34-f754-42a0-a82b-8a139e316aa2\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 05:53:58 crc kubenswrapper[4809]: I1206 05:53:58.295589 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3b196d34-f754-42a0-a82b-8a139e316aa2-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"3b196d34-f754-42a0-a82b-8a139e316aa2\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 05:53:58 crc kubenswrapper[4809]: I1206 05:53:58.346539 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 05:53:58 crc kubenswrapper[4809]: I1206 05:53:58.368065 4809 generic.go:334] "Generic (PLEG): container finished" podID="c892e996-27b7-4b8d-bd92-7eac2df7702b" containerID="7ba2c355548a958343d1e97f6c244b4c1cb70b0083c1fb59afb79e37ce2ee82c" exitCode=0 Dec 06 05:53:58 crc kubenswrapper[4809]: I1206 05:53:58.368152 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416665-lsfkp" event={"ID":"c892e996-27b7-4b8d-bd92-7eac2df7702b","Type":"ContainerDied","Data":"7ba2c355548a958343d1e97f6c244b4c1cb70b0083c1fb59afb79e37ce2ee82c"} Dec 06 05:53:58 crc kubenswrapper[4809]: I1206 05:53:58.374262 4809 generic.go:334] "Generic (PLEG): container finished" podID="e356b822-4e99-42f8-99ba-7aea197f3d95" containerID="59b3a77acd59b69589a0f06f5fa141f0a8538dc3c3422d7932a033031e3e5d80" exitCode=0 Dec 06 05:53:58 crc kubenswrapper[4809]: I1206 05:53:58.374360 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hfgk4" event={"ID":"e356b822-4e99-42f8-99ba-7aea197f3d95","Type":"ContainerDied","Data":"59b3a77acd59b69589a0f06f5fa141f0a8538dc3c3422d7932a033031e3e5d80"} Dec 06 05:53:58 crc kubenswrapper[4809]: I1206 05:53:58.381880 4809 generic.go:334] "Generic (PLEG): container finished" podID="40a90e94-16ee-42ba-beff-5103476432c1" containerID="fb25f0d646c14b2a4d5d1288817d466f004acaaa4d3fc0aa90dae543dd397b6b" exitCode=0 Dec 06 05:53:58 crc kubenswrapper[4809]: I1206 05:53:58.381925 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lpb28" event={"ID":"40a90e94-16ee-42ba-beff-5103476432c1","Type":"ContainerDied","Data":"fb25f0d646c14b2a4d5d1288817d466f004acaaa4d3fc0aa90dae543dd397b6b"} Dec 06 05:53:58 crc kubenswrapper[4809]: I1206 05:53:58.389776 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-9k8zx" event={"ID":"a4f75fd3-03e8-41e7-8926-f6ff21a5b681","Type":"ContainerStarted","Data":"dd9c9981731d86558aa612125b8a4abcc720058b21734a6007d998d4273e9f5e"} Dec 06 05:53:58 crc kubenswrapper[4809]: I1206 05:53:58.390985 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wj2h7" event={"ID":"2c4b2f56-0b03-41ae-af23-6192b8bd3d34","Type":"ContainerStarted","Data":"53c987bae54781114f4630739e892e70b687516ae3183ac01861392512f94375"} Dec 06 05:53:58 crc kubenswrapper[4809]: I1206 05:53:58.392317 4809 generic.go:334] "Generic (PLEG): container finished" podID="a83cbcc6-62ac-4814-9fce-eb7ea465bcbf" containerID="6c8f901f4b4a3252b4273b5561ce69cc87f1ebacbc0de5a67bd1e9a705eff52f" exitCode=0 Dec 06 05:53:58 crc kubenswrapper[4809]: I1206 05:53:58.392398 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dvwjq" event={"ID":"a83cbcc6-62ac-4814-9fce-eb7ea465bcbf","Type":"ContainerDied","Data":"6c8f901f4b4a3252b4273b5561ce69cc87f1ebacbc0de5a67bd1e9a705eff52f"} Dec 06 05:53:58 crc kubenswrapper[4809]: I1206 05:53:58.394417 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-tphf5" event={"ID":"09a7af3d-6bf2-4b7a-93d9-cecc73d993a1","Type":"ContainerStarted","Data":"77fd9eb631c7c48a1ddd4683ea48b7a2597ee08a0594fe858673dd0a18eec617"} Dec 06 05:53:58 crc kubenswrapper[4809]: I1206 05:53:58.396541 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-p2lkt" event={"ID":"c598f580-a438-4015-9a8a-f98b3b2b0b3d","Type":"ContainerStarted","Data":"6f17e7584bc2a683f6b053e3ded398fe8997cfaaa4bb53b28b0e9b6f0dfe352e"} Dec 06 05:53:58 crc kubenswrapper[4809]: I1206 05:53:58.557580 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 06 05:53:58 crc kubenswrapper[4809]: I1206 05:53:58.621120 4809 patch_prober.go:28] interesting pod/router-default-5444994796-n6z57 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 05:53:58 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Dec 06 05:53:58 crc kubenswrapper[4809]: [+]process-running ok Dec 06 05:53:58 crc kubenswrapper[4809]: healthz check failed Dec 06 05:53:58 crc kubenswrapper[4809]: I1206 05:53:58.621208 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n6z57" podUID="e5356f1d-5afd-47bd-89af-72abfa8f901f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 05:53:58 crc kubenswrapper[4809]: I1206 05:53:58.697759 4809 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 06 05:53:58 crc kubenswrapper[4809]: I1206 05:53:58.697822 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:58 crc kubenswrapper[4809]: I1206 05:53:58.725108 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kwzrc\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:58 crc kubenswrapper[4809]: I1206 05:53:58.780590 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:58 crc kubenswrapper[4809]: I1206 05:53:58.971923 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kwzrc"] Dec 06 05:53:58 crc kubenswrapper[4809]: W1206 05:53:58.979601 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc8703a5f_60fb_40f5_be0c_d1025852c05a.slice/crio-e50dd982694f146cca4fadd24a64db7898a56020f167f6645c5fcaa8d16c37b4 WatchSource:0}: Error finding container e50dd982694f146cca4fadd24a64db7898a56020f167f6645c5fcaa8d16c37b4: Status 404 returned error can't find the container with id e50dd982694f146cca4fadd24a64db7898a56020f167f6645c5fcaa8d16c37b4 Dec 06 05:53:59 crc kubenswrapper[4809]: I1206 05:53:59.403571 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 06 05:53:59 crc kubenswrapper[4809]: I1206 05:53:59.424627 4809 generic.go:334] "Generic (PLEG): container finished" podID="b475ab56-d949-44e5-a480-f59d8a1d737e" containerID="3162d1b7bddb463502601ccee5bfef8b3ccc2d1403ec4a883b60fe68fa80bbb1" exitCode=0 Dec 06 05:53:59 crc kubenswrapper[4809]: I1206 05:53:59.424703 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ls5nx" event={"ID":"b475ab56-d949-44e5-a480-f59d8a1d737e","Type":"ContainerDied","Data":"3162d1b7bddb463502601ccee5bfef8b3ccc2d1403ec4a883b60fe68fa80bbb1"} Dec 06 05:53:59 crc kubenswrapper[4809]: I1206 05:53:59.453371 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" event={"ID":"c8703a5f-60fb-40f5-be0c-d1025852c05a","Type":"ContainerStarted","Data":"7b7947c95ae0c2b087e5964c2a4f311424f6e49c17d8601b507e5c18e01992c9"} Dec 06 05:53:59 crc kubenswrapper[4809]: I1206 05:53:59.453439 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" event={"ID":"c8703a5f-60fb-40f5-be0c-d1025852c05a","Type":"ContainerStarted","Data":"e50dd982694f146cca4fadd24a64db7898a56020f167f6645c5fcaa8d16c37b4"} Dec 06 05:53:59 crc kubenswrapper[4809]: I1206 05:53:59.453626 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:53:59 crc kubenswrapper[4809]: I1206 05:53:59.465451 4809 generic.go:334] "Generic (PLEG): container finished" podID="2c4b2f56-0b03-41ae-af23-6192b8bd3d34" containerID="53c987bae54781114f4630739e892e70b687516ae3183ac01861392512f94375" exitCode=0 Dec 06 05:53:59 crc kubenswrapper[4809]: I1206 05:53:59.465513 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wj2h7" event={"ID":"2c4b2f56-0b03-41ae-af23-6192b8bd3d34","Type":"ContainerDied","Data":"53c987bae54781114f4630739e892e70b687516ae3183ac01861392512f94375"} Dec 06 05:53:59 crc kubenswrapper[4809]: I1206 05:53:59.473821 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"3b196d34-f754-42a0-a82b-8a139e316aa2","Type":"ContainerStarted","Data":"6aca914b627feff21f991d038f746be15a6f1d49b21eae88c62f2f006454f22f"} Dec 06 05:53:59 crc kubenswrapper[4809]: I1206 05:53:59.473850 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"3b196d34-f754-42a0-a82b-8a139e316aa2","Type":"ContainerStarted","Data":"1d1e9d23bf9c8cb529dba845b6a6839e16e74c197b7135a38713283ac826a26b"} Dec 06 05:53:59 crc kubenswrapper[4809]: I1206 05:53:59.475785 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" podStartSLOduration=145.475775084 podStartE2EDuration="2m25.475775084s" podCreationTimestamp="2025-12-06 05:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:53:59.475266291 +0000 UTC m=+164.364249243" watchObservedRunningTime="2025-12-06 05:53:59.475775084 +0000 UTC m=+164.364758026" Dec 06 05:53:59 crc kubenswrapper[4809]: I1206 05:53:59.500309 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-9k8zx" event={"ID":"a4f75fd3-03e8-41e7-8926-f6ff21a5b681","Type":"ContainerStarted","Data":"63d80de5c8d7289b28e3d985ab94b413b95701c56891f7b228f8288bd5308612"} Dec 06 05:53:59 crc kubenswrapper[4809]: I1206 05:53:59.510477 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.510461104 podStartE2EDuration="2.510461104s" podCreationTimestamp="2025-12-06 05:53:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:53:59.508392398 +0000 UTC m=+164.397375340" watchObservedRunningTime="2025-12-06 05:53:59.510461104 +0000 UTC m=+164.399444046" Dec 06 05:53:59 crc kubenswrapper[4809]: I1206 05:53:59.572450 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-tphf5" podStartSLOduration=145.572427159 podStartE2EDuration="2m25.572427159s" podCreationTimestamp="2025-12-06 05:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:53:59.569636542 +0000 UTC m=+164.458619494" watchObservedRunningTime="2025-12-06 05:53:59.572427159 +0000 UTC m=+164.461410101" Dec 06 05:53:59 crc kubenswrapper[4809]: I1206 05:53:59.633274 4809 patch_prober.go:28] interesting pod/router-default-5444994796-n6z57 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 05:53:59 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Dec 06 05:53:59 crc kubenswrapper[4809]: [+]process-running ok Dec 06 05:53:59 crc kubenswrapper[4809]: healthz check failed Dec 06 05:53:59 crc kubenswrapper[4809]: I1206 05:53:59.633562 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n6z57" podUID="e5356f1d-5afd-47bd-89af-72abfa8f901f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 05:54:00 crc kubenswrapper[4809]: I1206 05:54:00.063344 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416665-lsfkp" Dec 06 05:54:00 crc kubenswrapper[4809]: I1206 05:54:00.169740 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-s7px9" Dec 06 05:54:00 crc kubenswrapper[4809]: I1206 05:54:00.216295 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxk8g\" (UniqueName: \"kubernetes.io/projected/c892e996-27b7-4b8d-bd92-7eac2df7702b-kube-api-access-vxk8g\") pod \"c892e996-27b7-4b8d-bd92-7eac2df7702b\" (UID: \"c892e996-27b7-4b8d-bd92-7eac2df7702b\") " Dec 06 05:54:00 crc kubenswrapper[4809]: I1206 05:54:00.216404 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c892e996-27b7-4b8d-bd92-7eac2df7702b-secret-volume\") pod \"c892e996-27b7-4b8d-bd92-7eac2df7702b\" (UID: \"c892e996-27b7-4b8d-bd92-7eac2df7702b\") " Dec 06 05:54:00 crc kubenswrapper[4809]: I1206 05:54:00.216478 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c892e996-27b7-4b8d-bd92-7eac2df7702b-config-volume\") pod \"c892e996-27b7-4b8d-bd92-7eac2df7702b\" (UID: \"c892e996-27b7-4b8d-bd92-7eac2df7702b\") " Dec 06 05:54:00 crc kubenswrapper[4809]: I1206 05:54:00.218559 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c892e996-27b7-4b8d-bd92-7eac2df7702b-config-volume" (OuterVolumeSpecName: "config-volume") pod "c892e996-27b7-4b8d-bd92-7eac2df7702b" (UID: "c892e996-27b7-4b8d-bd92-7eac2df7702b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:54:00 crc kubenswrapper[4809]: I1206 05:54:00.218671 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-tphf5" Dec 06 05:54:00 crc kubenswrapper[4809]: I1206 05:54:00.219022 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-tphf5" Dec 06 05:54:00 crc kubenswrapper[4809]: I1206 05:54:00.223560 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c892e996-27b7-4b8d-bd92-7eac2df7702b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c892e996-27b7-4b8d-bd92-7eac2df7702b" (UID: "c892e996-27b7-4b8d-bd92-7eac2df7702b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:54:00 crc kubenswrapper[4809]: I1206 05:54:00.224754 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c892e996-27b7-4b8d-bd92-7eac2df7702b-kube-api-access-vxk8g" (OuterVolumeSpecName: "kube-api-access-vxk8g") pod "c892e996-27b7-4b8d-bd92-7eac2df7702b" (UID: "c892e996-27b7-4b8d-bd92-7eac2df7702b"). InnerVolumeSpecName "kube-api-access-vxk8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:54:00 crc kubenswrapper[4809]: I1206 05:54:00.246072 4809 patch_prober.go:28] interesting pod/apiserver-76f77b778f-tphf5 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 06 05:54:00 crc kubenswrapper[4809]: [+]log ok Dec 06 05:54:00 crc kubenswrapper[4809]: [+]etcd ok Dec 06 05:54:00 crc kubenswrapper[4809]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 06 05:54:00 crc kubenswrapper[4809]: [+]poststarthook/generic-apiserver-start-informers ok Dec 06 05:54:00 crc kubenswrapper[4809]: [+]poststarthook/max-in-flight-filter ok Dec 06 05:54:00 crc kubenswrapper[4809]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 06 05:54:00 crc kubenswrapper[4809]: [+]poststarthook/image.openshift.io-apiserver-caches ok Dec 06 05:54:00 crc kubenswrapper[4809]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Dec 06 05:54:00 crc kubenswrapper[4809]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Dec 06 05:54:00 crc kubenswrapper[4809]: [+]poststarthook/project.openshift.io-projectcache ok Dec 06 05:54:00 crc kubenswrapper[4809]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Dec 06 05:54:00 crc kubenswrapper[4809]: [+]poststarthook/openshift.io-startinformers ok Dec 06 05:54:00 crc kubenswrapper[4809]: [+]poststarthook/openshift.io-restmapperupdater ok Dec 06 05:54:00 crc kubenswrapper[4809]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 06 05:54:00 crc kubenswrapper[4809]: livez check failed Dec 06 05:54:00 crc kubenswrapper[4809]: I1206 05:54:00.246195 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-tphf5" podUID="09a7af3d-6bf2-4b7a-93d9-cecc73d993a1" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 05:54:00 crc kubenswrapper[4809]: I1206 05:54:00.319077 4809 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c892e996-27b7-4b8d-bd92-7eac2df7702b-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 05:54:00 crc kubenswrapper[4809]: I1206 05:54:00.319239 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxk8g\" (UniqueName: \"kubernetes.io/projected/c892e996-27b7-4b8d-bd92-7eac2df7702b-kube-api-access-vxk8g\") on node \"crc\" DevicePath \"\"" Dec 06 05:54:00 crc kubenswrapper[4809]: I1206 05:54:00.319256 4809 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c892e996-27b7-4b8d-bd92-7eac2df7702b-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 05:54:00 crc kubenswrapper[4809]: I1206 05:54:00.513308 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-tphf5" event={"ID":"09a7af3d-6bf2-4b7a-93d9-cecc73d993a1","Type":"ContainerStarted","Data":"c9d1eb9dd4a35f2aeba03cfe298833cc920886a3a084151f0d7a9b5d2e812f8b"} Dec 06 05:54:00 crc kubenswrapper[4809]: I1206 05:54:00.516729 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-p2lkt" event={"ID":"c598f580-a438-4015-9a8a-f98b3b2b0b3d","Type":"ContainerStarted","Data":"da27debccf839f53581ab5357700db3ad6dc646a24621bb226798397fed2b342"} Dec 06 05:54:00 crc kubenswrapper[4809]: I1206 05:54:00.516772 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-p2lkt" event={"ID":"c598f580-a438-4015-9a8a-f98b3b2b0b3d","Type":"ContainerStarted","Data":"22a125e47a10e7cc74157649c3195f94b238b36efe6ef8cdb1fda2a2f729b5be"} Dec 06 05:54:00 crc kubenswrapper[4809]: I1206 05:54:00.519768 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416665-lsfkp" event={"ID":"c892e996-27b7-4b8d-bd92-7eac2df7702b","Type":"ContainerDied","Data":"b73033163a7fa503fbc6be3c2b6de4efe81ecc843d0c92400f5406b9a9e7e344"} Dec 06 05:54:00 crc kubenswrapper[4809]: I1206 05:54:00.519802 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b73033163a7fa503fbc6be3c2b6de4efe81ecc843d0c92400f5406b9a9e7e344" Dec 06 05:54:00 crc kubenswrapper[4809]: I1206 05:54:00.519840 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416665-lsfkp" Dec 06 05:54:00 crc kubenswrapper[4809]: I1206 05:54:00.523666 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-9k8zx" event={"ID":"a4f75fd3-03e8-41e7-8926-f6ff21a5b681","Type":"ContainerStarted","Data":"6a2266bd60d66b63c582c65fd20237e3d836f37692f934b87a42b5b90d854c68"} Dec 06 05:54:00 crc kubenswrapper[4809]: I1206 05:54:00.525793 4809 generic.go:334] "Generic (PLEG): container finished" podID="3b196d34-f754-42a0-a82b-8a139e316aa2" containerID="6aca914b627feff21f991d038f746be15a6f1d49b21eae88c62f2f006454f22f" exitCode=0 Dec 06 05:54:00 crc kubenswrapper[4809]: I1206 05:54:00.526194 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"3b196d34-f754-42a0-a82b-8a139e316aa2","Type":"ContainerDied","Data":"6aca914b627feff21f991d038f746be15a6f1d49b21eae88c62f2f006454f22f"} Dec 06 05:54:00 crc kubenswrapper[4809]: I1206 05:54:00.541541 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-p2lkt" podStartSLOduration=19.541524112 podStartE2EDuration="19.541524112s" podCreationTimestamp="2025-12-06 05:53:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:54:00.537071751 +0000 UTC m=+165.426054693" watchObservedRunningTime="2025-12-06 05:54:00.541524112 +0000 UTC m=+165.430507054" Dec 06 05:54:00 crc kubenswrapper[4809]: I1206 05:54:00.580821 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-9k8zx" podStartSLOduration=146.580801357 podStartE2EDuration="2m26.580801357s" podCreationTimestamp="2025-12-06 05:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:54:00.558668842 +0000 UTC m=+165.447651784" watchObservedRunningTime="2025-12-06 05:54:00.580801357 +0000 UTC m=+165.469784309" Dec 06 05:54:00 crc kubenswrapper[4809]: I1206 05:54:00.622153 4809 patch_prober.go:28] interesting pod/router-default-5444994796-n6z57 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 05:54:00 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Dec 06 05:54:00 crc kubenswrapper[4809]: [+]process-running ok Dec 06 05:54:00 crc kubenswrapper[4809]: healthz check failed Dec 06 05:54:00 crc kubenswrapper[4809]: I1206 05:54:00.622491 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n6z57" podUID="e5356f1d-5afd-47bd-89af-72abfa8f901f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 05:54:01 crc kubenswrapper[4809]: I1206 05:54:01.622844 4809 patch_prober.go:28] interesting pod/router-default-5444994796-n6z57 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 05:54:01 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Dec 06 05:54:01 crc kubenswrapper[4809]: [+]process-running ok Dec 06 05:54:01 crc kubenswrapper[4809]: healthz check failed Dec 06 05:54:01 crc kubenswrapper[4809]: I1206 05:54:01.622913 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n6z57" podUID="e5356f1d-5afd-47bd-89af-72abfa8f901f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 05:54:01 crc kubenswrapper[4809]: I1206 05:54:01.841284 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 05:54:01 crc kubenswrapper[4809]: I1206 05:54:01.950187 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3b196d34-f754-42a0-a82b-8a139e316aa2-kubelet-dir\") pod \"3b196d34-f754-42a0-a82b-8a139e316aa2\" (UID: \"3b196d34-f754-42a0-a82b-8a139e316aa2\") " Dec 06 05:54:01 crc kubenswrapper[4809]: I1206 05:54:01.950245 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3b196d34-f754-42a0-a82b-8a139e316aa2-kube-api-access\") pod \"3b196d34-f754-42a0-a82b-8a139e316aa2\" (UID: \"3b196d34-f754-42a0-a82b-8a139e316aa2\") " Dec 06 05:54:01 crc kubenswrapper[4809]: I1206 05:54:01.950329 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3b196d34-f754-42a0-a82b-8a139e316aa2-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "3b196d34-f754-42a0-a82b-8a139e316aa2" (UID: "3b196d34-f754-42a0-a82b-8a139e316aa2"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 05:54:01 crc kubenswrapper[4809]: I1206 05:54:01.950606 4809 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3b196d34-f754-42a0-a82b-8a139e316aa2-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 06 05:54:01 crc kubenswrapper[4809]: I1206 05:54:01.955491 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b196d34-f754-42a0-a82b-8a139e316aa2-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "3b196d34-f754-42a0-a82b-8a139e316aa2" (UID: "3b196d34-f754-42a0-a82b-8a139e316aa2"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:54:02 crc kubenswrapper[4809]: I1206 05:54:02.051976 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3b196d34-f754-42a0-a82b-8a139e316aa2-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 05:54:02 crc kubenswrapper[4809]: I1206 05:54:02.543721 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"3b196d34-f754-42a0-a82b-8a139e316aa2","Type":"ContainerDied","Data":"1d1e9d23bf9c8cb529dba845b6a6839e16e74c197b7135a38713283ac826a26b"} Dec 06 05:54:02 crc kubenswrapper[4809]: I1206 05:54:02.543862 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d1e9d23bf9c8cb529dba845b6a6839e16e74c197b7135a38713283ac826a26b" Dec 06 05:54:02 crc kubenswrapper[4809]: I1206 05:54:02.543763 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 05:54:02 crc kubenswrapper[4809]: I1206 05:54:02.620435 4809 patch_prober.go:28] interesting pod/router-default-5444994796-n6z57 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 05:54:02 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Dec 06 05:54:02 crc kubenswrapper[4809]: [+]process-running ok Dec 06 05:54:02 crc kubenswrapper[4809]: healthz check failed Dec 06 05:54:02 crc kubenswrapper[4809]: I1206 05:54:02.620505 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n6z57" podUID="e5356f1d-5afd-47bd-89af-72abfa8f901f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 05:54:03 crc kubenswrapper[4809]: I1206 05:54:03.621625 4809 patch_prober.go:28] interesting pod/router-default-5444994796-n6z57 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 05:54:03 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Dec 06 05:54:03 crc kubenswrapper[4809]: [+]process-running ok Dec 06 05:54:03 crc kubenswrapper[4809]: healthz check failed Dec 06 05:54:03 crc kubenswrapper[4809]: I1206 05:54:03.621681 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n6z57" podUID="e5356f1d-5afd-47bd-89af-72abfa8f901f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 05:54:04 crc kubenswrapper[4809]: I1206 05:54:04.040238 4809 patch_prober.go:28] interesting pod/console-f9d7485db-4pwcw container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.6:8443/health\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Dec 06 05:54:04 crc kubenswrapper[4809]: I1206 05:54:04.040298 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-4pwcw" podUID="8e03364a-1840-4c91-8b2b-c7e2071fc73b" containerName="console" probeResult="failure" output="Get \"https://10.217.0.6:8443/health\": dial tcp 10.217.0.6:8443: connect: connection refused" Dec 06 05:54:04 crc kubenswrapper[4809]: I1206 05:54:04.483884 4809 patch_prober.go:28] interesting pod/downloads-7954f5f757-224mv container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.37:8080/\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Dec 06 05:54:04 crc kubenswrapper[4809]: I1206 05:54:04.483957 4809 patch_prober.go:28] interesting pod/downloads-7954f5f757-224mv container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Dec 06 05:54:04 crc kubenswrapper[4809]: I1206 05:54:04.483977 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-224mv" podUID="8e290142-b2fc-492e-8763-ff41224579fc" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.37:8080/\": dial tcp 10.217.0.37:8080: connect: connection refused" Dec 06 05:54:04 crc kubenswrapper[4809]: I1206 05:54:04.484007 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-224mv" podUID="8e290142-b2fc-492e-8763-ff41224579fc" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.37:8080/\": dial tcp 10.217.0.37:8080: connect: connection refused" Dec 06 05:54:04 crc kubenswrapper[4809]: I1206 05:54:04.496104 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 05:54:04 crc kubenswrapper[4809]: I1206 05:54:04.496182 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 05:54:04 crc kubenswrapper[4809]: I1206 05:54:04.620613 4809 patch_prober.go:28] interesting pod/router-default-5444994796-n6z57 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 05:54:04 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Dec 06 05:54:04 crc kubenswrapper[4809]: [+]process-running ok Dec 06 05:54:04 crc kubenswrapper[4809]: healthz check failed Dec 06 05:54:04 crc kubenswrapper[4809]: I1206 05:54:04.620672 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n6z57" podUID="e5356f1d-5afd-47bd-89af-72abfa8f901f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 05:54:05 crc kubenswrapper[4809]: I1206 05:54:05.212847 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-tphf5" Dec 06 05:54:05 crc kubenswrapper[4809]: I1206 05:54:05.219151 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-tphf5" Dec 06 05:54:05 crc kubenswrapper[4809]: I1206 05:54:05.623252 4809 patch_prober.go:28] interesting pod/router-default-5444994796-n6z57 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 05:54:05 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Dec 06 05:54:05 crc kubenswrapper[4809]: [+]process-running ok Dec 06 05:54:05 crc kubenswrapper[4809]: healthz check failed Dec 06 05:54:05 crc kubenswrapper[4809]: I1206 05:54:05.623895 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n6z57" podUID="e5356f1d-5afd-47bd-89af-72abfa8f901f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 05:54:06 crc kubenswrapper[4809]: I1206 05:54:06.619625 4809 patch_prober.go:28] interesting pod/router-default-5444994796-n6z57 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 05:54:06 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Dec 06 05:54:06 crc kubenswrapper[4809]: [+]process-running ok Dec 06 05:54:06 crc kubenswrapper[4809]: healthz check failed Dec 06 05:54:06 crc kubenswrapper[4809]: I1206 05:54:06.619687 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n6z57" podUID="e5356f1d-5afd-47bd-89af-72abfa8f901f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 05:54:07 crc kubenswrapper[4809]: I1206 05:54:07.619221 4809 patch_prober.go:28] interesting pod/router-default-5444994796-n6z57 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 05:54:07 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Dec 06 05:54:07 crc kubenswrapper[4809]: [+]process-running ok Dec 06 05:54:07 crc kubenswrapper[4809]: healthz check failed Dec 06 05:54:07 crc kubenswrapper[4809]: I1206 05:54:07.619279 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n6z57" podUID="e5356f1d-5afd-47bd-89af-72abfa8f901f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 05:54:08 crc kubenswrapper[4809]: I1206 05:54:08.619713 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-n6z57" Dec 06 05:54:08 crc kubenswrapper[4809]: I1206 05:54:08.625762 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-n6z57" Dec 06 05:54:14 crc kubenswrapper[4809]: I1206 05:54:14.426854 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-4pwcw" Dec 06 05:54:14 crc kubenswrapper[4809]: I1206 05:54:14.443299 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-4pwcw" Dec 06 05:54:14 crc kubenswrapper[4809]: I1206 05:54:14.495454 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-224mv" Dec 06 05:54:18 crc kubenswrapper[4809]: I1206 05:54:18.787153 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:54:23 crc kubenswrapper[4809]: I1206 05:54:23.526092 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:54:25 crc kubenswrapper[4809]: I1206 05:54:25.000358 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ch964" Dec 06 05:54:31 crc kubenswrapper[4809]: I1206 05:54:31.201237 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 06 05:54:31 crc kubenswrapper[4809]: E1206 05:54:31.202840 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b196d34-f754-42a0-a82b-8a139e316aa2" containerName="pruner" Dec 06 05:54:31 crc kubenswrapper[4809]: I1206 05:54:31.202963 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b196d34-f754-42a0-a82b-8a139e316aa2" containerName="pruner" Dec 06 05:54:31 crc kubenswrapper[4809]: E1206 05:54:31.203065 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c892e996-27b7-4b8d-bd92-7eac2df7702b" containerName="collect-profiles" Dec 06 05:54:31 crc kubenswrapper[4809]: I1206 05:54:31.203142 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c892e996-27b7-4b8d-bd92-7eac2df7702b" containerName="collect-profiles" Dec 06 05:54:31 crc kubenswrapper[4809]: I1206 05:54:31.203362 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="c892e996-27b7-4b8d-bd92-7eac2df7702b" containerName="collect-profiles" Dec 06 05:54:31 crc kubenswrapper[4809]: I1206 05:54:31.203460 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b196d34-f754-42a0-a82b-8a139e316aa2" containerName="pruner" Dec 06 05:54:31 crc kubenswrapper[4809]: I1206 05:54:31.204046 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 05:54:31 crc kubenswrapper[4809]: I1206 05:54:31.250096 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 06 05:54:31 crc kubenswrapper[4809]: I1206 05:54:31.258660 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 06 05:54:31 crc kubenswrapper[4809]: I1206 05:54:31.267226 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 06 05:54:31 crc kubenswrapper[4809]: I1206 05:54:31.271298 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7d53c879-8535-4785-8743-3c272df4d0cc-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7d53c879-8535-4785-8743-3c272df4d0cc\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 05:54:31 crc kubenswrapper[4809]: I1206 05:54:31.271501 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7d53c879-8535-4785-8743-3c272df4d0cc-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7d53c879-8535-4785-8743-3c272df4d0cc\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 05:54:31 crc kubenswrapper[4809]: I1206 05:54:31.372673 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7d53c879-8535-4785-8743-3c272df4d0cc-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7d53c879-8535-4785-8743-3c272df4d0cc\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 05:54:31 crc kubenswrapper[4809]: I1206 05:54:31.373030 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7d53c879-8535-4785-8743-3c272df4d0cc-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7d53c879-8535-4785-8743-3c272df4d0cc\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 05:54:31 crc kubenswrapper[4809]: I1206 05:54:31.373195 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7d53c879-8535-4785-8743-3c272df4d0cc-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7d53c879-8535-4785-8743-3c272df4d0cc\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 05:54:31 crc kubenswrapper[4809]: I1206 05:54:31.408305 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7d53c879-8535-4785-8743-3c272df4d0cc-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7d53c879-8535-4785-8743-3c272df4d0cc\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 05:54:31 crc kubenswrapper[4809]: I1206 05:54:31.583571 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 05:54:34 crc kubenswrapper[4809]: I1206 05:54:34.496472 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 05:54:34 crc kubenswrapper[4809]: I1206 05:54:34.496570 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 05:54:36 crc kubenswrapper[4809]: I1206 05:54:36.190982 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 06 05:54:36 crc kubenswrapper[4809]: I1206 05:54:36.192827 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 06 05:54:36 crc kubenswrapper[4809]: I1206 05:54:36.196717 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 06 05:54:36 crc kubenswrapper[4809]: I1206 05:54:36.238739 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/001a0f27-e6f3-4d62-9c25-43858654d2b6-var-lock\") pod \"installer-9-crc\" (UID: \"001a0f27-e6f3-4d62-9c25-43858654d2b6\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 05:54:36 crc kubenswrapper[4809]: I1206 05:54:36.238875 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/001a0f27-e6f3-4d62-9c25-43858654d2b6-kubelet-dir\") pod \"installer-9-crc\" (UID: \"001a0f27-e6f3-4d62-9c25-43858654d2b6\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 05:54:36 crc kubenswrapper[4809]: I1206 05:54:36.238906 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/001a0f27-e6f3-4d62-9c25-43858654d2b6-kube-api-access\") pod \"installer-9-crc\" (UID: \"001a0f27-e6f3-4d62-9c25-43858654d2b6\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 05:54:36 crc kubenswrapper[4809]: I1206 05:54:36.339640 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/001a0f27-e6f3-4d62-9c25-43858654d2b6-var-lock\") pod \"installer-9-crc\" (UID: \"001a0f27-e6f3-4d62-9c25-43858654d2b6\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 05:54:36 crc kubenswrapper[4809]: I1206 05:54:36.339753 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/001a0f27-e6f3-4d62-9c25-43858654d2b6-var-lock\") pod \"installer-9-crc\" (UID: \"001a0f27-e6f3-4d62-9c25-43858654d2b6\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 05:54:36 crc kubenswrapper[4809]: I1206 05:54:36.339783 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/001a0f27-e6f3-4d62-9c25-43858654d2b6-kubelet-dir\") pod \"installer-9-crc\" (UID: \"001a0f27-e6f3-4d62-9c25-43858654d2b6\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 05:54:36 crc kubenswrapper[4809]: I1206 05:54:36.339814 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/001a0f27-e6f3-4d62-9c25-43858654d2b6-kubelet-dir\") pod \"installer-9-crc\" (UID: \"001a0f27-e6f3-4d62-9c25-43858654d2b6\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 05:54:36 crc kubenswrapper[4809]: I1206 05:54:36.339814 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/001a0f27-e6f3-4d62-9c25-43858654d2b6-kube-api-access\") pod \"installer-9-crc\" (UID: \"001a0f27-e6f3-4d62-9c25-43858654d2b6\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 05:54:36 crc kubenswrapper[4809]: I1206 05:54:36.359825 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/001a0f27-e6f3-4d62-9c25-43858654d2b6-kube-api-access\") pod \"installer-9-crc\" (UID: \"001a0f27-e6f3-4d62-9c25-43858654d2b6\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 05:54:36 crc kubenswrapper[4809]: I1206 05:54:36.524862 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 06 05:55:04 crc kubenswrapper[4809]: I1206 05:55:04.496442 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 05:55:04 crc kubenswrapper[4809]: I1206 05:55:04.497131 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 05:55:04 crc kubenswrapper[4809]: I1206 05:55:04.497191 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-npms2" Dec 06 05:55:04 crc kubenswrapper[4809]: I1206 05:55:04.497839 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"930788ec9b4d97fdc39353a93111b84707037ce445f1930aa7737a20a44e1d37"} pod="openshift-machine-config-operator/machine-config-daemon-npms2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 05:55:04 crc kubenswrapper[4809]: I1206 05:55:04.497979 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" containerID="cri-o://930788ec9b4d97fdc39353a93111b84707037ce445f1930aa7737a20a44e1d37" gracePeriod=600 Dec 06 05:55:05 crc kubenswrapper[4809]: I1206 05:55:05.905113 4809 generic.go:334] "Generic (PLEG): container finished" podID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerID="930788ec9b4d97fdc39353a93111b84707037ce445f1930aa7737a20a44e1d37" exitCode=0 Dec 06 05:55:05 crc kubenswrapper[4809]: I1206 05:55:05.905216 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" event={"ID":"cbf4cf62-024e-4703-a8b8-9aecda9cd26a","Type":"ContainerDied","Data":"930788ec9b4d97fdc39353a93111b84707037ce445f1930aa7737a20a44e1d37"} Dec 06 05:55:21 crc kubenswrapper[4809]: E1206 05:55:21.356627 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 06 05:55:21 crc kubenswrapper[4809]: E1206 05:55:21.357482 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ffd5h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-ls5nx_openshift-marketplace(b475ab56-d949-44e5-a480-f59d8a1d737e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 05:55:21 crc kubenswrapper[4809]: E1206 05:55:21.358737 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-ls5nx" podUID="b475ab56-d949-44e5-a480-f59d8a1d737e" Dec 06 05:55:26 crc kubenswrapper[4809]: E1206 05:55:26.160981 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-ls5nx" podUID="b475ab56-d949-44e5-a480-f59d8a1d737e" Dec 06 05:55:26 crc kubenswrapper[4809]: E1206 05:55:26.378751 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 06 05:55:26 crc kubenswrapper[4809]: E1206 05:55:26.378977 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p7f6b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-9zm48_openshift-marketplace(c822a806-352f-4b4f-8549-051ff20c4ca2): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 05:55:26 crc kubenswrapper[4809]: E1206 05:55:26.380260 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-9zm48" podUID="c822a806-352f-4b4f-8549-051ff20c4ca2" Dec 06 05:55:29 crc kubenswrapper[4809]: E1206 05:55:29.498318 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-9zm48" podUID="c822a806-352f-4b4f-8549-051ff20c4ca2" Dec 06 05:55:29 crc kubenswrapper[4809]: E1206 05:55:29.802659 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 06 05:55:29 crc kubenswrapper[4809]: E1206 05:55:29.802793 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-z5kmb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-hfgk4_openshift-marketplace(e356b822-4e99-42f8-99ba-7aea197f3d95): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 05:55:29 crc kubenswrapper[4809]: E1206 05:55:29.804072 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-hfgk4" podUID="e356b822-4e99-42f8-99ba-7aea197f3d95" Dec 06 05:55:30 crc kubenswrapper[4809]: E1206 05:55:30.021642 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 06 05:55:30 crc kubenswrapper[4809]: E1206 05:55:30.021815 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ld72t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-24qxt_openshift-marketplace(893020b2-fa30-4371-86f9-a585517839ce): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 05:55:30 crc kubenswrapper[4809]: E1206 05:55:30.023011 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-24qxt" podUID="893020b2-fa30-4371-86f9-a585517839ce" Dec 06 05:55:31 crc kubenswrapper[4809]: E1206 05:55:31.218784 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-hfgk4" podUID="e356b822-4e99-42f8-99ba-7aea197f3d95" Dec 06 05:55:31 crc kubenswrapper[4809]: E1206 05:55:31.218838 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-24qxt" podUID="893020b2-fa30-4371-86f9-a585517839ce" Dec 06 05:55:31 crc kubenswrapper[4809]: E1206 05:55:31.243167 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 06 05:55:31 crc kubenswrapper[4809]: E1206 05:55:31.243625 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dssmc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-lpb28_openshift-marketplace(40a90e94-16ee-42ba-beff-5103476432c1): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 05:55:31 crc kubenswrapper[4809]: E1206 05:55:31.245117 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-lpb28" podUID="40a90e94-16ee-42ba-beff-5103476432c1" Dec 06 05:55:31 crc kubenswrapper[4809]: I1206 05:55:31.481709 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 06 05:55:31 crc kubenswrapper[4809]: W1206 05:55:31.485797 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod001a0f27_e6f3_4d62_9c25_43858654d2b6.slice/crio-875ca06589b2bb882a691d83ec8796748443f3cff33e3a2eea843c9736908ff1 WatchSource:0}: Error finding container 875ca06589b2bb882a691d83ec8796748443f3cff33e3a2eea843c9736908ff1: Status 404 returned error can't find the container with id 875ca06589b2bb882a691d83ec8796748443f3cff33e3a2eea843c9736908ff1 Dec 06 05:55:31 crc kubenswrapper[4809]: E1206 05:55:31.600514 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 06 05:55:31 crc kubenswrapper[4809]: E1206 05:55:31.600674 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-j9b9l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-dvwjq_openshift-marketplace(a83cbcc6-62ac-4814-9fce-eb7ea465bcbf): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 05:55:31 crc kubenswrapper[4809]: E1206 05:55:31.601798 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-dvwjq" podUID="a83cbcc6-62ac-4814-9fce-eb7ea465bcbf" Dec 06 05:55:31 crc kubenswrapper[4809]: I1206 05:55:31.606718 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 06 05:55:31 crc kubenswrapper[4809]: W1206 05:55:31.616515 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod7d53c879_8535_4785_8743_3c272df4d0cc.slice/crio-4458ed8a8a5950da5fa8761f56291cd30d59fee1c02faab0772c1c24a5277e43 WatchSource:0}: Error finding container 4458ed8a8a5950da5fa8761f56291cd30d59fee1c02faab0772c1c24a5277e43: Status 404 returned error can't find the container with id 4458ed8a8a5950da5fa8761f56291cd30d59fee1c02faab0772c1c24a5277e43 Dec 06 05:55:31 crc kubenswrapper[4809]: E1206 05:55:31.841925 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 06 05:55:31 crc kubenswrapper[4809]: E1206 05:55:31.842460 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qwbq2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-zk7tm_openshift-marketplace(5e67c2b8-8bd1-466c-98d8-e93d411162ba): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 05:55:31 crc kubenswrapper[4809]: E1206 05:55:31.844024 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-zk7tm" podUID="5e67c2b8-8bd1-466c-98d8-e93d411162ba" Dec 06 05:55:32 crc kubenswrapper[4809]: I1206 05:55:32.046572 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" event={"ID":"cbf4cf62-024e-4703-a8b8-9aecda9cd26a","Type":"ContainerStarted","Data":"2442b241343358fa541b07551125d7c63b82fcdcd98be18afedabbdc8da517a8"} Dec 06 05:55:32 crc kubenswrapper[4809]: I1206 05:55:32.047747 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"001a0f27-e6f3-4d62-9c25-43858654d2b6","Type":"ContainerStarted","Data":"875ca06589b2bb882a691d83ec8796748443f3cff33e3a2eea843c9736908ff1"} Dec 06 05:55:32 crc kubenswrapper[4809]: I1206 05:55:32.048900 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"7d53c879-8535-4785-8743-3c272df4d0cc","Type":"ContainerStarted","Data":"4458ed8a8a5950da5fa8761f56291cd30d59fee1c02faab0772c1c24a5277e43"} Dec 06 05:55:32 crc kubenswrapper[4809]: E1206 05:55:32.050966 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-zk7tm" podUID="5e67c2b8-8bd1-466c-98d8-e93d411162ba" Dec 06 05:55:32 crc kubenswrapper[4809]: E1206 05:55:32.051129 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-lpb28" podUID="40a90e94-16ee-42ba-beff-5103476432c1" Dec 06 05:55:32 crc kubenswrapper[4809]: E1206 05:55:32.051205 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-dvwjq" podUID="a83cbcc6-62ac-4814-9fce-eb7ea465bcbf" Dec 06 05:55:32 crc kubenswrapper[4809]: E1206 05:55:32.698018 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 06 05:55:32 crc kubenswrapper[4809]: E1206 05:55:32.698442 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gglkq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-wj2h7_openshift-marketplace(2c4b2f56-0b03-41ae-af23-6192b8bd3d34): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 05:55:32 crc kubenswrapper[4809]: E1206 05:55:32.699605 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-wj2h7" podUID="2c4b2f56-0b03-41ae-af23-6192b8bd3d34" Dec 06 05:55:33 crc kubenswrapper[4809]: I1206 05:55:33.057465 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"7d53c879-8535-4785-8743-3c272df4d0cc","Type":"ContainerStarted","Data":"537a5f71ebdd77425c4d4fbbd1147383d36a68d5bccfe17fa62af2234ef3b3b3"} Dec 06 05:55:33 crc kubenswrapper[4809]: I1206 05:55:33.059596 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"001a0f27-e6f3-4d62-9c25-43858654d2b6","Type":"ContainerStarted","Data":"2ec322a0f85b3bd7cc2e0b41458a3e30fc7803aec81398f0c39be0aa2b803b8c"} Dec 06 05:55:33 crc kubenswrapper[4809]: E1206 05:55:33.065753 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wj2h7" podUID="2c4b2f56-0b03-41ae-af23-6192b8bd3d34" Dec 06 05:55:33 crc kubenswrapper[4809]: I1206 05:55:33.081123 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=62.081096891 podStartE2EDuration="1m2.081096891s" podCreationTimestamp="2025-12-06 05:54:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:55:33.070259958 +0000 UTC m=+257.959242900" watchObservedRunningTime="2025-12-06 05:55:33.081096891 +0000 UTC m=+257.970079833" Dec 06 05:55:33 crc kubenswrapper[4809]: I1206 05:55:33.092898 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=57.092878851 podStartE2EDuration="57.092878851s" podCreationTimestamp="2025-12-06 05:54:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:55:33.091051599 +0000 UTC m=+257.980034541" watchObservedRunningTime="2025-12-06 05:55:33.092878851 +0000 UTC m=+257.981861793" Dec 06 05:55:34 crc kubenswrapper[4809]: I1206 05:55:34.063713 4809 generic.go:334] "Generic (PLEG): container finished" podID="7d53c879-8535-4785-8743-3c272df4d0cc" containerID="537a5f71ebdd77425c4d4fbbd1147383d36a68d5bccfe17fa62af2234ef3b3b3" exitCode=0 Dec 06 05:55:34 crc kubenswrapper[4809]: I1206 05:55:34.065447 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"7d53c879-8535-4785-8743-3c272df4d0cc","Type":"ContainerDied","Data":"537a5f71ebdd77425c4d4fbbd1147383d36a68d5bccfe17fa62af2234ef3b3b3"} Dec 06 05:55:35 crc kubenswrapper[4809]: I1206 05:55:35.315156 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 05:55:35 crc kubenswrapper[4809]: I1206 05:55:35.364896 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7d53c879-8535-4785-8743-3c272df4d0cc-kube-api-access\") pod \"7d53c879-8535-4785-8743-3c272df4d0cc\" (UID: \"7d53c879-8535-4785-8743-3c272df4d0cc\") " Dec 06 05:55:35 crc kubenswrapper[4809]: I1206 05:55:35.365094 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7d53c879-8535-4785-8743-3c272df4d0cc-kubelet-dir\") pod \"7d53c879-8535-4785-8743-3c272df4d0cc\" (UID: \"7d53c879-8535-4785-8743-3c272df4d0cc\") " Dec 06 05:55:35 crc kubenswrapper[4809]: I1206 05:55:35.365323 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7d53c879-8535-4785-8743-3c272df4d0cc-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "7d53c879-8535-4785-8743-3c272df4d0cc" (UID: "7d53c879-8535-4785-8743-3c272df4d0cc"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 05:55:35 crc kubenswrapper[4809]: I1206 05:55:35.372286 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d53c879-8535-4785-8743-3c272df4d0cc-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "7d53c879-8535-4785-8743-3c272df4d0cc" (UID: "7d53c879-8535-4785-8743-3c272df4d0cc"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:55:35 crc kubenswrapper[4809]: I1206 05:55:35.466889 4809 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7d53c879-8535-4785-8743-3c272df4d0cc-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 06 05:55:35 crc kubenswrapper[4809]: I1206 05:55:35.467350 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7d53c879-8535-4785-8743-3c272df4d0cc-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 05:55:36 crc kubenswrapper[4809]: I1206 05:55:36.080801 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"7d53c879-8535-4785-8743-3c272df4d0cc","Type":"ContainerDied","Data":"4458ed8a8a5950da5fa8761f56291cd30d59fee1c02faab0772c1c24a5277e43"} Dec 06 05:55:36 crc kubenswrapper[4809]: I1206 05:55:36.080844 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4458ed8a8a5950da5fa8761f56291cd30d59fee1c02faab0772c1c24a5277e43" Dec 06 05:55:36 crc kubenswrapper[4809]: I1206 05:55:36.080869 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 05:55:43 crc kubenswrapper[4809]: I1206 05:55:43.124529 4809 generic.go:334] "Generic (PLEG): container finished" podID="b475ab56-d949-44e5-a480-f59d8a1d737e" containerID="a467d9fc608d31088c67940d3f2b4e00f273fceb66151a86992caba5f306857d" exitCode=0 Dec 06 05:55:43 crc kubenswrapper[4809]: I1206 05:55:43.124586 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ls5nx" event={"ID":"b475ab56-d949-44e5-a480-f59d8a1d737e","Type":"ContainerDied","Data":"a467d9fc608d31088c67940d3f2b4e00f273fceb66151a86992caba5f306857d"} Dec 06 05:55:45 crc kubenswrapper[4809]: I1206 05:55:45.139925 4809 generic.go:334] "Generic (PLEG): container finished" podID="5e67c2b8-8bd1-466c-98d8-e93d411162ba" containerID="ca19224db7d03a2501accfdaa1cf9b66092ad9c0d0ea63def51a482eff77fc61" exitCode=0 Dec 06 05:55:45 crc kubenswrapper[4809]: I1206 05:55:45.139978 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zk7tm" event={"ID":"5e67c2b8-8bd1-466c-98d8-e93d411162ba","Type":"ContainerDied","Data":"ca19224db7d03a2501accfdaa1cf9b66092ad9c0d0ea63def51a482eff77fc61"} Dec 06 05:55:45 crc kubenswrapper[4809]: I1206 05:55:45.143329 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ls5nx" event={"ID":"b475ab56-d949-44e5-a480-f59d8a1d737e","Type":"ContainerStarted","Data":"b37ea71fa86c9f42445ffa12712950964a2ca002475e4d87b6c0bf1e14ac18be"} Dec 06 05:55:45 crc kubenswrapper[4809]: I1206 05:55:45.419227 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ls5nx" podStartSLOduration=7.828032755 podStartE2EDuration="1m52.419210359s" podCreationTimestamp="2025-12-06 05:53:53 +0000 UTC" firstStartedPulling="2025-12-06 05:53:59.428384598 +0000 UTC m=+164.317367540" lastFinishedPulling="2025-12-06 05:55:44.019562202 +0000 UTC m=+268.908545144" observedRunningTime="2025-12-06 05:55:45.18299854 +0000 UTC m=+270.071981482" watchObservedRunningTime="2025-12-06 05:55:45.419210359 +0000 UTC m=+270.308193301" Dec 06 05:55:45 crc kubenswrapper[4809]: E1206 05:55:45.462385 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[networking-console-plugin-cert nginx-conf], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:55:46 crc kubenswrapper[4809]: I1206 05:55:46.150834 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lpb28" event={"ID":"40a90e94-16ee-42ba-beff-5103476432c1","Type":"ContainerStarted","Data":"29dd3f6f4e238c2aaea07bfb17de884e652c4fc6d66a928725123e15777f54c0"} Dec 06 05:55:46 crc kubenswrapper[4809]: I1206 05:55:46.427994 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:55:46 crc kubenswrapper[4809]: I1206 05:55:46.428732 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:55:46 crc kubenswrapper[4809]: I1206 05:55:46.429524 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:55:46 crc kubenswrapper[4809]: I1206 05:55:46.436149 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:55:47 crc kubenswrapper[4809]: I1206 05:55:47.156370 4809 generic.go:334] "Generic (PLEG): container finished" podID="40a90e94-16ee-42ba-beff-5103476432c1" containerID="29dd3f6f4e238c2aaea07bfb17de884e652c4fc6d66a928725123e15777f54c0" exitCode=0 Dec 06 05:55:47 crc kubenswrapper[4809]: I1206 05:55:47.156433 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lpb28" event={"ID":"40a90e94-16ee-42ba-beff-5103476432c1","Type":"ContainerDied","Data":"29dd3f6f4e238c2aaea07bfb17de884e652c4fc6d66a928725123e15777f54c0"} Dec 06 05:55:47 crc kubenswrapper[4809]: I1206 05:55:47.159802 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hfgk4" event={"ID":"e356b822-4e99-42f8-99ba-7aea197f3d95","Type":"ContainerStarted","Data":"d9effd39b1fa4f05d488767fd74860ce039f4a882d6439ccb3930ab59a749496"} Dec 06 05:55:47 crc kubenswrapper[4809]: I1206 05:55:47.162566 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-24qxt" event={"ID":"893020b2-fa30-4371-86f9-a585517839ce","Type":"ContainerStarted","Data":"8cc7549375e24b0c91ac812b6424f92854ecac7d3424bdee33188d237ec03f3c"} Dec 06 05:55:47 crc kubenswrapper[4809]: I1206 05:55:47.164590 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zk7tm" event={"ID":"5e67c2b8-8bd1-466c-98d8-e93d411162ba","Type":"ContainerStarted","Data":"20b718263143dd67f4af46e1f388040c1db9cc32b952a8cc6f35f0c6f8039868"} Dec 06 05:55:47 crc kubenswrapper[4809]: I1206 05:55:47.222428 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zk7tm" podStartSLOduration=7.020064919 podStartE2EDuration="1m56.222412121s" podCreationTimestamp="2025-12-06 05:53:51 +0000 UTC" firstStartedPulling="2025-12-06 05:53:57.361077288 +0000 UTC m=+162.250060230" lastFinishedPulling="2025-12-06 05:55:46.56342449 +0000 UTC m=+271.452407432" observedRunningTime="2025-12-06 05:55:47.219345587 +0000 UTC m=+272.108328529" watchObservedRunningTime="2025-12-06 05:55:47.222412121 +0000 UTC m=+272.111395063" Dec 06 05:55:48 crc kubenswrapper[4809]: I1206 05:55:48.172550 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9zm48" event={"ID":"c822a806-352f-4b4f-8549-051ff20c4ca2","Type":"ContainerStarted","Data":"f9d032773c1b7216ba187e731f469d9970a1224cf425291d216f7e609ce79973"} Dec 06 05:55:48 crc kubenswrapper[4809]: I1206 05:55:48.174597 4809 generic.go:334] "Generic (PLEG): container finished" podID="e356b822-4e99-42f8-99ba-7aea197f3d95" containerID="d9effd39b1fa4f05d488767fd74860ce039f4a882d6439ccb3930ab59a749496" exitCode=0 Dec 06 05:55:48 crc kubenswrapper[4809]: I1206 05:55:48.174636 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hfgk4" event={"ID":"e356b822-4e99-42f8-99ba-7aea197f3d95","Type":"ContainerDied","Data":"d9effd39b1fa4f05d488767fd74860ce039f4a882d6439ccb3930ab59a749496"} Dec 06 05:55:48 crc kubenswrapper[4809]: I1206 05:55:48.177074 4809 generic.go:334] "Generic (PLEG): container finished" podID="893020b2-fa30-4371-86f9-a585517839ce" containerID="8cc7549375e24b0c91ac812b6424f92854ecac7d3424bdee33188d237ec03f3c" exitCode=0 Dec 06 05:55:48 crc kubenswrapper[4809]: I1206 05:55:48.177109 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-24qxt" event={"ID":"893020b2-fa30-4371-86f9-a585517839ce","Type":"ContainerDied","Data":"8cc7549375e24b0c91ac812b6424f92854ecac7d3424bdee33188d237ec03f3c"} Dec 06 05:55:48 crc kubenswrapper[4809]: E1206 05:55:48.348991 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc822a806_352f_4b4f_8549_051ff20c4ca2.slice/crio-conmon-f9d032773c1b7216ba187e731f469d9970a1224cf425291d216f7e609ce79973.scope\": RecentStats: unable to find data in memory cache]" Dec 06 05:55:49 crc kubenswrapper[4809]: I1206 05:55:49.185166 4809 generic.go:334] "Generic (PLEG): container finished" podID="c822a806-352f-4b4f-8549-051ff20c4ca2" containerID="f9d032773c1b7216ba187e731f469d9970a1224cf425291d216f7e609ce79973" exitCode=0 Dec 06 05:55:49 crc kubenswrapper[4809]: I1206 05:55:49.185217 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9zm48" event={"ID":"c822a806-352f-4b4f-8549-051ff20c4ca2","Type":"ContainerDied","Data":"f9d032773c1b7216ba187e731f469d9970a1224cf425291d216f7e609ce79973"} Dec 06 05:55:52 crc kubenswrapper[4809]: I1206 05:55:52.167044 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zk7tm" Dec 06 05:55:52 crc kubenswrapper[4809]: I1206 05:55:52.167621 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zk7tm" Dec 06 05:55:53 crc kubenswrapper[4809]: I1206 05:55:53.118148 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zk7tm" Dec 06 05:55:53 crc kubenswrapper[4809]: I1206 05:55:53.160534 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zk7tm" Dec 06 05:55:54 crc kubenswrapper[4809]: I1206 05:55:54.221710 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ls5nx" Dec 06 05:55:54 crc kubenswrapper[4809]: I1206 05:55:54.222500 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ls5nx" Dec 06 05:55:54 crc kubenswrapper[4809]: I1206 05:55:54.274779 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ls5nx" Dec 06 05:55:54 crc kubenswrapper[4809]: I1206 05:55:54.660703 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jn446"] Dec 06 05:55:55 crc kubenswrapper[4809]: I1206 05:55:55.219588 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lpb28" event={"ID":"40a90e94-16ee-42ba-beff-5103476432c1","Type":"ContainerStarted","Data":"12e34109bddd8e50e1a581763ffb23958cb523a827d40e9a9dd070c15da134dc"} Dec 06 05:55:55 crc kubenswrapper[4809]: I1206 05:55:55.266251 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ls5nx" Dec 06 05:55:55 crc kubenswrapper[4809]: I1206 05:55:55.280619 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lpb28" podStartSLOduration=9.42179708 podStartE2EDuration="2m1.280595644s" podCreationTimestamp="2025-12-06 05:53:54 +0000 UTC" firstStartedPulling="2025-12-06 05:53:59.521163137 +0000 UTC m=+164.410146079" lastFinishedPulling="2025-12-06 05:55:51.379961701 +0000 UTC m=+276.268944643" observedRunningTime="2025-12-06 05:55:55.241344349 +0000 UTC m=+280.130327291" watchObservedRunningTime="2025-12-06 05:55:55.280595644 +0000 UTC m=+280.169578606" Dec 06 05:55:55 crc kubenswrapper[4809]: I1206 05:55:55.958341 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lpb28" Dec 06 05:55:55 crc kubenswrapper[4809]: I1206 05:55:55.958405 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lpb28" Dec 06 05:55:57 crc kubenswrapper[4809]: I1206 05:55:57.042881 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lpb28" podUID="40a90e94-16ee-42ba-beff-5103476432c1" containerName="registry-server" probeResult="failure" output=< Dec 06 05:55:57 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Dec 06 05:55:57 crc kubenswrapper[4809]: > Dec 06 05:55:59 crc kubenswrapper[4809]: I1206 05:55:57.355180 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ls5nx"] Dec 06 05:55:59 crc kubenswrapper[4809]: I1206 05:55:58.234256 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ls5nx" podUID="b475ab56-d949-44e5-a480-f59d8a1d737e" containerName="registry-server" containerID="cri-o://b37ea71fa86c9f42445ffa12712950964a2ca002475e4d87b6c0bf1e14ac18be" gracePeriod=2 Dec 06 05:55:59 crc kubenswrapper[4809]: I1206 05:55:59.388183 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:55:59 crc kubenswrapper[4809]: I1206 05:55:59.388687 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:56:00 crc kubenswrapper[4809]: I1206 05:56:00.248411 4809 generic.go:334] "Generic (PLEG): container finished" podID="b475ab56-d949-44e5-a480-f59d8a1d737e" containerID="b37ea71fa86c9f42445ffa12712950964a2ca002475e4d87b6c0bf1e14ac18be" exitCode=0 Dec 06 05:56:00 crc kubenswrapper[4809]: I1206 05:56:00.248561 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ls5nx" event={"ID":"b475ab56-d949-44e5-a480-f59d8a1d737e","Type":"ContainerDied","Data":"b37ea71fa86c9f42445ffa12712950964a2ca002475e4d87b6c0bf1e14ac18be"} Dec 06 05:56:00 crc kubenswrapper[4809]: I1206 05:56:00.523580 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ls5nx" Dec 06 05:56:00 crc kubenswrapper[4809]: I1206 05:56:00.705813 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffd5h\" (UniqueName: \"kubernetes.io/projected/b475ab56-d949-44e5-a480-f59d8a1d737e-kube-api-access-ffd5h\") pod \"b475ab56-d949-44e5-a480-f59d8a1d737e\" (UID: \"b475ab56-d949-44e5-a480-f59d8a1d737e\") " Dec 06 05:56:00 crc kubenswrapper[4809]: I1206 05:56:00.707157 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b475ab56-d949-44e5-a480-f59d8a1d737e-catalog-content\") pod \"b475ab56-d949-44e5-a480-f59d8a1d737e\" (UID: \"b475ab56-d949-44e5-a480-f59d8a1d737e\") " Dec 06 05:56:00 crc kubenswrapper[4809]: I1206 05:56:00.707192 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b475ab56-d949-44e5-a480-f59d8a1d737e-utilities\") pod \"b475ab56-d949-44e5-a480-f59d8a1d737e\" (UID: \"b475ab56-d949-44e5-a480-f59d8a1d737e\") " Dec 06 05:56:00 crc kubenswrapper[4809]: I1206 05:56:00.707993 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b475ab56-d949-44e5-a480-f59d8a1d737e-utilities" (OuterVolumeSpecName: "utilities") pod "b475ab56-d949-44e5-a480-f59d8a1d737e" (UID: "b475ab56-d949-44e5-a480-f59d8a1d737e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:56:00 crc kubenswrapper[4809]: I1206 05:56:00.718982 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b475ab56-d949-44e5-a480-f59d8a1d737e-kube-api-access-ffd5h" (OuterVolumeSpecName: "kube-api-access-ffd5h") pod "b475ab56-d949-44e5-a480-f59d8a1d737e" (UID: "b475ab56-d949-44e5-a480-f59d8a1d737e"). InnerVolumeSpecName "kube-api-access-ffd5h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:56:00 crc kubenswrapper[4809]: I1206 05:56:00.729377 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b475ab56-d949-44e5-a480-f59d8a1d737e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b475ab56-d949-44e5-a480-f59d8a1d737e" (UID: "b475ab56-d949-44e5-a480-f59d8a1d737e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:56:00 crc kubenswrapper[4809]: W1206 05:56:00.762459 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-124456cbc04713ff2318a8310f60b00466c18efaab00101bcc8be8cee0666b83 WatchSource:0}: Error finding container 124456cbc04713ff2318a8310f60b00466c18efaab00101bcc8be8cee0666b83: Status 404 returned error can't find the container with id 124456cbc04713ff2318a8310f60b00466c18efaab00101bcc8be8cee0666b83 Dec 06 05:56:00 crc kubenswrapper[4809]: I1206 05:56:00.809296 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b475ab56-d949-44e5-a480-f59d8a1d737e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 05:56:00 crc kubenswrapper[4809]: I1206 05:56:00.809703 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b475ab56-d949-44e5-a480-f59d8a1d737e-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 05:56:00 crc kubenswrapper[4809]: I1206 05:56:00.809716 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffd5h\" (UniqueName: \"kubernetes.io/projected/b475ab56-d949-44e5-a480-f59d8a1d737e-kube-api-access-ffd5h\") on node \"crc\" DevicePath \"\"" Dec 06 05:56:01 crc kubenswrapper[4809]: I1206 05:56:01.260478 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-24qxt" event={"ID":"893020b2-fa30-4371-86f9-a585517839ce","Type":"ContainerStarted","Data":"6ad127ae57fe37bc2ab09e2d03a74dd79af68385bbfee5069273b6c01f037ccf"} Dec 06 05:56:01 crc kubenswrapper[4809]: I1206 05:56:01.262449 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ls5nx" event={"ID":"b475ab56-d949-44e5-a480-f59d8a1d737e","Type":"ContainerDied","Data":"f413b3d3dd61ba7fc1ac276f14b507e33ab610d693fceb2d105f6d66bcfbb406"} Dec 06 05:56:01 crc kubenswrapper[4809]: I1206 05:56:01.262486 4809 scope.go:117] "RemoveContainer" containerID="b37ea71fa86c9f42445ffa12712950964a2ca002475e4d87b6c0bf1e14ac18be" Dec 06 05:56:01 crc kubenswrapper[4809]: I1206 05:56:01.262493 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ls5nx" Dec 06 05:56:01 crc kubenswrapper[4809]: I1206 05:56:01.264606 4809 generic.go:334] "Generic (PLEG): container finished" podID="2c4b2f56-0b03-41ae-af23-6192b8bd3d34" containerID="fbf084ae1a873d7f4aa9bb837f4ea622d95d00fc7f6da3f8aecf04c74bf90da9" exitCode=0 Dec 06 05:56:01 crc kubenswrapper[4809]: I1206 05:56:01.264657 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wj2h7" event={"ID":"2c4b2f56-0b03-41ae-af23-6192b8bd3d34","Type":"ContainerDied","Data":"fbf084ae1a873d7f4aa9bb837f4ea622d95d00fc7f6da3f8aecf04c74bf90da9"} Dec 06 05:56:01 crc kubenswrapper[4809]: I1206 05:56:01.268728 4809 generic.go:334] "Generic (PLEG): container finished" podID="a83cbcc6-62ac-4814-9fce-eb7ea465bcbf" containerID="d9cb75f3a92a568740557049d70f009040650c11c5fe919255949f3f9d01b275" exitCode=0 Dec 06 05:56:01 crc kubenswrapper[4809]: I1206 05:56:01.268788 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dvwjq" event={"ID":"a83cbcc6-62ac-4814-9fce-eb7ea465bcbf","Type":"ContainerDied","Data":"d9cb75f3a92a568740557049d70f009040650c11c5fe919255949f3f9d01b275"} Dec 06 05:56:01 crc kubenswrapper[4809]: I1206 05:56:01.272676 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9zm48" event={"ID":"c822a806-352f-4b4f-8549-051ff20c4ca2","Type":"ContainerStarted","Data":"ee55f33e692c658c851dc535b2c6cf2b3c68342211988109544307f25874ea9e"} Dec 06 05:56:01 crc kubenswrapper[4809]: I1206 05:56:01.273863 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"bfa32bbb42e171ec6dd5d268f7979b86e134c67ef76f5245d1f534cd2bb9a410"} Dec 06 05:56:01 crc kubenswrapper[4809]: I1206 05:56:01.273896 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"124456cbc04713ff2318a8310f60b00466c18efaab00101bcc8be8cee0666b83"} Dec 06 05:56:01 crc kubenswrapper[4809]: I1206 05:56:01.280456 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hfgk4" event={"ID":"e356b822-4e99-42f8-99ba-7aea197f3d95","Type":"ContainerStarted","Data":"284f02413d4d6e0a2fe38ddb39686def4715cf410535ea504dc82e53c186ee64"} Dec 06 05:56:01 crc kubenswrapper[4809]: I1206 05:56:01.286759 4809 scope.go:117] "RemoveContainer" containerID="a467d9fc608d31088c67940d3f2b4e00f273fceb66151a86992caba5f306857d" Dec 06 05:56:01 crc kubenswrapper[4809]: I1206 05:56:01.305886 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-24qxt" podStartSLOduration=3.121408485 podStartE2EDuration="2m10.305864833s" podCreationTimestamp="2025-12-06 05:53:51 +0000 UTC" firstStartedPulling="2025-12-06 05:53:53.19397469 +0000 UTC m=+158.082957632" lastFinishedPulling="2025-12-06 05:56:00.378431038 +0000 UTC m=+285.267413980" observedRunningTime="2025-12-06 05:56:01.284892448 +0000 UTC m=+286.173875390" watchObservedRunningTime="2025-12-06 05:56:01.305864833 +0000 UTC m=+286.194847775" Dec 06 05:56:01 crc kubenswrapper[4809]: I1206 05:56:01.308984 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hfgk4" podStartSLOduration=5.529400453 podStartE2EDuration="2m6.308973437s" podCreationTimestamp="2025-12-06 05:53:55 +0000 UTC" firstStartedPulling="2025-12-06 05:53:59.522836142 +0000 UTC m=+164.411819084" lastFinishedPulling="2025-12-06 05:56:00.302409126 +0000 UTC m=+285.191392068" observedRunningTime="2025-12-06 05:56:01.302619353 +0000 UTC m=+286.191602295" watchObservedRunningTime="2025-12-06 05:56:01.308973437 +0000 UTC m=+286.197956389" Dec 06 05:56:01 crc kubenswrapper[4809]: I1206 05:56:01.333194 4809 scope.go:117] "RemoveContainer" containerID="3162d1b7bddb463502601ccee5bfef8b3ccc2d1403ec4a883b60fe68fa80bbb1" Dec 06 05:56:01 crc kubenswrapper[4809]: I1206 05:56:01.394029 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9zm48" podStartSLOduration=8.240933287 podStartE2EDuration="2m10.394015107s" podCreationTimestamp="2025-12-06 05:53:51 +0000 UTC" firstStartedPulling="2025-12-06 05:53:58.400061744 +0000 UTC m=+163.289044696" lastFinishedPulling="2025-12-06 05:56:00.553143574 +0000 UTC m=+285.442126516" observedRunningTime="2025-12-06 05:56:01.390568072 +0000 UTC m=+286.279551014" watchObservedRunningTime="2025-12-06 05:56:01.394015107 +0000 UTC m=+286.282998049" Dec 06 05:56:01 crc kubenswrapper[4809]: I1206 05:56:01.406002 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ls5nx"] Dec 06 05:56:01 crc kubenswrapper[4809]: I1206 05:56:01.408684 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ls5nx"] Dec 06 05:56:01 crc kubenswrapper[4809]: I1206 05:56:01.806883 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-24qxt" Dec 06 05:56:01 crc kubenswrapper[4809]: I1206 05:56:01.807341 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-24qxt" Dec 06 05:56:02 crc kubenswrapper[4809]: I1206 05:56:02.276431 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9zm48" Dec 06 05:56:02 crc kubenswrapper[4809]: I1206 05:56:02.276473 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9zm48" Dec 06 05:56:02 crc kubenswrapper[4809]: I1206 05:56:02.848372 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-24qxt" podUID="893020b2-fa30-4371-86f9-a585517839ce" containerName="registry-server" probeResult="failure" output=< Dec 06 05:56:02 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Dec 06 05:56:02 crc kubenswrapper[4809]: > Dec 06 05:56:03 crc kubenswrapper[4809]: I1206 05:56:03.306264 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wj2h7" event={"ID":"2c4b2f56-0b03-41ae-af23-6192b8bd3d34","Type":"ContainerStarted","Data":"69983aab74d142a4d79a6b927b484e1faff5ba399abafd3c6e39c4f4c1805346"} Dec 06 05:56:03 crc kubenswrapper[4809]: I1206 05:56:03.309264 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dvwjq" event={"ID":"a83cbcc6-62ac-4814-9fce-eb7ea465bcbf","Type":"ContainerStarted","Data":"50137c90f10698b2438f0408ede009051a0a7d8ecb4527e96df10128ea127b02"} Dec 06 05:56:03 crc kubenswrapper[4809]: I1206 05:56:03.326695 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wj2h7" podStartSLOduration=7.616383526 podStartE2EDuration="2m10.326675225s" podCreationTimestamp="2025-12-06 05:53:53 +0000 UTC" firstStartedPulling="2025-12-06 05:53:59.522203274 +0000 UTC m=+164.411186216" lastFinishedPulling="2025-12-06 05:56:02.232494973 +0000 UTC m=+287.121477915" observedRunningTime="2025-12-06 05:56:03.325196464 +0000 UTC m=+288.214179406" watchObservedRunningTime="2025-12-06 05:56:03.326675225 +0000 UTC m=+288.215658167" Dec 06 05:56:03 crc kubenswrapper[4809]: I1206 05:56:03.343090 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-9zm48" podUID="c822a806-352f-4b4f-8549-051ff20c4ca2" containerName="registry-server" probeResult="failure" output=< Dec 06 05:56:03 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Dec 06 05:56:03 crc kubenswrapper[4809]: > Dec 06 05:56:03 crc kubenswrapper[4809]: I1206 05:56:03.343365 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dvwjq" podStartSLOduration=8.69100019 podStartE2EDuration="2m11.343348271s" podCreationTimestamp="2025-12-06 05:53:52 +0000 UTC" firstStartedPulling="2025-12-06 05:53:59.523181732 +0000 UTC m=+164.412164674" lastFinishedPulling="2025-12-06 05:56:02.175529813 +0000 UTC m=+287.064512755" observedRunningTime="2025-12-06 05:56:03.341262904 +0000 UTC m=+288.230245866" watchObservedRunningTime="2025-12-06 05:56:03.343348271 +0000 UTC m=+288.232331223" Dec 06 05:56:03 crc kubenswrapper[4809]: I1206 05:56:03.394271 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b475ab56-d949-44e5-a480-f59d8a1d737e" path="/var/lib/kubelet/pods/b475ab56-d949-44e5-a480-f59d8a1d737e/volumes" Dec 06 05:56:03 crc kubenswrapper[4809]: I1206 05:56:03.867124 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wj2h7" Dec 06 05:56:03 crc kubenswrapper[4809]: I1206 05:56:03.867195 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wj2h7" Dec 06 05:56:03 crc kubenswrapper[4809]: I1206 05:56:03.926433 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wj2h7" Dec 06 05:56:05 crc kubenswrapper[4809]: I1206 05:56:05.399542 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hfgk4" Dec 06 05:56:05 crc kubenswrapper[4809]: I1206 05:56:05.399867 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hfgk4" Dec 06 05:56:05 crc kubenswrapper[4809]: I1206 05:56:05.992503 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lpb28" Dec 06 05:56:06 crc kubenswrapper[4809]: I1206 05:56:06.031038 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lpb28" Dec 06 05:56:06 crc kubenswrapper[4809]: I1206 05:56:06.444890 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hfgk4" podUID="e356b822-4e99-42f8-99ba-7aea197f3d95" containerName="registry-server" probeResult="failure" output=< Dec 06 05:56:06 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Dec 06 05:56:06 crc kubenswrapper[4809]: > Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.276913 4809 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 06 05:56:11 crc kubenswrapper[4809]: E1206 05:56:11.277445 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b475ab56-d949-44e5-a480-f59d8a1d737e" containerName="extract-content" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.277456 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b475ab56-d949-44e5-a480-f59d8a1d737e" containerName="extract-content" Dec 06 05:56:11 crc kubenswrapper[4809]: E1206 05:56:11.277471 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d53c879-8535-4785-8743-3c272df4d0cc" containerName="pruner" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.277477 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d53c879-8535-4785-8743-3c272df4d0cc" containerName="pruner" Dec 06 05:56:11 crc kubenswrapper[4809]: E1206 05:56:11.277486 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b475ab56-d949-44e5-a480-f59d8a1d737e" containerName="extract-utilities" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.277492 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b475ab56-d949-44e5-a480-f59d8a1d737e" containerName="extract-utilities" Dec 06 05:56:11 crc kubenswrapper[4809]: E1206 05:56:11.277506 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b475ab56-d949-44e5-a480-f59d8a1d737e" containerName="registry-server" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.277512 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b475ab56-d949-44e5-a480-f59d8a1d737e" containerName="registry-server" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.277619 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d53c879-8535-4785-8743-3c272df4d0cc" containerName="pruner" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.277636 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b475ab56-d949-44e5-a480-f59d8a1d737e" containerName="registry-server" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.278025 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.281279 4809 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.281903 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d" gracePeriod=15 Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.281950 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36" gracePeriod=15 Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.282015 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad" gracePeriod=15 Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.282028 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48" gracePeriod=15 Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.281998 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f" gracePeriod=15 Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.283870 4809 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 06 05:56:11 crc kubenswrapper[4809]: E1206 05:56:11.284138 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.284161 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 05:56:11 crc kubenswrapper[4809]: E1206 05:56:11.284184 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.284192 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 06 05:56:11 crc kubenswrapper[4809]: E1206 05:56:11.284201 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.284208 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 06 05:56:11 crc kubenswrapper[4809]: E1206 05:56:11.284229 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.284241 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 06 05:56:11 crc kubenswrapper[4809]: E1206 05:56:11.284249 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.284256 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 06 05:56:11 crc kubenswrapper[4809]: E1206 05:56:11.284265 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.284271 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.284380 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.284392 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.284402 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.284412 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.284422 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.320535 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.459892 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.459971 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.460022 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.460057 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.460100 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.460153 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.460181 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.460265 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.561076 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.561137 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.561159 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.561185 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.561210 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.561226 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.561229 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.561274 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.561304 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.561330 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.561352 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.561332 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.561280 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.561373 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.561391 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.561415 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.614313 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 05:56:11 crc kubenswrapper[4809]: E1206 05:56:11.642308 4809 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.154:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187e8aab250b9d75 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-06 05:56:11.641372021 +0000 UTC m=+296.530354963,LastTimestamp:2025-12-06 05:56:11.641372021 +0000 UTC m=+296.530354963,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.872718 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-24qxt" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.874343 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.874712 4809 status_manager.go:851] "Failed to get status for pod" podUID="893020b2-fa30-4371-86f9-a585517839ce" pod="openshift-marketplace/community-operators-24qxt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-24qxt\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.920478 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-24qxt" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.921235 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:11 crc kubenswrapper[4809]: I1206 05:56:11.921891 4809 status_manager.go:851] "Failed to get status for pod" podUID="893020b2-fa30-4371-86f9-a585517839ce" pod="openshift-marketplace/community-operators-24qxt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-24qxt\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:12 crc kubenswrapper[4809]: I1206 05:56:12.322497 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9zm48" Dec 06 05:56:12 crc kubenswrapper[4809]: I1206 05:56:12.324238 4809 status_manager.go:851] "Failed to get status for pod" podUID="893020b2-fa30-4371-86f9-a585517839ce" pod="openshift-marketplace/community-operators-24qxt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-24qxt\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:12 crc kubenswrapper[4809]: I1206 05:56:12.324813 4809 status_manager.go:851] "Failed to get status for pod" podUID="c822a806-352f-4b4f-8549-051ff20c4ca2" pod="openshift-marketplace/community-operators-9zm48" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9zm48\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:12 crc kubenswrapper[4809]: I1206 05:56:12.325212 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:12 crc kubenswrapper[4809]: I1206 05:56:12.371303 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 06 05:56:12 crc kubenswrapper[4809]: I1206 05:56:12.372084 4809 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36" exitCode=0 Dec 06 05:56:12 crc kubenswrapper[4809]: I1206 05:56:12.372166 4809 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad" exitCode=0 Dec 06 05:56:12 crc kubenswrapper[4809]: I1206 05:56:12.372176 4809 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48" exitCode=0 Dec 06 05:56:12 crc kubenswrapper[4809]: I1206 05:56:12.372183 4809 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f" exitCode=2 Dec 06 05:56:12 crc kubenswrapper[4809]: I1206 05:56:12.373893 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"7eb71090d0b795b8d65e3853cd4fae206875ec874c6e70eddaaf8b431b7889fa"} Dec 06 05:56:12 crc kubenswrapper[4809]: I1206 05:56:12.374028 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"cbaec32b9ccccbbc4d2e4d5c1608fd6051b2d9474800abc6be07dd5f90cb0b01"} Dec 06 05:56:12 crc kubenswrapper[4809]: I1206 05:56:12.375915 4809 generic.go:334] "Generic (PLEG): container finished" podID="001a0f27-e6f3-4d62-9c25-43858654d2b6" containerID="2ec322a0f85b3bd7cc2e0b41458a3e30fc7803aec81398f0c39be0aa2b803b8c" exitCode=0 Dec 06 05:56:12 crc kubenswrapper[4809]: I1206 05:56:12.376949 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"001a0f27-e6f3-4d62-9c25-43858654d2b6","Type":"ContainerDied","Data":"2ec322a0f85b3bd7cc2e0b41458a3e30fc7803aec81398f0c39be0aa2b803b8c"} Dec 06 05:56:12 crc kubenswrapper[4809]: I1206 05:56:12.377620 4809 status_manager.go:851] "Failed to get status for pod" podUID="893020b2-fa30-4371-86f9-a585517839ce" pod="openshift-marketplace/community-operators-24qxt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-24qxt\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:12 crc kubenswrapper[4809]: I1206 05:56:12.377860 4809 status_manager.go:851] "Failed to get status for pod" podUID="c822a806-352f-4b4f-8549-051ff20c4ca2" pod="openshift-marketplace/community-operators-9zm48" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9zm48\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:12 crc kubenswrapper[4809]: I1206 05:56:12.378105 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:12 crc kubenswrapper[4809]: I1206 05:56:12.378329 4809 status_manager.go:851] "Failed to get status for pod" podUID="001a0f27-e6f3-4d62-9c25-43858654d2b6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:12 crc kubenswrapper[4809]: I1206 05:56:12.381235 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9zm48" Dec 06 05:56:12 crc kubenswrapper[4809]: I1206 05:56:12.382752 4809 status_manager.go:851] "Failed to get status for pod" podUID="893020b2-fa30-4371-86f9-a585517839ce" pod="openshift-marketplace/community-operators-24qxt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-24qxt\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:12 crc kubenswrapper[4809]: I1206 05:56:12.383166 4809 status_manager.go:851] "Failed to get status for pod" podUID="c822a806-352f-4b4f-8549-051ff20c4ca2" pod="openshift-marketplace/community-operators-9zm48" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9zm48\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:12 crc kubenswrapper[4809]: I1206 05:56:12.383741 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:12 crc kubenswrapper[4809]: I1206 05:56:12.384166 4809 status_manager.go:851] "Failed to get status for pod" podUID="001a0f27-e6f3-4d62-9c25-43858654d2b6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:12 crc kubenswrapper[4809]: I1206 05:56:12.495542 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dvwjq" Dec 06 05:56:12 crc kubenswrapper[4809]: I1206 05:56:12.495603 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dvwjq" Dec 06 05:56:12 crc kubenswrapper[4809]: I1206 05:56:12.534670 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dvwjq" Dec 06 05:56:12 crc kubenswrapper[4809]: I1206 05:56:12.535242 4809 status_manager.go:851] "Failed to get status for pod" podUID="001a0f27-e6f3-4d62-9c25-43858654d2b6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:12 crc kubenswrapper[4809]: I1206 05:56:12.535757 4809 status_manager.go:851] "Failed to get status for pod" podUID="893020b2-fa30-4371-86f9-a585517839ce" pod="openshift-marketplace/community-operators-24qxt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-24qxt\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:12 crc kubenswrapper[4809]: I1206 05:56:12.536271 4809 status_manager.go:851] "Failed to get status for pod" podUID="a83cbcc6-62ac-4814-9fce-eb7ea465bcbf" pod="openshift-marketplace/certified-operators-dvwjq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-dvwjq\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:12 crc kubenswrapper[4809]: I1206 05:56:12.536575 4809 status_manager.go:851] "Failed to get status for pod" podUID="c822a806-352f-4b4f-8549-051ff20c4ca2" pod="openshift-marketplace/community-operators-9zm48" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9zm48\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:12 crc kubenswrapper[4809]: I1206 05:56:12.536834 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.383380 4809 status_manager.go:851] "Failed to get status for pod" podUID="001a0f27-e6f3-4d62-9c25-43858654d2b6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.383681 4809 status_manager.go:851] "Failed to get status for pod" podUID="893020b2-fa30-4371-86f9-a585517839ce" pod="openshift-marketplace/community-operators-24qxt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-24qxt\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.383905 4809 status_manager.go:851] "Failed to get status for pod" podUID="a83cbcc6-62ac-4814-9fce-eb7ea465bcbf" pod="openshift-marketplace/certified-operators-dvwjq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-dvwjq\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.397855 4809 status_manager.go:851] "Failed to get status for pod" podUID="c822a806-352f-4b4f-8549-051ff20c4ca2" pod="openshift-marketplace/community-operators-9zm48" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9zm48\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.398556 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.466475 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dvwjq" Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.469192 4809 status_manager.go:851] "Failed to get status for pod" podUID="893020b2-fa30-4371-86f9-a585517839ce" pod="openshift-marketplace/community-operators-24qxt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-24qxt\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.469740 4809 status_manager.go:851] "Failed to get status for pod" podUID="a83cbcc6-62ac-4814-9fce-eb7ea465bcbf" pod="openshift-marketplace/certified-operators-dvwjq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-dvwjq\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.471679 4809 status_manager.go:851] "Failed to get status for pod" podUID="c822a806-352f-4b4f-8549-051ff20c4ca2" pod="openshift-marketplace/community-operators-9zm48" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9zm48\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.472280 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.472703 4809 status_manager.go:851] "Failed to get status for pod" podUID="001a0f27-e6f3-4d62-9c25-43858654d2b6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:13 crc kubenswrapper[4809]: E1206 05:56:13.612080 4809 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.154:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187e8aab250b9d75 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-06 05:56:11.641372021 +0000 UTC m=+296.530354963,LastTimestamp:2025-12-06 05:56:11.641372021 +0000 UTC m=+296.530354963,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.676563 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.677322 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.677794 4809 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.678116 4809 status_manager.go:851] "Failed to get status for pod" podUID="c822a806-352f-4b4f-8549-051ff20c4ca2" pod="openshift-marketplace/community-operators-9zm48" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9zm48\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.678494 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.678790 4809 status_manager.go:851] "Failed to get status for pod" podUID="001a0f27-e6f3-4d62-9c25-43858654d2b6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.679037 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.679517 4809 status_manager.go:851] "Failed to get status for pod" podUID="893020b2-fa30-4371-86f9-a585517839ce" pod="openshift-marketplace/community-operators-24qxt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-24qxt\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.679765 4809 status_manager.go:851] "Failed to get status for pod" podUID="a83cbcc6-62ac-4814-9fce-eb7ea465bcbf" pod="openshift-marketplace/certified-operators-dvwjq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-dvwjq\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.680159 4809 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.680410 4809 status_manager.go:851] "Failed to get status for pod" podUID="c822a806-352f-4b4f-8549-051ff20c4ca2" pod="openshift-marketplace/community-operators-9zm48" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9zm48\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.680680 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.680964 4809 status_manager.go:851] "Failed to get status for pod" podUID="001a0f27-e6f3-4d62-9c25-43858654d2b6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.681220 4809 status_manager.go:851] "Failed to get status for pod" podUID="893020b2-fa30-4371-86f9-a585517839ce" pod="openshift-marketplace/community-operators-24qxt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-24qxt\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.681451 4809 status_manager.go:851] "Failed to get status for pod" podUID="a83cbcc6-62ac-4814-9fce-eb7ea465bcbf" pod="openshift-marketplace/certified-operators-dvwjq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-dvwjq\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.790575 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/001a0f27-e6f3-4d62-9c25-43858654d2b6-var-lock\") pod \"001a0f27-e6f3-4d62-9c25-43858654d2b6\" (UID: \"001a0f27-e6f3-4d62-9c25-43858654d2b6\") " Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.790631 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.790665 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.790681 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/001a0f27-e6f3-4d62-9c25-43858654d2b6-var-lock" (OuterVolumeSpecName: "var-lock") pod "001a0f27-e6f3-4d62-9c25-43858654d2b6" (UID: "001a0f27-e6f3-4d62-9c25-43858654d2b6"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.790708 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/001a0f27-e6f3-4d62-9c25-43858654d2b6-kube-api-access\") pod \"001a0f27-e6f3-4d62-9c25-43858654d2b6\" (UID: \"001a0f27-e6f3-4d62-9c25-43858654d2b6\") " Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.790719 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.790764 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.790804 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.790788 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.790880 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/001a0f27-e6f3-4d62-9c25-43858654d2b6-kubelet-dir\") pod \"001a0f27-e6f3-4d62-9c25-43858654d2b6\" (UID: \"001a0f27-e6f3-4d62-9c25-43858654d2b6\") " Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.790905 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/001a0f27-e6f3-4d62-9c25-43858654d2b6-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "001a0f27-e6f3-4d62-9c25-43858654d2b6" (UID: "001a0f27-e6f3-4d62-9c25-43858654d2b6"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.791080 4809 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.791091 4809 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.791098 4809 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/001a0f27-e6f3-4d62-9c25-43858654d2b6-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.791106 4809 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/001a0f27-e6f3-4d62-9c25-43858654d2b6-var-lock\") on node \"crc\" DevicePath \"\"" Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.791114 4809 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.795178 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/001a0f27-e6f3-4d62-9c25-43858654d2b6-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "001a0f27-e6f3-4d62-9c25-43858654d2b6" (UID: "001a0f27-e6f3-4d62-9c25-43858654d2b6"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.893457 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/001a0f27-e6f3-4d62-9c25-43858654d2b6-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.912470 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wj2h7" Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.913110 4809 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.913348 4809 status_manager.go:851] "Failed to get status for pod" podUID="c822a806-352f-4b4f-8549-051ff20c4ca2" pod="openshift-marketplace/community-operators-9zm48" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9zm48\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.913560 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.913789 4809 status_manager.go:851] "Failed to get status for pod" podUID="001a0f27-e6f3-4d62-9c25-43858654d2b6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.914022 4809 status_manager.go:851] "Failed to get status for pod" podUID="893020b2-fa30-4371-86f9-a585517839ce" pod="openshift-marketplace/community-operators-24qxt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-24qxt\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.914230 4809 status_manager.go:851] "Failed to get status for pod" podUID="a83cbcc6-62ac-4814-9fce-eb7ea465bcbf" pod="openshift-marketplace/certified-operators-dvwjq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-dvwjq\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:13 crc kubenswrapper[4809]: I1206 05:56:13.914421 4809 status_manager.go:851] "Failed to get status for pod" podUID="2c4b2f56-0b03-41ae-af23-6192b8bd3d34" pod="openshift-marketplace/redhat-marketplace-wj2h7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-wj2h7\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:14 crc kubenswrapper[4809]: I1206 05:56:14.393034 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 06 05:56:14 crc kubenswrapper[4809]: I1206 05:56:14.397546 4809 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d" exitCode=0 Dec 06 05:56:14 crc kubenswrapper[4809]: I1206 05:56:14.397737 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:56:14 crc kubenswrapper[4809]: I1206 05:56:14.397949 4809 scope.go:117] "RemoveContainer" containerID="5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36" Dec 06 05:56:14 crc kubenswrapper[4809]: I1206 05:56:14.400135 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"001a0f27-e6f3-4d62-9c25-43858654d2b6","Type":"ContainerDied","Data":"875ca06589b2bb882a691d83ec8796748443f3cff33e3a2eea843c9736908ff1"} Dec 06 05:56:14 crc kubenswrapper[4809]: I1206 05:56:14.400201 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="875ca06589b2bb882a691d83ec8796748443f3cff33e3a2eea843c9736908ff1" Dec 06 05:56:14 crc kubenswrapper[4809]: I1206 05:56:14.400367 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 06 05:56:14 crc kubenswrapper[4809]: I1206 05:56:14.412824 4809 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:14 crc kubenswrapper[4809]: I1206 05:56:14.414115 4809 status_manager.go:851] "Failed to get status for pod" podUID="c822a806-352f-4b4f-8549-051ff20c4ca2" pod="openshift-marketplace/community-operators-9zm48" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9zm48\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:14 crc kubenswrapper[4809]: I1206 05:56:14.414335 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:14 crc kubenswrapper[4809]: I1206 05:56:14.414600 4809 status_manager.go:851] "Failed to get status for pod" podUID="001a0f27-e6f3-4d62-9c25-43858654d2b6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:14 crc kubenswrapper[4809]: I1206 05:56:14.414880 4809 status_manager.go:851] "Failed to get status for pod" podUID="2c4b2f56-0b03-41ae-af23-6192b8bd3d34" pod="openshift-marketplace/redhat-marketplace-wj2h7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-wj2h7\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:14 crc kubenswrapper[4809]: I1206 05:56:14.415119 4809 status_manager.go:851] "Failed to get status for pod" podUID="893020b2-fa30-4371-86f9-a585517839ce" pod="openshift-marketplace/community-operators-24qxt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-24qxt\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:14 crc kubenswrapper[4809]: I1206 05:56:14.415314 4809 status_manager.go:851] "Failed to get status for pod" podUID="a83cbcc6-62ac-4814-9fce-eb7ea465bcbf" pod="openshift-marketplace/certified-operators-dvwjq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-dvwjq\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:14 crc kubenswrapper[4809]: I1206 05:56:14.419439 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:14 crc kubenswrapper[4809]: I1206 05:56:14.419771 4809 status_manager.go:851] "Failed to get status for pod" podUID="001a0f27-e6f3-4d62-9c25-43858654d2b6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:14 crc kubenswrapper[4809]: I1206 05:56:14.420065 4809 status_manager.go:851] "Failed to get status for pod" podUID="2c4b2f56-0b03-41ae-af23-6192b8bd3d34" pod="openshift-marketplace/redhat-marketplace-wj2h7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-wj2h7\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:14 crc kubenswrapper[4809]: I1206 05:56:14.420335 4809 status_manager.go:851] "Failed to get status for pod" podUID="893020b2-fa30-4371-86f9-a585517839ce" pod="openshift-marketplace/community-operators-24qxt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-24qxt\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:14 crc kubenswrapper[4809]: I1206 05:56:14.421158 4809 status_manager.go:851] "Failed to get status for pod" podUID="a83cbcc6-62ac-4814-9fce-eb7ea465bcbf" pod="openshift-marketplace/certified-operators-dvwjq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-dvwjq\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:14 crc kubenswrapper[4809]: I1206 05:56:14.421511 4809 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:14 crc kubenswrapper[4809]: I1206 05:56:14.421776 4809 status_manager.go:851] "Failed to get status for pod" podUID="c822a806-352f-4b4f-8549-051ff20c4ca2" pod="openshift-marketplace/community-operators-9zm48" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9zm48\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:14 crc kubenswrapper[4809]: I1206 05:56:14.424375 4809 scope.go:117] "RemoveContainer" containerID="eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad" Dec 06 05:56:14 crc kubenswrapper[4809]: I1206 05:56:14.443636 4809 scope.go:117] "RemoveContainer" containerID="332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48" Dec 06 05:56:14 crc kubenswrapper[4809]: I1206 05:56:14.458124 4809 scope.go:117] "RemoveContainer" containerID="74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f" Dec 06 05:56:14 crc kubenswrapper[4809]: I1206 05:56:14.472060 4809 scope.go:117] "RemoveContainer" containerID="b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d" Dec 06 05:56:14 crc kubenswrapper[4809]: I1206 05:56:14.494055 4809 scope.go:117] "RemoveContainer" containerID="30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7" Dec 06 05:56:14 crc kubenswrapper[4809]: I1206 05:56:14.513035 4809 scope.go:117] "RemoveContainer" containerID="5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36" Dec 06 05:56:14 crc kubenswrapper[4809]: E1206 05:56:14.514608 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36\": container with ID starting with 5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36 not found: ID does not exist" containerID="5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36" Dec 06 05:56:14 crc kubenswrapper[4809]: I1206 05:56:14.514650 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36"} err="failed to get container status \"5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36\": rpc error: code = NotFound desc = could not find container \"5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36\": container with ID starting with 5daa0e07a3537c03efc296d00d3e779c50b27ee3de1c537dc5e0c9068c96ea36 not found: ID does not exist" Dec 06 05:56:14 crc kubenswrapper[4809]: I1206 05:56:14.514676 4809 scope.go:117] "RemoveContainer" containerID="eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad" Dec 06 05:56:14 crc kubenswrapper[4809]: E1206 05:56:14.515088 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad\": container with ID starting with eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad not found: ID does not exist" containerID="eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad" Dec 06 05:56:14 crc kubenswrapper[4809]: I1206 05:56:14.515128 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad"} err="failed to get container status \"eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad\": rpc error: code = NotFound desc = could not find container \"eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad\": container with ID starting with eab9743b782a69be89866c195f9219a71eae0fc1585979da033e15de2545ddad not found: ID does not exist" Dec 06 05:56:14 crc kubenswrapper[4809]: I1206 05:56:14.515155 4809 scope.go:117] "RemoveContainer" containerID="332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48" Dec 06 05:56:14 crc kubenswrapper[4809]: E1206 05:56:14.515436 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48\": container with ID starting with 332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48 not found: ID does not exist" containerID="332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48" Dec 06 05:56:14 crc kubenswrapper[4809]: I1206 05:56:14.515463 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48"} err="failed to get container status \"332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48\": rpc error: code = NotFound desc = could not find container \"332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48\": container with ID starting with 332cbbb677b3fe2715f4ce829efca34d6c28aed7f12970ba0d03b3602dcdbb48 not found: ID does not exist" Dec 06 05:56:14 crc kubenswrapper[4809]: I1206 05:56:14.515476 4809 scope.go:117] "RemoveContainer" containerID="74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f" Dec 06 05:56:14 crc kubenswrapper[4809]: E1206 05:56:14.515718 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f\": container with ID starting with 74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f not found: ID does not exist" containerID="74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f" Dec 06 05:56:14 crc kubenswrapper[4809]: I1206 05:56:14.515759 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f"} err="failed to get container status \"74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f\": rpc error: code = NotFound desc = could not find container \"74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f\": container with ID starting with 74d46465e3d6f8a427411d523538bd6a751554db1ac75a7415741bb9c78a446f not found: ID does not exist" Dec 06 05:56:14 crc kubenswrapper[4809]: I1206 05:56:14.516517 4809 scope.go:117] "RemoveContainer" containerID="b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d" Dec 06 05:56:14 crc kubenswrapper[4809]: E1206 05:56:14.516815 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d\": container with ID starting with b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d not found: ID does not exist" containerID="b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d" Dec 06 05:56:14 crc kubenswrapper[4809]: I1206 05:56:14.516843 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d"} err="failed to get container status \"b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d\": rpc error: code = NotFound desc = could not find container \"b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d\": container with ID starting with b749f15269caf6977fbdbcab2b19d5718556c6c5d50c237bcc97f2df167d224d not found: ID does not exist" Dec 06 05:56:14 crc kubenswrapper[4809]: I1206 05:56:14.516858 4809 scope.go:117] "RemoveContainer" containerID="30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7" Dec 06 05:56:14 crc kubenswrapper[4809]: E1206 05:56:14.517088 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\": container with ID starting with 30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7 not found: ID does not exist" containerID="30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7" Dec 06 05:56:14 crc kubenswrapper[4809]: I1206 05:56:14.517127 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7"} err="failed to get container status \"30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\": rpc error: code = NotFound desc = could not find container \"30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7\": container with ID starting with 30847224b4b3dcd350a06a1420c4b62f6cc214220de562e866c8d3cb40064dd7 not found: ID does not exist" Dec 06 05:56:15 crc kubenswrapper[4809]: I1206 05:56:15.397333 4809 status_manager.go:851] "Failed to get status for pod" podUID="893020b2-fa30-4371-86f9-a585517839ce" pod="openshift-marketplace/community-operators-24qxt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-24qxt\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:15 crc kubenswrapper[4809]: I1206 05:56:15.397698 4809 status_manager.go:851] "Failed to get status for pod" podUID="a83cbcc6-62ac-4814-9fce-eb7ea465bcbf" pod="openshift-marketplace/certified-operators-dvwjq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-dvwjq\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:15 crc kubenswrapper[4809]: I1206 05:56:15.398029 4809 status_manager.go:851] "Failed to get status for pod" podUID="2c4b2f56-0b03-41ae-af23-6192b8bd3d34" pod="openshift-marketplace/redhat-marketplace-wj2h7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-wj2h7\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:15 crc kubenswrapper[4809]: I1206 05:56:15.398413 4809 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:15 crc kubenswrapper[4809]: I1206 05:56:15.398958 4809 status_manager.go:851] "Failed to get status for pod" podUID="c822a806-352f-4b4f-8549-051ff20c4ca2" pod="openshift-marketplace/community-operators-9zm48" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9zm48\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:15 crc kubenswrapper[4809]: I1206 05:56:15.399223 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:15 crc kubenswrapper[4809]: I1206 05:56:15.399576 4809 status_manager.go:851] "Failed to get status for pod" podUID="001a0f27-e6f3-4d62-9c25-43858654d2b6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:15 crc kubenswrapper[4809]: I1206 05:56:15.408411 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 06 05:56:15 crc kubenswrapper[4809]: I1206 05:56:15.438680 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hfgk4" Dec 06 05:56:15 crc kubenswrapper[4809]: I1206 05:56:15.439619 4809 status_manager.go:851] "Failed to get status for pod" podUID="c822a806-352f-4b4f-8549-051ff20c4ca2" pod="openshift-marketplace/community-operators-9zm48" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9zm48\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:15 crc kubenswrapper[4809]: I1206 05:56:15.439977 4809 status_manager.go:851] "Failed to get status for pod" podUID="e356b822-4e99-42f8-99ba-7aea197f3d95" pod="openshift-marketplace/redhat-operators-hfgk4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hfgk4\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:15 crc kubenswrapper[4809]: I1206 05:56:15.440330 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:15 crc kubenswrapper[4809]: I1206 05:56:15.440705 4809 status_manager.go:851] "Failed to get status for pod" podUID="001a0f27-e6f3-4d62-9c25-43858654d2b6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:15 crc kubenswrapper[4809]: I1206 05:56:15.441146 4809 status_manager.go:851] "Failed to get status for pod" podUID="893020b2-fa30-4371-86f9-a585517839ce" pod="openshift-marketplace/community-operators-24qxt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-24qxt\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:15 crc kubenswrapper[4809]: I1206 05:56:15.441443 4809 status_manager.go:851] "Failed to get status for pod" podUID="a83cbcc6-62ac-4814-9fce-eb7ea465bcbf" pod="openshift-marketplace/certified-operators-dvwjq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-dvwjq\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:15 crc kubenswrapper[4809]: I1206 05:56:15.441775 4809 status_manager.go:851] "Failed to get status for pod" podUID="2c4b2f56-0b03-41ae-af23-6192b8bd3d34" pod="openshift-marketplace/redhat-marketplace-wj2h7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-wj2h7\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:15 crc kubenswrapper[4809]: I1206 05:56:15.478629 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hfgk4" Dec 06 05:56:15 crc kubenswrapper[4809]: I1206 05:56:15.479299 4809 status_manager.go:851] "Failed to get status for pod" podUID="c822a806-352f-4b4f-8549-051ff20c4ca2" pod="openshift-marketplace/community-operators-9zm48" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9zm48\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:15 crc kubenswrapper[4809]: I1206 05:56:15.479778 4809 status_manager.go:851] "Failed to get status for pod" podUID="e356b822-4e99-42f8-99ba-7aea197f3d95" pod="openshift-marketplace/redhat-operators-hfgk4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hfgk4\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:15 crc kubenswrapper[4809]: I1206 05:56:15.480270 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:15 crc kubenswrapper[4809]: I1206 05:56:15.480574 4809 status_manager.go:851] "Failed to get status for pod" podUID="001a0f27-e6f3-4d62-9c25-43858654d2b6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:15 crc kubenswrapper[4809]: I1206 05:56:15.480869 4809 status_manager.go:851] "Failed to get status for pod" podUID="893020b2-fa30-4371-86f9-a585517839ce" pod="openshift-marketplace/community-operators-24qxt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-24qxt\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:15 crc kubenswrapper[4809]: I1206 05:56:15.481229 4809 status_manager.go:851] "Failed to get status for pod" podUID="a83cbcc6-62ac-4814-9fce-eb7ea465bcbf" pod="openshift-marketplace/certified-operators-dvwjq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-dvwjq\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:15 crc kubenswrapper[4809]: I1206 05:56:15.481522 4809 status_manager.go:851] "Failed to get status for pod" podUID="2c4b2f56-0b03-41ae-af23-6192b8bd3d34" pod="openshift-marketplace/redhat-marketplace-wj2h7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-wj2h7\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:19 crc kubenswrapper[4809]: I1206 05:56:19.702349 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-jn446" podUID="0a9baebe-98e6-44be-ad16-882e47f09248" containerName="oauth-openshift" containerID="cri-o://65e9f6726d60aa3c96833be77b3799ee27896dbd41ffe685acd2f6a01d1fc947" gracePeriod=15 Dec 06 05:56:20 crc kubenswrapper[4809]: E1206 05:56:20.254075 4809 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:20 crc kubenswrapper[4809]: E1206 05:56:20.254772 4809 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:20 crc kubenswrapper[4809]: E1206 05:56:20.255274 4809 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:20 crc kubenswrapper[4809]: E1206 05:56:20.255579 4809 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:20 crc kubenswrapper[4809]: E1206 05:56:20.255869 4809 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:20 crc kubenswrapper[4809]: I1206 05:56:20.255899 4809 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 06 05:56:20 crc kubenswrapper[4809]: E1206 05:56:20.256104 4809 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.154:6443: connect: connection refused" interval="200ms" Dec 06 05:56:20 crc kubenswrapper[4809]: E1206 05:56:20.457250 4809 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.154:6443: connect: connection refused" interval="400ms" Dec 06 05:56:20 crc kubenswrapper[4809]: E1206 05:56:20.858402 4809 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.154:6443: connect: connection refused" interval="800ms" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.190264 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.191266 4809 status_manager.go:851] "Failed to get status for pod" podUID="c822a806-352f-4b4f-8549-051ff20c4ca2" pod="openshift-marketplace/community-operators-9zm48" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9zm48\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.191780 4809 status_manager.go:851] "Failed to get status for pod" podUID="e356b822-4e99-42f8-99ba-7aea197f3d95" pod="openshift-marketplace/redhat-operators-hfgk4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hfgk4\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.192382 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.192899 4809 status_manager.go:851] "Failed to get status for pod" podUID="001a0f27-e6f3-4d62-9c25-43858654d2b6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.193241 4809 status_manager.go:851] "Failed to get status for pod" podUID="893020b2-fa30-4371-86f9-a585517839ce" pod="openshift-marketplace/community-operators-24qxt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-24qxt\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.193749 4809 status_manager.go:851] "Failed to get status for pod" podUID="a83cbcc6-62ac-4814-9fce-eb7ea465bcbf" pod="openshift-marketplace/certified-operators-dvwjq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-dvwjq\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.194088 4809 status_manager.go:851] "Failed to get status for pod" podUID="2c4b2f56-0b03-41ae-af23-6192b8bd3d34" pod="openshift-marketplace/redhat-marketplace-wj2h7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-wj2h7\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.194364 4809 status_manager.go:851] "Failed to get status for pod" podUID="0a9baebe-98e6-44be-ad16-882e47f09248" pod="openshift-authentication/oauth-openshift-558db77b4-jn446" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-jn446\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.379562 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-system-serving-cert\") pod \"0a9baebe-98e6-44be-ad16-882e47f09248\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.379618 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-user-template-provider-selection\") pod \"0a9baebe-98e6-44be-ad16-882e47f09248\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.379649 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-system-cliconfig\") pod \"0a9baebe-98e6-44be-ad16-882e47f09248\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.379687 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nczzr\" (UniqueName: \"kubernetes.io/projected/0a9baebe-98e6-44be-ad16-882e47f09248-kube-api-access-nczzr\") pod \"0a9baebe-98e6-44be-ad16-882e47f09248\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.379713 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-user-template-login\") pod \"0a9baebe-98e6-44be-ad16-882e47f09248\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.379740 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-user-idp-0-file-data\") pod \"0a9baebe-98e6-44be-ad16-882e47f09248\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.380560 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "0a9baebe-98e6-44be-ad16-882e47f09248" (UID: "0a9baebe-98e6-44be-ad16-882e47f09248"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.380837 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-user-template-error\") pod \"0a9baebe-98e6-44be-ad16-882e47f09248\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.380880 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-system-trusted-ca-bundle\") pod \"0a9baebe-98e6-44be-ad16-882e47f09248\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.380917 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-system-service-ca\") pod \"0a9baebe-98e6-44be-ad16-882e47f09248\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.380966 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-system-router-certs\") pod \"0a9baebe-98e6-44be-ad16-882e47f09248\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.381030 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-system-ocp-branding-template\") pod \"0a9baebe-98e6-44be-ad16-882e47f09248\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.381062 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0a9baebe-98e6-44be-ad16-882e47f09248-audit-dir\") pod \"0a9baebe-98e6-44be-ad16-882e47f09248\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.381089 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0a9baebe-98e6-44be-ad16-882e47f09248-audit-policies\") pod \"0a9baebe-98e6-44be-ad16-882e47f09248\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.381118 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-system-session\") pod \"0a9baebe-98e6-44be-ad16-882e47f09248\" (UID: \"0a9baebe-98e6-44be-ad16-882e47f09248\") " Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.381366 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.381434 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "0a9baebe-98e6-44be-ad16-882e47f09248" (UID: "0a9baebe-98e6-44be-ad16-882e47f09248"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.381453 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "0a9baebe-98e6-44be-ad16-882e47f09248" (UID: "0a9baebe-98e6-44be-ad16-882e47f09248"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.381556 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0a9baebe-98e6-44be-ad16-882e47f09248-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "0a9baebe-98e6-44be-ad16-882e47f09248" (UID: "0a9baebe-98e6-44be-ad16-882e47f09248"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.381984 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a9baebe-98e6-44be-ad16-882e47f09248-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "0a9baebe-98e6-44be-ad16-882e47f09248" (UID: "0a9baebe-98e6-44be-ad16-882e47f09248"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.385781 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "0a9baebe-98e6-44be-ad16-882e47f09248" (UID: "0a9baebe-98e6-44be-ad16-882e47f09248"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.386271 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "0a9baebe-98e6-44be-ad16-882e47f09248" (UID: "0a9baebe-98e6-44be-ad16-882e47f09248"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.386426 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a9baebe-98e6-44be-ad16-882e47f09248-kube-api-access-nczzr" (OuterVolumeSpecName: "kube-api-access-nczzr") pod "0a9baebe-98e6-44be-ad16-882e47f09248" (UID: "0a9baebe-98e6-44be-ad16-882e47f09248"). InnerVolumeSpecName "kube-api-access-nczzr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.386515 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "0a9baebe-98e6-44be-ad16-882e47f09248" (UID: "0a9baebe-98e6-44be-ad16-882e47f09248"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.387165 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "0a9baebe-98e6-44be-ad16-882e47f09248" (UID: "0a9baebe-98e6-44be-ad16-882e47f09248"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.387981 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "0a9baebe-98e6-44be-ad16-882e47f09248" (UID: "0a9baebe-98e6-44be-ad16-882e47f09248"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.390376 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "0a9baebe-98e6-44be-ad16-882e47f09248" (UID: "0a9baebe-98e6-44be-ad16-882e47f09248"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.390670 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "0a9baebe-98e6-44be-ad16-882e47f09248" (UID: "0a9baebe-98e6-44be-ad16-882e47f09248"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.390806 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "0a9baebe-98e6-44be-ad16-882e47f09248" (UID: "0a9baebe-98e6-44be-ad16-882e47f09248"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.444521 4809 generic.go:334] "Generic (PLEG): container finished" podID="0a9baebe-98e6-44be-ad16-882e47f09248" containerID="65e9f6726d60aa3c96833be77b3799ee27896dbd41ffe685acd2f6a01d1fc947" exitCode=0 Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.444575 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-jn446" event={"ID":"0a9baebe-98e6-44be-ad16-882e47f09248","Type":"ContainerDied","Data":"65e9f6726d60aa3c96833be77b3799ee27896dbd41ffe685acd2f6a01d1fc947"} Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.444612 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-jn446" event={"ID":"0a9baebe-98e6-44be-ad16-882e47f09248","Type":"ContainerDied","Data":"dad7c50235935b49033383b51383d232bb7c41a9c13c3c777faba482031af985"} Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.444632 4809 scope.go:117] "RemoveContainer" containerID="65e9f6726d60aa3c96833be77b3799ee27896dbd41ffe685acd2f6a01d1fc947" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.444667 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-jn446" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.452497 4809 status_manager.go:851] "Failed to get status for pod" podUID="e356b822-4e99-42f8-99ba-7aea197f3d95" pod="openshift-marketplace/redhat-operators-hfgk4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hfgk4\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.452718 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.453098 4809 status_manager.go:851] "Failed to get status for pod" podUID="001a0f27-e6f3-4d62-9c25-43858654d2b6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.453564 4809 status_manager.go:851] "Failed to get status for pod" podUID="893020b2-fa30-4371-86f9-a585517839ce" pod="openshift-marketplace/community-operators-24qxt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-24qxt\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.453888 4809 status_manager.go:851] "Failed to get status for pod" podUID="a83cbcc6-62ac-4814-9fce-eb7ea465bcbf" pod="openshift-marketplace/certified-operators-dvwjq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-dvwjq\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.454665 4809 status_manager.go:851] "Failed to get status for pod" podUID="2c4b2f56-0b03-41ae-af23-6192b8bd3d34" pod="openshift-marketplace/redhat-marketplace-wj2h7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-wj2h7\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.454908 4809 status_manager.go:851] "Failed to get status for pod" podUID="0a9baebe-98e6-44be-ad16-882e47f09248" pod="openshift-authentication/oauth-openshift-558db77b4-jn446" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-jn446\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.455123 4809 status_manager.go:851] "Failed to get status for pod" podUID="c822a806-352f-4b4f-8549-051ff20c4ca2" pod="openshift-marketplace/community-operators-9zm48" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9zm48\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.456013 4809 status_manager.go:851] "Failed to get status for pod" podUID="001a0f27-e6f3-4d62-9c25-43858654d2b6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.456407 4809 status_manager.go:851] "Failed to get status for pod" podUID="2c4b2f56-0b03-41ae-af23-6192b8bd3d34" pod="openshift-marketplace/redhat-marketplace-wj2h7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-wj2h7\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.456831 4809 status_manager.go:851] "Failed to get status for pod" podUID="893020b2-fa30-4371-86f9-a585517839ce" pod="openshift-marketplace/community-operators-24qxt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-24qxt\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.457105 4809 status_manager.go:851] "Failed to get status for pod" podUID="a83cbcc6-62ac-4814-9fce-eb7ea465bcbf" pod="openshift-marketplace/certified-operators-dvwjq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-dvwjq\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.457329 4809 status_manager.go:851] "Failed to get status for pod" podUID="0a9baebe-98e6-44be-ad16-882e47f09248" pod="openshift-authentication/oauth-openshift-558db77b4-jn446" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-jn446\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.457588 4809 status_manager.go:851] "Failed to get status for pod" podUID="c822a806-352f-4b4f-8549-051ff20c4ca2" pod="openshift-marketplace/community-operators-9zm48" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9zm48\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.457874 4809 status_manager.go:851] "Failed to get status for pod" podUID="e356b822-4e99-42f8-99ba-7aea197f3d95" pod="openshift-marketplace/redhat-operators-hfgk4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hfgk4\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.458455 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.467985 4809 scope.go:117] "RemoveContainer" containerID="65e9f6726d60aa3c96833be77b3799ee27896dbd41ffe685acd2f6a01d1fc947" Dec 06 05:56:21 crc kubenswrapper[4809]: E1206 05:56:21.468469 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65e9f6726d60aa3c96833be77b3799ee27896dbd41ffe685acd2f6a01d1fc947\": container with ID starting with 65e9f6726d60aa3c96833be77b3799ee27896dbd41ffe685acd2f6a01d1fc947 not found: ID does not exist" containerID="65e9f6726d60aa3c96833be77b3799ee27896dbd41ffe685acd2f6a01d1fc947" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.468509 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65e9f6726d60aa3c96833be77b3799ee27896dbd41ffe685acd2f6a01d1fc947"} err="failed to get container status \"65e9f6726d60aa3c96833be77b3799ee27896dbd41ffe685acd2f6a01d1fc947\": rpc error: code = NotFound desc = could not find container \"65e9f6726d60aa3c96833be77b3799ee27896dbd41ffe685acd2f6a01d1fc947\": container with ID starting with 65e9f6726d60aa3c96833be77b3799ee27896dbd41ffe685acd2f6a01d1fc947 not found: ID does not exist" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.483465 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nczzr\" (UniqueName: \"kubernetes.io/projected/0a9baebe-98e6-44be-ad16-882e47f09248-kube-api-access-nczzr\") on node \"crc\" DevicePath \"\"" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.483510 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.483531 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.483551 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.483569 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.483588 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.483606 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.483626 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.483647 4809 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0a9baebe-98e6-44be-ad16-882e47f09248-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.483664 4809 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0a9baebe-98e6-44be-ad16-882e47f09248-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.483683 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.483700 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:56:21 crc kubenswrapper[4809]: I1206 05:56:21.483719 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0a9baebe-98e6-44be-ad16-882e47f09248-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 06 05:56:21 crc kubenswrapper[4809]: E1206 05:56:21.659295 4809 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.154:6443: connect: connection refused" interval="1.6s" Dec 06 05:56:23 crc kubenswrapper[4809]: E1206 05:56:23.259741 4809 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.154:6443: connect: connection refused" interval="3.2s" Dec 06 05:56:23 crc kubenswrapper[4809]: E1206 05:56:23.613961 4809 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.154:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187e8aab250b9d75 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-06 05:56:11.641372021 +0000 UTC m=+296.530354963,LastTimestamp:2025-12-06 05:56:11.641372021 +0000 UTC m=+296.530354963,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 06 05:56:25 crc kubenswrapper[4809]: I1206 05:56:25.395126 4809 status_manager.go:851] "Failed to get status for pod" podUID="a83cbcc6-62ac-4814-9fce-eb7ea465bcbf" pod="openshift-marketplace/certified-operators-dvwjq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-dvwjq\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:25 crc kubenswrapper[4809]: I1206 05:56:25.395272 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:56:25 crc kubenswrapper[4809]: I1206 05:56:25.396901 4809 status_manager.go:851] "Failed to get status for pod" podUID="2c4b2f56-0b03-41ae-af23-6192b8bd3d34" pod="openshift-marketplace/redhat-marketplace-wj2h7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-wj2h7\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:25 crc kubenswrapper[4809]: I1206 05:56:25.397360 4809 status_manager.go:851] "Failed to get status for pod" podUID="893020b2-fa30-4371-86f9-a585517839ce" pod="openshift-marketplace/community-operators-24qxt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-24qxt\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:25 crc kubenswrapper[4809]: I1206 05:56:25.398029 4809 status_manager.go:851] "Failed to get status for pod" podUID="0a9baebe-98e6-44be-ad16-882e47f09248" pod="openshift-authentication/oauth-openshift-558db77b4-jn446" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-jn446\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:25 crc kubenswrapper[4809]: I1206 05:56:25.399285 4809 status_manager.go:851] "Failed to get status for pod" podUID="c822a806-352f-4b4f-8549-051ff20c4ca2" pod="openshift-marketplace/community-operators-9zm48" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9zm48\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:25 crc kubenswrapper[4809]: I1206 05:56:25.399473 4809 status_manager.go:851] "Failed to get status for pod" podUID="e356b822-4e99-42f8-99ba-7aea197f3d95" pod="openshift-marketplace/redhat-operators-hfgk4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hfgk4\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:25 crc kubenswrapper[4809]: I1206 05:56:25.399657 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:25 crc kubenswrapper[4809]: I1206 05:56:25.399836 4809 status_manager.go:851] "Failed to get status for pod" podUID="001a0f27-e6f3-4d62-9c25-43858654d2b6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:25 crc kubenswrapper[4809]: I1206 05:56:25.400115 4809 status_manager.go:851] "Failed to get status for pod" podUID="c822a806-352f-4b4f-8549-051ff20c4ca2" pod="openshift-marketplace/community-operators-9zm48" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9zm48\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:25 crc kubenswrapper[4809]: I1206 05:56:25.400375 4809 status_manager.go:851] "Failed to get status for pod" podUID="e356b822-4e99-42f8-99ba-7aea197f3d95" pod="openshift-marketplace/redhat-operators-hfgk4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hfgk4\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:25 crc kubenswrapper[4809]: I1206 05:56:25.400546 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:25 crc kubenswrapper[4809]: I1206 05:56:25.400730 4809 status_manager.go:851] "Failed to get status for pod" podUID="001a0f27-e6f3-4d62-9c25-43858654d2b6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:25 crc kubenswrapper[4809]: I1206 05:56:25.400918 4809 status_manager.go:851] "Failed to get status for pod" podUID="893020b2-fa30-4371-86f9-a585517839ce" pod="openshift-marketplace/community-operators-24qxt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-24qxt\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:25 crc kubenswrapper[4809]: I1206 05:56:25.401117 4809 status_manager.go:851] "Failed to get status for pod" podUID="a83cbcc6-62ac-4814-9fce-eb7ea465bcbf" pod="openshift-marketplace/certified-operators-dvwjq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-dvwjq\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:25 crc kubenswrapper[4809]: I1206 05:56:25.401291 4809 status_manager.go:851] "Failed to get status for pod" podUID="2c4b2f56-0b03-41ae-af23-6192b8bd3d34" pod="openshift-marketplace/redhat-marketplace-wj2h7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-wj2h7\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:25 crc kubenswrapper[4809]: I1206 05:56:25.401696 4809 status_manager.go:851] "Failed to get status for pod" podUID="0a9baebe-98e6-44be-ad16-882e47f09248" pod="openshift-authentication/oauth-openshift-558db77b4-jn446" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-jn446\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:25 crc kubenswrapper[4809]: I1206 05:56:25.409129 4809 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5636dd78-3d61-4c6a-b8fd-6d2457e19234" Dec 06 05:56:25 crc kubenswrapper[4809]: I1206 05:56:25.409534 4809 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5636dd78-3d61-4c6a-b8fd-6d2457e19234" Dec 06 05:56:25 crc kubenswrapper[4809]: E1206 05:56:25.410065 4809 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:56:25 crc kubenswrapper[4809]: I1206 05:56:25.410734 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:56:25 crc kubenswrapper[4809]: W1206 05:56:25.436708 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-be987b264a59bee2c3ba41a6ad8689f5e10c8ed8c957d7e5b6b4be504f5661ab WatchSource:0}: Error finding container be987b264a59bee2c3ba41a6ad8689f5e10c8ed8c957d7e5b6b4be504f5661ab: Status 404 returned error can't find the container with id be987b264a59bee2c3ba41a6ad8689f5e10c8ed8c957d7e5b6b4be504f5661ab Dec 06 05:56:25 crc kubenswrapper[4809]: I1206 05:56:25.466792 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"be987b264a59bee2c3ba41a6ad8689f5e10c8ed8c957d7e5b6b4be504f5661ab"} Dec 06 05:56:26 crc kubenswrapper[4809]: E1206 05:56:26.461766 4809 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.154:6443: connect: connection refused" interval="6.4s" Dec 06 05:56:26 crc kubenswrapper[4809]: I1206 05:56:26.476605 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 06 05:56:26 crc kubenswrapper[4809]: I1206 05:56:26.476669 4809 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="5720f7f2977d9af6ed5324a30d08d34a53d55460951b3875b7d7d753d202f349" exitCode=1 Dec 06 05:56:26 crc kubenswrapper[4809]: I1206 05:56:26.476736 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"5720f7f2977d9af6ed5324a30d08d34a53d55460951b3875b7d7d753d202f349"} Dec 06 05:56:26 crc kubenswrapper[4809]: I1206 05:56:26.477544 4809 scope.go:117] "RemoveContainer" containerID="5720f7f2977d9af6ed5324a30d08d34a53d55460951b3875b7d7d753d202f349" Dec 06 05:56:26 crc kubenswrapper[4809]: I1206 05:56:26.477598 4809 status_manager.go:851] "Failed to get status for pod" podUID="001a0f27-e6f3-4d62-9c25-43858654d2b6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:26 crc kubenswrapper[4809]: I1206 05:56:26.477850 4809 status_manager.go:851] "Failed to get status for pod" podUID="a83cbcc6-62ac-4814-9fce-eb7ea465bcbf" pod="openshift-marketplace/certified-operators-dvwjq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-dvwjq\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:26 crc kubenswrapper[4809]: I1206 05:56:26.478080 4809 status_manager.go:851] "Failed to get status for pod" podUID="2c4b2f56-0b03-41ae-af23-6192b8bd3d34" pod="openshift-marketplace/redhat-marketplace-wj2h7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-wj2h7\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:26 crc kubenswrapper[4809]: I1206 05:56:26.478336 4809 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:26 crc kubenswrapper[4809]: I1206 05:56:26.478547 4809 status_manager.go:851] "Failed to get status for pod" podUID="893020b2-fa30-4371-86f9-a585517839ce" pod="openshift-marketplace/community-operators-24qxt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-24qxt\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:26 crc kubenswrapper[4809]: I1206 05:56:26.478757 4809 status_manager.go:851] "Failed to get status for pod" podUID="0a9baebe-98e6-44be-ad16-882e47f09248" pod="openshift-authentication/oauth-openshift-558db77b4-jn446" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-jn446\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:26 crc kubenswrapper[4809]: I1206 05:56:26.478764 4809 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="25f753934cd9b4c6b1569e3fb87533bf08125c07c50fe33596ac1a41ee6554ea" exitCode=0 Dec 06 05:56:26 crc kubenswrapper[4809]: I1206 05:56:26.478777 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"25f753934cd9b4c6b1569e3fb87533bf08125c07c50fe33596ac1a41ee6554ea"} Dec 06 05:56:26 crc kubenswrapper[4809]: I1206 05:56:26.479031 4809 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5636dd78-3d61-4c6a-b8fd-6d2457e19234" Dec 06 05:56:26 crc kubenswrapper[4809]: I1206 05:56:26.479051 4809 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5636dd78-3d61-4c6a-b8fd-6d2457e19234" Dec 06 05:56:26 crc kubenswrapper[4809]: I1206 05:56:26.479055 4809 status_manager.go:851] "Failed to get status for pod" podUID="c822a806-352f-4b4f-8549-051ff20c4ca2" pod="openshift-marketplace/community-operators-9zm48" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9zm48\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:26 crc kubenswrapper[4809]: E1206 05:56:26.479282 4809 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:56:26 crc kubenswrapper[4809]: I1206 05:56:26.479481 4809 status_manager.go:851] "Failed to get status for pod" podUID="e356b822-4e99-42f8-99ba-7aea197f3d95" pod="openshift-marketplace/redhat-operators-hfgk4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hfgk4\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:26 crc kubenswrapper[4809]: I1206 05:56:26.479707 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:26 crc kubenswrapper[4809]: I1206 05:56:26.479994 4809 status_manager.go:851] "Failed to get status for pod" podUID="c822a806-352f-4b4f-8549-051ff20c4ca2" pod="openshift-marketplace/community-operators-9zm48" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9zm48\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:26 crc kubenswrapper[4809]: I1206 05:56:26.480224 4809 status_manager.go:851] "Failed to get status for pod" podUID="e356b822-4e99-42f8-99ba-7aea197f3d95" pod="openshift-marketplace/redhat-operators-hfgk4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hfgk4\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:26 crc kubenswrapper[4809]: I1206 05:56:26.480432 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:26 crc kubenswrapper[4809]: I1206 05:56:26.480644 4809 status_manager.go:851] "Failed to get status for pod" podUID="001a0f27-e6f3-4d62-9c25-43858654d2b6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:26 crc kubenswrapper[4809]: I1206 05:56:26.480833 4809 status_manager.go:851] "Failed to get status for pod" podUID="893020b2-fa30-4371-86f9-a585517839ce" pod="openshift-marketplace/community-operators-24qxt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-24qxt\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:26 crc kubenswrapper[4809]: I1206 05:56:26.481065 4809 status_manager.go:851] "Failed to get status for pod" podUID="a83cbcc6-62ac-4814-9fce-eb7ea465bcbf" pod="openshift-marketplace/certified-operators-dvwjq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-dvwjq\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:26 crc kubenswrapper[4809]: I1206 05:56:26.481342 4809 status_manager.go:851] "Failed to get status for pod" podUID="2c4b2f56-0b03-41ae-af23-6192b8bd3d34" pod="openshift-marketplace/redhat-marketplace-wj2h7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-wj2h7\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:26 crc kubenswrapper[4809]: I1206 05:56:26.481518 4809 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:26 crc kubenswrapper[4809]: I1206 05:56:26.481710 4809 status_manager.go:851] "Failed to get status for pod" podUID="0a9baebe-98e6-44be-ad16-882e47f09248" pod="openshift-authentication/oauth-openshift-558db77b4-jn446" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-jn446\": dial tcp 38.102.83.154:6443: connect: connection refused" Dec 06 05:56:27 crc kubenswrapper[4809]: I1206 05:56:27.495091 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"5ec292df97bfd7d38ce895fda329da71f8284ab3617484fae7934070a6bf0a7d"} Dec 06 05:56:27 crc kubenswrapper[4809]: I1206 05:56:27.495458 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"e2838764a2f78d17f6c4a69a8cf64669e7b9991c0847f0212426f55302caa770"} Dec 06 05:56:27 crc kubenswrapper[4809]: I1206 05:56:27.495475 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"650c9b988233d5919baa33b6787eeff43a2b8429db035dd2d8cfa068799c5331"} Dec 06 05:56:27 crc kubenswrapper[4809]: I1206 05:56:27.495486 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"23f05415cb7880128765827fdb95caba0ca868f01d7749b8c7ac39a5aef4dd1e"} Dec 06 05:56:27 crc kubenswrapper[4809]: I1206 05:56:27.497920 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 06 05:56:27 crc kubenswrapper[4809]: I1206 05:56:27.497992 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"838a7a996f090c68dc35677c87c102eed28d8117b273e3c5260df2933cc7a001"} Dec 06 05:56:28 crc kubenswrapper[4809]: I1206 05:56:28.509802 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b00505b929ad9c6ef9ab88c75b34997219a328e0df0dfcea737abc5702ff36f5"} Dec 06 05:56:28 crc kubenswrapper[4809]: I1206 05:56:28.510090 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:56:28 crc kubenswrapper[4809]: I1206 05:56:28.510167 4809 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5636dd78-3d61-4c6a-b8fd-6d2457e19234" Dec 06 05:56:28 crc kubenswrapper[4809]: I1206 05:56:28.510196 4809 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5636dd78-3d61-4c6a-b8fd-6d2457e19234" Dec 06 05:56:30 crc kubenswrapper[4809]: I1206 05:56:30.411774 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:56:30 crc kubenswrapper[4809]: I1206 05:56:30.412110 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:56:30 crc kubenswrapper[4809]: I1206 05:56:30.416593 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:56:30 crc kubenswrapper[4809]: I1206 05:56:30.548224 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 05:56:30 crc kubenswrapper[4809]: I1206 05:56:30.548514 4809 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 06 05:56:30 crc kubenswrapper[4809]: I1206 05:56:30.548560 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 06 05:56:33 crc kubenswrapper[4809]: I1206 05:56:33.520387 4809 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:56:34 crc kubenswrapper[4809]: I1206 05:56:34.544753 4809 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5636dd78-3d61-4c6a-b8fd-6d2457e19234" Dec 06 05:56:34 crc kubenswrapper[4809]: I1206 05:56:34.544804 4809 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5636dd78-3d61-4c6a-b8fd-6d2457e19234" Dec 06 05:56:34 crc kubenswrapper[4809]: I1206 05:56:34.552924 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:56:34 crc kubenswrapper[4809]: I1206 05:56:34.561629 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 05:56:35 crc kubenswrapper[4809]: I1206 05:56:35.433404 4809 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="931dd383-dc93-4a1c-9f7d-16523d0806ee" Dec 06 05:56:35 crc kubenswrapper[4809]: I1206 05:56:35.551112 4809 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5636dd78-3d61-4c6a-b8fd-6d2457e19234" Dec 06 05:56:35 crc kubenswrapper[4809]: I1206 05:56:35.551154 4809 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5636dd78-3d61-4c6a-b8fd-6d2457e19234" Dec 06 05:56:35 crc kubenswrapper[4809]: I1206 05:56:35.554435 4809 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="931dd383-dc93-4a1c-9f7d-16523d0806ee" Dec 06 05:56:40 crc kubenswrapper[4809]: I1206 05:56:40.552386 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 05:56:40 crc kubenswrapper[4809]: I1206 05:56:40.556903 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 05:56:42 crc kubenswrapper[4809]: I1206 05:56:42.966911 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 06 05:56:43 crc kubenswrapper[4809]: I1206 05:56:43.920134 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 06 05:56:44 crc kubenswrapper[4809]: I1206 05:56:44.197647 4809 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 06 05:56:44 crc kubenswrapper[4809]: I1206 05:56:44.248419 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 06 05:56:44 crc kubenswrapper[4809]: I1206 05:56:44.378670 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 06 05:56:44 crc kubenswrapper[4809]: I1206 05:56:44.400175 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 06 05:56:44 crc kubenswrapper[4809]: I1206 05:56:44.723121 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 06 05:56:44 crc kubenswrapper[4809]: I1206 05:56:44.974418 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 06 05:56:44 crc kubenswrapper[4809]: I1206 05:56:44.988254 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 06 05:56:45 crc kubenswrapper[4809]: I1206 05:56:45.014661 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 06 05:56:45 crc kubenswrapper[4809]: I1206 05:56:45.035160 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 06 05:56:45 crc kubenswrapper[4809]: I1206 05:56:45.322651 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 06 05:56:45 crc kubenswrapper[4809]: I1206 05:56:45.350663 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 06 05:56:45 crc kubenswrapper[4809]: I1206 05:56:45.398218 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 06 05:56:45 crc kubenswrapper[4809]: I1206 05:56:45.409493 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 06 05:56:45 crc kubenswrapper[4809]: I1206 05:56:45.523221 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 06 05:56:45 crc kubenswrapper[4809]: I1206 05:56:45.532765 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 06 05:56:45 crc kubenswrapper[4809]: I1206 05:56:45.554091 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 06 05:56:45 crc kubenswrapper[4809]: I1206 05:56:45.567813 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 06 05:56:45 crc kubenswrapper[4809]: I1206 05:56:45.581604 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 06 05:56:45 crc kubenswrapper[4809]: I1206 05:56:45.596959 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 06 05:56:45 crc kubenswrapper[4809]: I1206 05:56:45.601744 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 06 05:56:45 crc kubenswrapper[4809]: I1206 05:56:45.635717 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 06 05:56:45 crc kubenswrapper[4809]: I1206 05:56:45.730202 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 06 05:56:45 crc kubenswrapper[4809]: I1206 05:56:45.831699 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 06 05:56:45 crc kubenswrapper[4809]: I1206 05:56:45.884573 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 06 05:56:45 crc kubenswrapper[4809]: I1206 05:56:45.924081 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 06 05:56:46 crc kubenswrapper[4809]: I1206 05:56:46.136212 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 06 05:56:46 crc kubenswrapper[4809]: I1206 05:56:46.141681 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 06 05:56:46 crc kubenswrapper[4809]: I1206 05:56:46.213560 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 06 05:56:46 crc kubenswrapper[4809]: I1206 05:56:46.256375 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 06 05:56:46 crc kubenswrapper[4809]: I1206 05:56:46.260281 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 06 05:56:46 crc kubenswrapper[4809]: I1206 05:56:46.383899 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 06 05:56:46 crc kubenswrapper[4809]: I1206 05:56:46.497874 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 06 05:56:46 crc kubenswrapper[4809]: I1206 05:56:46.508612 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 06 05:56:46 crc kubenswrapper[4809]: I1206 05:56:46.545586 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 06 05:56:46 crc kubenswrapper[4809]: I1206 05:56:46.625129 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 06 05:56:46 crc kubenswrapper[4809]: I1206 05:56:46.658374 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 06 05:56:46 crc kubenswrapper[4809]: I1206 05:56:46.745885 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 06 05:56:46 crc kubenswrapper[4809]: I1206 05:56:46.759187 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 06 05:56:46 crc kubenswrapper[4809]: I1206 05:56:46.793913 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 06 05:56:46 crc kubenswrapper[4809]: I1206 05:56:46.941366 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 06 05:56:46 crc kubenswrapper[4809]: I1206 05:56:46.973000 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 06 05:56:47 crc kubenswrapper[4809]: I1206 05:56:47.109231 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 06 05:56:47 crc kubenswrapper[4809]: I1206 05:56:47.169766 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 06 05:56:47 crc kubenswrapper[4809]: I1206 05:56:47.200803 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 06 05:56:47 crc kubenswrapper[4809]: I1206 05:56:47.256089 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 06 05:56:47 crc kubenswrapper[4809]: I1206 05:56:47.276922 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 06 05:56:47 crc kubenswrapper[4809]: I1206 05:56:47.288540 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 06 05:56:47 crc kubenswrapper[4809]: I1206 05:56:47.293000 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 06 05:56:47 crc kubenswrapper[4809]: I1206 05:56:47.297062 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 06 05:56:47 crc kubenswrapper[4809]: I1206 05:56:47.422168 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 06 05:56:47 crc kubenswrapper[4809]: I1206 05:56:47.509063 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 06 05:56:47 crc kubenswrapper[4809]: I1206 05:56:47.537869 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 06 05:56:47 crc kubenswrapper[4809]: I1206 05:56:47.547223 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 06 05:56:47 crc kubenswrapper[4809]: I1206 05:56:47.549559 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 06 05:56:47 crc kubenswrapper[4809]: I1206 05:56:47.590210 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 06 05:56:47 crc kubenswrapper[4809]: I1206 05:56:47.599248 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 06 05:56:47 crc kubenswrapper[4809]: I1206 05:56:47.624166 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 06 05:56:47 crc kubenswrapper[4809]: I1206 05:56:47.662674 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 06 05:56:47 crc kubenswrapper[4809]: I1206 05:56:47.685999 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 06 05:56:47 crc kubenswrapper[4809]: I1206 05:56:47.735596 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 06 05:56:47 crc kubenswrapper[4809]: I1206 05:56:47.805834 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 06 05:56:47 crc kubenswrapper[4809]: I1206 05:56:47.824190 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 06 05:56:47 crc kubenswrapper[4809]: I1206 05:56:47.830775 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 06 05:56:47 crc kubenswrapper[4809]: I1206 05:56:47.879074 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 06 05:56:47 crc kubenswrapper[4809]: I1206 05:56:47.886608 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 06 05:56:47 crc kubenswrapper[4809]: I1206 05:56:47.917708 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 06 05:56:47 crc kubenswrapper[4809]: I1206 05:56:47.944792 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 06 05:56:47 crc kubenswrapper[4809]: I1206 05:56:47.946437 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 06 05:56:48 crc kubenswrapper[4809]: I1206 05:56:48.072119 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 06 05:56:48 crc kubenswrapper[4809]: I1206 05:56:48.217944 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 06 05:56:48 crc kubenswrapper[4809]: I1206 05:56:48.230569 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 06 05:56:48 crc kubenswrapper[4809]: I1206 05:56:48.238689 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 06 05:56:48 crc kubenswrapper[4809]: I1206 05:56:48.244202 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 06 05:56:48 crc kubenswrapper[4809]: I1206 05:56:48.560667 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 06 05:56:48 crc kubenswrapper[4809]: I1206 05:56:48.601343 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 06 05:56:48 crc kubenswrapper[4809]: I1206 05:56:48.607319 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 06 05:56:48 crc kubenswrapper[4809]: I1206 05:56:48.630158 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 06 05:56:48 crc kubenswrapper[4809]: I1206 05:56:48.635058 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 06 05:56:48 crc kubenswrapper[4809]: I1206 05:56:48.644861 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 06 05:56:48 crc kubenswrapper[4809]: I1206 05:56:48.649921 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 06 05:56:48 crc kubenswrapper[4809]: I1206 05:56:48.673808 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 06 05:56:48 crc kubenswrapper[4809]: I1206 05:56:48.682625 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 06 05:56:48 crc kubenswrapper[4809]: I1206 05:56:48.821862 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 06 05:56:48 crc kubenswrapper[4809]: I1206 05:56:48.876654 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 06 05:56:48 crc kubenswrapper[4809]: I1206 05:56:48.898899 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 06 05:56:48 crc kubenswrapper[4809]: I1206 05:56:48.955120 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 06 05:56:48 crc kubenswrapper[4809]: I1206 05:56:48.956025 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 06 05:56:48 crc kubenswrapper[4809]: I1206 05:56:48.969454 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 06 05:56:49 crc kubenswrapper[4809]: I1206 05:56:49.074567 4809 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 06 05:56:49 crc kubenswrapper[4809]: I1206 05:56:49.133754 4809 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 06 05:56:49 crc kubenswrapper[4809]: I1206 05:56:49.181321 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 06 05:56:49 crc kubenswrapper[4809]: I1206 05:56:49.194459 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 06 05:56:49 crc kubenswrapper[4809]: I1206 05:56:49.262247 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 06 05:56:49 crc kubenswrapper[4809]: I1206 05:56:49.382017 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 06 05:56:49 crc kubenswrapper[4809]: I1206 05:56:49.389914 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 06 05:56:49 crc kubenswrapper[4809]: I1206 05:56:49.417507 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 06 05:56:49 crc kubenswrapper[4809]: I1206 05:56:49.435025 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 06 05:56:49 crc kubenswrapper[4809]: I1206 05:56:49.486825 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 06 05:56:49 crc kubenswrapper[4809]: I1206 05:56:49.575503 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 06 05:56:49 crc kubenswrapper[4809]: I1206 05:56:49.612642 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 06 05:56:49 crc kubenswrapper[4809]: I1206 05:56:49.642209 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 06 05:56:49 crc kubenswrapper[4809]: I1206 05:56:49.887594 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 06 05:56:49 crc kubenswrapper[4809]: I1206 05:56:49.962563 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 06 05:56:50 crc kubenswrapper[4809]: I1206 05:56:50.224747 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 06 05:56:50 crc kubenswrapper[4809]: I1206 05:56:50.247671 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 06 05:56:50 crc kubenswrapper[4809]: I1206 05:56:50.267972 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 06 05:56:50 crc kubenswrapper[4809]: I1206 05:56:50.282412 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 06 05:56:50 crc kubenswrapper[4809]: I1206 05:56:50.313283 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 06 05:56:50 crc kubenswrapper[4809]: I1206 05:56:50.335374 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 06 05:56:50 crc kubenswrapper[4809]: I1206 05:56:50.350302 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 06 05:56:50 crc kubenswrapper[4809]: I1206 05:56:50.495016 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 06 05:56:50 crc kubenswrapper[4809]: I1206 05:56:50.525415 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 06 05:56:50 crc kubenswrapper[4809]: I1206 05:56:50.674816 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 06 05:56:50 crc kubenswrapper[4809]: I1206 05:56:50.712481 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 06 05:56:50 crc kubenswrapper[4809]: I1206 05:56:50.736069 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 06 05:56:50 crc kubenswrapper[4809]: I1206 05:56:50.904631 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 06 05:56:51 crc kubenswrapper[4809]: I1206 05:56:51.025606 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 06 05:56:51 crc kubenswrapper[4809]: I1206 05:56:51.034549 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 06 05:56:51 crc kubenswrapper[4809]: I1206 05:56:51.035089 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 06 05:56:51 crc kubenswrapper[4809]: I1206 05:56:51.040477 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 06 05:56:51 crc kubenswrapper[4809]: I1206 05:56:51.162845 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 06 05:56:51 crc kubenswrapper[4809]: I1206 05:56:51.196070 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 06 05:56:51 crc kubenswrapper[4809]: I1206 05:56:51.367839 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 06 05:56:51 crc kubenswrapper[4809]: I1206 05:56:51.489661 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 06 05:56:51 crc kubenswrapper[4809]: I1206 05:56:51.577420 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 06 05:56:51 crc kubenswrapper[4809]: I1206 05:56:51.582258 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 06 05:56:51 crc kubenswrapper[4809]: I1206 05:56:51.618961 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 06 05:56:51 crc kubenswrapper[4809]: I1206 05:56:51.625021 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 06 05:56:51 crc kubenswrapper[4809]: I1206 05:56:51.635975 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 06 05:56:51 crc kubenswrapper[4809]: I1206 05:56:51.648422 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 06 05:56:51 crc kubenswrapper[4809]: I1206 05:56:51.695768 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 06 05:56:51 crc kubenswrapper[4809]: I1206 05:56:51.728750 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 06 05:56:51 crc kubenswrapper[4809]: I1206 05:56:51.743783 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 06 05:56:51 crc kubenswrapper[4809]: I1206 05:56:51.776648 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 06 05:56:51 crc kubenswrapper[4809]: I1206 05:56:51.820230 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 06 05:56:51 crc kubenswrapper[4809]: I1206 05:56:51.978895 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 06 05:56:52 crc kubenswrapper[4809]: I1206 05:56:52.217845 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 06 05:56:52 crc kubenswrapper[4809]: I1206 05:56:52.369389 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 06 05:56:52 crc kubenswrapper[4809]: I1206 05:56:52.431162 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 06 05:56:52 crc kubenswrapper[4809]: I1206 05:56:52.447518 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 06 05:56:52 crc kubenswrapper[4809]: I1206 05:56:52.483486 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 06 05:56:52 crc kubenswrapper[4809]: I1206 05:56:52.529805 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 06 05:56:52 crc kubenswrapper[4809]: I1206 05:56:52.698287 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 06 05:56:52 crc kubenswrapper[4809]: I1206 05:56:52.758683 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 06 05:56:52 crc kubenswrapper[4809]: I1206 05:56:52.811095 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 06 05:56:52 crc kubenswrapper[4809]: I1206 05:56:52.830965 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 06 05:56:52 crc kubenswrapper[4809]: I1206 05:56:52.945752 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 06 05:56:52 crc kubenswrapper[4809]: I1206 05:56:52.973505 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 06 05:56:52 crc kubenswrapper[4809]: I1206 05:56:52.980053 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 06 05:56:53 crc kubenswrapper[4809]: I1206 05:56:53.097863 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 06 05:56:53 crc kubenswrapper[4809]: I1206 05:56:53.110624 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 06 05:56:53 crc kubenswrapper[4809]: I1206 05:56:53.152755 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 06 05:56:53 crc kubenswrapper[4809]: I1206 05:56:53.331442 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 06 05:56:53 crc kubenswrapper[4809]: I1206 05:56:53.397124 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 06 05:56:53 crc kubenswrapper[4809]: I1206 05:56:53.469341 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 06 05:56:53 crc kubenswrapper[4809]: I1206 05:56:53.571906 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 06 05:56:53 crc kubenswrapper[4809]: I1206 05:56:53.582726 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 06 05:56:53 crc kubenswrapper[4809]: I1206 05:56:53.703513 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 06 05:56:53 crc kubenswrapper[4809]: I1206 05:56:53.746177 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 06 05:56:53 crc kubenswrapper[4809]: I1206 05:56:53.766898 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 06 05:56:53 crc kubenswrapper[4809]: I1206 05:56:53.788234 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 06 05:56:53 crc kubenswrapper[4809]: I1206 05:56:53.942039 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 06 05:56:54 crc kubenswrapper[4809]: I1206 05:56:54.035881 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 06 05:56:54 crc kubenswrapper[4809]: I1206 05:56:54.091611 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 06 05:56:54 crc kubenswrapper[4809]: I1206 05:56:54.264791 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 06 05:56:54 crc kubenswrapper[4809]: I1206 05:56:54.289699 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 06 05:56:54 crc kubenswrapper[4809]: I1206 05:56:54.320272 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 06 05:56:54 crc kubenswrapper[4809]: I1206 05:56:54.509609 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 06 05:56:54 crc kubenswrapper[4809]: I1206 05:56:54.532763 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 06 05:56:54 crc kubenswrapper[4809]: I1206 05:56:54.742894 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 06 05:56:54 crc kubenswrapper[4809]: I1206 05:56:54.751622 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 06 05:56:54 crc kubenswrapper[4809]: I1206 05:56:54.791659 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 06 05:56:54 crc kubenswrapper[4809]: I1206 05:56:54.869995 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 06 05:56:54 crc kubenswrapper[4809]: I1206 05:56:54.983198 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 06 05:56:55 crc kubenswrapper[4809]: I1206 05:56:55.047560 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 06 05:56:55 crc kubenswrapper[4809]: I1206 05:56:55.060046 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 06 05:56:55 crc kubenswrapper[4809]: I1206 05:56:55.193320 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 06 05:56:55 crc kubenswrapper[4809]: I1206 05:56:55.268582 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 06 05:56:55 crc kubenswrapper[4809]: I1206 05:56:55.330371 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 06 05:56:55 crc kubenswrapper[4809]: I1206 05:56:55.402722 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 06 05:56:55 crc kubenswrapper[4809]: I1206 05:56:55.452859 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 06 05:56:55 crc kubenswrapper[4809]: I1206 05:56:55.472615 4809 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 06 05:56:55 crc kubenswrapper[4809]: I1206 05:56:55.535885 4809 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 06 05:56:55 crc kubenswrapper[4809]: I1206 05:56:55.554426 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 06 05:56:55 crc kubenswrapper[4809]: I1206 05:56:55.661621 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 06 05:56:55 crc kubenswrapper[4809]: I1206 05:56:55.677136 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 06 05:56:55 crc kubenswrapper[4809]: I1206 05:56:55.679605 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 06 05:56:55 crc kubenswrapper[4809]: I1206 05:56:55.723958 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 06 05:56:55 crc kubenswrapper[4809]: I1206 05:56:55.761391 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.038364 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.063914 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.105335 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.131461 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.224234 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.370014 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.377682 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.418185 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.418404 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.495191 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.539087 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.608952 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.632331 4809 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.635513 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=45.635495789 podStartE2EDuration="45.635495789s" podCreationTimestamp="2025-12-06 05:56:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:56:33.063344263 +0000 UTC m=+317.952327205" watchObservedRunningTime="2025-12-06 05:56:56.635495789 +0000 UTC m=+341.524478731" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.636854 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jn446","openshift-kube-apiserver/kube-apiserver-crc"] Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.636905 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-55889b984c-nxw8t"] Dec 06 05:56:56 crc kubenswrapper[4809]: E1206 05:56:56.637101 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a9baebe-98e6-44be-ad16-882e47f09248" containerName="oauth-openshift" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.637116 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a9baebe-98e6-44be-ad16-882e47f09248" containerName="oauth-openshift" Dec 06 05:56:56 crc kubenswrapper[4809]: E1206 05:56:56.637136 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="001a0f27-e6f3-4d62-9c25-43858654d2b6" containerName="installer" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.637142 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="001a0f27-e6f3-4d62-9c25-43858654d2b6" containerName="installer" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.637221 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a9baebe-98e6-44be-ad16-882e47f09248" containerName="oauth-openshift" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.637231 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="001a0f27-e6f3-4d62-9c25-43858654d2b6" containerName="installer" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.637246 4809 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5636dd78-3d61-4c6a-b8fd-6d2457e19234" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.637266 4809 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5636dd78-3d61-4c6a-b8fd-6d2457e19234" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.637604 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.640769 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.641657 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.641988 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.642622 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.642715 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.642729 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.642750 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.642852 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.644666 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.644731 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.644802 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.644883 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.645042 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.653050 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.657995 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.661822 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.681574 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=23.681554866 podStartE2EDuration="23.681554866s" podCreationTimestamp="2025-12-06 05:56:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:56:56.676303981 +0000 UTC m=+341.565286943" watchObservedRunningTime="2025-12-06 05:56:56.681554866 +0000 UTC m=+341.570537818" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.715443 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.718638 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/404214e0-30aa-44ff-b2d0-fb3c127068f9-v4-0-config-system-session\") pod \"oauth-openshift-55889b984c-nxw8t\" (UID: \"404214e0-30aa-44ff-b2d0-fb3c127068f9\") " pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.718699 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/404214e0-30aa-44ff-b2d0-fb3c127068f9-v4-0-config-system-router-certs\") pod \"oauth-openshift-55889b984c-nxw8t\" (UID: \"404214e0-30aa-44ff-b2d0-fb3c127068f9\") " pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.718734 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/404214e0-30aa-44ff-b2d0-fb3c127068f9-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-55889b984c-nxw8t\" (UID: \"404214e0-30aa-44ff-b2d0-fb3c127068f9\") " pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.718802 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/404214e0-30aa-44ff-b2d0-fb3c127068f9-audit-dir\") pod \"oauth-openshift-55889b984c-nxw8t\" (UID: \"404214e0-30aa-44ff-b2d0-fb3c127068f9\") " pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.718842 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/404214e0-30aa-44ff-b2d0-fb3c127068f9-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-55889b984c-nxw8t\" (UID: \"404214e0-30aa-44ff-b2d0-fb3c127068f9\") " pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.718868 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/404214e0-30aa-44ff-b2d0-fb3c127068f9-audit-policies\") pod \"oauth-openshift-55889b984c-nxw8t\" (UID: \"404214e0-30aa-44ff-b2d0-fb3c127068f9\") " pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.718980 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/404214e0-30aa-44ff-b2d0-fb3c127068f9-v4-0-config-system-service-ca\") pod \"oauth-openshift-55889b984c-nxw8t\" (UID: \"404214e0-30aa-44ff-b2d0-fb3c127068f9\") " pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.719055 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/404214e0-30aa-44ff-b2d0-fb3c127068f9-v4-0-config-user-template-error\") pod \"oauth-openshift-55889b984c-nxw8t\" (UID: \"404214e0-30aa-44ff-b2d0-fb3c127068f9\") " pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.719097 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qcd5\" (UniqueName: \"kubernetes.io/projected/404214e0-30aa-44ff-b2d0-fb3c127068f9-kube-api-access-9qcd5\") pod \"oauth-openshift-55889b984c-nxw8t\" (UID: \"404214e0-30aa-44ff-b2d0-fb3c127068f9\") " pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.719141 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/404214e0-30aa-44ff-b2d0-fb3c127068f9-v4-0-config-user-template-login\") pod \"oauth-openshift-55889b984c-nxw8t\" (UID: \"404214e0-30aa-44ff-b2d0-fb3c127068f9\") " pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.719242 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/404214e0-30aa-44ff-b2d0-fb3c127068f9-v4-0-config-system-cliconfig\") pod \"oauth-openshift-55889b984c-nxw8t\" (UID: \"404214e0-30aa-44ff-b2d0-fb3c127068f9\") " pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.719271 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/404214e0-30aa-44ff-b2d0-fb3c127068f9-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-55889b984c-nxw8t\" (UID: \"404214e0-30aa-44ff-b2d0-fb3c127068f9\") " pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.719325 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/404214e0-30aa-44ff-b2d0-fb3c127068f9-v4-0-config-system-serving-cert\") pod \"oauth-openshift-55889b984c-nxw8t\" (UID: \"404214e0-30aa-44ff-b2d0-fb3c127068f9\") " pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.719352 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/404214e0-30aa-44ff-b2d0-fb3c127068f9-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-55889b984c-nxw8t\" (UID: \"404214e0-30aa-44ff-b2d0-fb3c127068f9\") " pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.740769 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.820548 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/404214e0-30aa-44ff-b2d0-fb3c127068f9-v4-0-config-system-cliconfig\") pod \"oauth-openshift-55889b984c-nxw8t\" (UID: \"404214e0-30aa-44ff-b2d0-fb3c127068f9\") " pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.820598 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/404214e0-30aa-44ff-b2d0-fb3c127068f9-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-55889b984c-nxw8t\" (UID: \"404214e0-30aa-44ff-b2d0-fb3c127068f9\") " pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.820617 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/404214e0-30aa-44ff-b2d0-fb3c127068f9-v4-0-config-system-serving-cert\") pod \"oauth-openshift-55889b984c-nxw8t\" (UID: \"404214e0-30aa-44ff-b2d0-fb3c127068f9\") " pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.820636 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/404214e0-30aa-44ff-b2d0-fb3c127068f9-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-55889b984c-nxw8t\" (UID: \"404214e0-30aa-44ff-b2d0-fb3c127068f9\") " pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.820689 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/404214e0-30aa-44ff-b2d0-fb3c127068f9-v4-0-config-system-session\") pod \"oauth-openshift-55889b984c-nxw8t\" (UID: \"404214e0-30aa-44ff-b2d0-fb3c127068f9\") " pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.820708 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/404214e0-30aa-44ff-b2d0-fb3c127068f9-v4-0-config-system-router-certs\") pod \"oauth-openshift-55889b984c-nxw8t\" (UID: \"404214e0-30aa-44ff-b2d0-fb3c127068f9\") " pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.820725 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/404214e0-30aa-44ff-b2d0-fb3c127068f9-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-55889b984c-nxw8t\" (UID: \"404214e0-30aa-44ff-b2d0-fb3c127068f9\") " pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.820759 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/404214e0-30aa-44ff-b2d0-fb3c127068f9-audit-dir\") pod \"oauth-openshift-55889b984c-nxw8t\" (UID: \"404214e0-30aa-44ff-b2d0-fb3c127068f9\") " pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.820781 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/404214e0-30aa-44ff-b2d0-fb3c127068f9-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-55889b984c-nxw8t\" (UID: \"404214e0-30aa-44ff-b2d0-fb3c127068f9\") " pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.820798 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/404214e0-30aa-44ff-b2d0-fb3c127068f9-audit-policies\") pod \"oauth-openshift-55889b984c-nxw8t\" (UID: \"404214e0-30aa-44ff-b2d0-fb3c127068f9\") " pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.820813 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/404214e0-30aa-44ff-b2d0-fb3c127068f9-v4-0-config-system-service-ca\") pod \"oauth-openshift-55889b984c-nxw8t\" (UID: \"404214e0-30aa-44ff-b2d0-fb3c127068f9\") " pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.820833 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/404214e0-30aa-44ff-b2d0-fb3c127068f9-v4-0-config-user-template-error\") pod \"oauth-openshift-55889b984c-nxw8t\" (UID: \"404214e0-30aa-44ff-b2d0-fb3c127068f9\") " pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.820851 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qcd5\" (UniqueName: \"kubernetes.io/projected/404214e0-30aa-44ff-b2d0-fb3c127068f9-kube-api-access-9qcd5\") pod \"oauth-openshift-55889b984c-nxw8t\" (UID: \"404214e0-30aa-44ff-b2d0-fb3c127068f9\") " pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.820869 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/404214e0-30aa-44ff-b2d0-fb3c127068f9-v4-0-config-user-template-login\") pod \"oauth-openshift-55889b984c-nxw8t\" (UID: \"404214e0-30aa-44ff-b2d0-fb3c127068f9\") " pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.822103 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/404214e0-30aa-44ff-b2d0-fb3c127068f9-audit-dir\") pod \"oauth-openshift-55889b984c-nxw8t\" (UID: \"404214e0-30aa-44ff-b2d0-fb3c127068f9\") " pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.822467 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/404214e0-30aa-44ff-b2d0-fb3c127068f9-audit-policies\") pod \"oauth-openshift-55889b984c-nxw8t\" (UID: \"404214e0-30aa-44ff-b2d0-fb3c127068f9\") " pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.824119 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/404214e0-30aa-44ff-b2d0-fb3c127068f9-v4-0-config-system-service-ca\") pod \"oauth-openshift-55889b984c-nxw8t\" (UID: \"404214e0-30aa-44ff-b2d0-fb3c127068f9\") " pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.824230 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/404214e0-30aa-44ff-b2d0-fb3c127068f9-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-55889b984c-nxw8t\" (UID: \"404214e0-30aa-44ff-b2d0-fb3c127068f9\") " pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.824248 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/404214e0-30aa-44ff-b2d0-fb3c127068f9-v4-0-config-system-cliconfig\") pod \"oauth-openshift-55889b984c-nxw8t\" (UID: \"404214e0-30aa-44ff-b2d0-fb3c127068f9\") " pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.826757 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/404214e0-30aa-44ff-b2d0-fb3c127068f9-v4-0-config-system-serving-cert\") pod \"oauth-openshift-55889b984c-nxw8t\" (UID: \"404214e0-30aa-44ff-b2d0-fb3c127068f9\") " pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.827000 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/404214e0-30aa-44ff-b2d0-fb3c127068f9-v4-0-config-user-template-error\") pod \"oauth-openshift-55889b984c-nxw8t\" (UID: \"404214e0-30aa-44ff-b2d0-fb3c127068f9\") " pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.828987 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/404214e0-30aa-44ff-b2d0-fb3c127068f9-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-55889b984c-nxw8t\" (UID: \"404214e0-30aa-44ff-b2d0-fb3c127068f9\") " pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.830155 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/404214e0-30aa-44ff-b2d0-fb3c127068f9-v4-0-config-system-session\") pod \"oauth-openshift-55889b984c-nxw8t\" (UID: \"404214e0-30aa-44ff-b2d0-fb3c127068f9\") " pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.831214 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/404214e0-30aa-44ff-b2d0-fb3c127068f9-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-55889b984c-nxw8t\" (UID: \"404214e0-30aa-44ff-b2d0-fb3c127068f9\") " pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.831650 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/404214e0-30aa-44ff-b2d0-fb3c127068f9-v4-0-config-system-router-certs\") pod \"oauth-openshift-55889b984c-nxw8t\" (UID: \"404214e0-30aa-44ff-b2d0-fb3c127068f9\") " pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.832423 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/404214e0-30aa-44ff-b2d0-fb3c127068f9-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-55889b984c-nxw8t\" (UID: \"404214e0-30aa-44ff-b2d0-fb3c127068f9\") " pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.834233 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/404214e0-30aa-44ff-b2d0-fb3c127068f9-v4-0-config-user-template-login\") pod \"oauth-openshift-55889b984c-nxw8t\" (UID: \"404214e0-30aa-44ff-b2d0-fb3c127068f9\") " pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.840564 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qcd5\" (UniqueName: \"kubernetes.io/projected/404214e0-30aa-44ff-b2d0-fb3c127068f9-kube-api-access-9qcd5\") pod \"oauth-openshift-55889b984c-nxw8t\" (UID: \"404214e0-30aa-44ff-b2d0-fb3c127068f9\") " pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.857989 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.921164 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 06 05:56:56 crc kubenswrapper[4809]: I1206 05:56:56.954281 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" Dec 06 05:56:57 crc kubenswrapper[4809]: I1206 05:56:57.052430 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 06 05:56:57 crc kubenswrapper[4809]: I1206 05:56:57.058994 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 06 05:56:57 crc kubenswrapper[4809]: I1206 05:56:57.173199 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-55889b984c-nxw8t"] Dec 06 05:56:57 crc kubenswrapper[4809]: I1206 05:56:57.211697 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 06 05:56:57 crc kubenswrapper[4809]: I1206 05:56:57.220612 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 06 05:56:57 crc kubenswrapper[4809]: I1206 05:56:57.230836 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 06 05:56:57 crc kubenswrapper[4809]: I1206 05:56:57.264161 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 06 05:56:57 crc kubenswrapper[4809]: I1206 05:56:57.351418 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 06 05:56:57 crc kubenswrapper[4809]: I1206 05:56:57.373340 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 06 05:56:57 crc kubenswrapper[4809]: I1206 05:56:57.396619 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a9baebe-98e6-44be-ad16-882e47f09248" path="/var/lib/kubelet/pods/0a9baebe-98e6-44be-ad16-882e47f09248/volumes" Dec 06 05:56:57 crc kubenswrapper[4809]: I1206 05:56:57.463473 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 06 05:56:57 crc kubenswrapper[4809]: I1206 05:56:57.473418 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 06 05:56:57 crc kubenswrapper[4809]: I1206 05:56:57.675792 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" event={"ID":"404214e0-30aa-44ff-b2d0-fb3c127068f9","Type":"ContainerStarted","Data":"ec18d9ee2d4ac08e29c13c1494a1ab1a12fa1e7f5f90278e958d5aed219963ff"} Dec 06 05:56:57 crc kubenswrapper[4809]: I1206 05:56:57.675852 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" event={"ID":"404214e0-30aa-44ff-b2d0-fb3c127068f9","Type":"ContainerStarted","Data":"0a1fa094e368b03aeb2273e1e7e9f549bffd0307f4f3bfa3001ba2f859d43c73"} Dec 06 05:56:57 crc kubenswrapper[4809]: I1206 05:56:57.676243 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" Dec 06 05:56:57 crc kubenswrapper[4809]: I1206 05:56:57.698421 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" podStartSLOduration=63.698406431 podStartE2EDuration="1m3.698406431s" podCreationTimestamp="2025-12-06 05:55:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:56:57.696285122 +0000 UTC m=+342.585268074" watchObservedRunningTime="2025-12-06 05:56:57.698406431 +0000 UTC m=+342.587389373" Dec 06 05:56:57 crc kubenswrapper[4809]: I1206 05:56:57.755555 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" Dec 06 05:56:57 crc kubenswrapper[4809]: I1206 05:56:57.961336 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 06 05:56:58 crc kubenswrapper[4809]: I1206 05:56:58.052975 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 06 05:56:58 crc kubenswrapper[4809]: I1206 05:56:58.194858 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 06 05:56:58 crc kubenswrapper[4809]: I1206 05:56:58.273722 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 06 05:56:58 crc kubenswrapper[4809]: I1206 05:56:58.324510 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 06 05:56:58 crc kubenswrapper[4809]: I1206 05:56:58.440678 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 06 05:56:58 crc kubenswrapper[4809]: I1206 05:56:58.441582 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 06 05:56:58 crc kubenswrapper[4809]: I1206 05:56:58.442883 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 06 05:56:58 crc kubenswrapper[4809]: I1206 05:56:58.501742 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 06 05:56:58 crc kubenswrapper[4809]: I1206 05:56:58.724851 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 06 05:56:58 crc kubenswrapper[4809]: I1206 05:56:58.752189 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 06 05:56:59 crc kubenswrapper[4809]: I1206 05:56:59.015132 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 06 05:56:59 crc kubenswrapper[4809]: I1206 05:56:59.199509 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 06 05:56:59 crc kubenswrapper[4809]: I1206 05:56:59.624758 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 06 05:56:59 crc kubenswrapper[4809]: I1206 05:56:59.653030 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 06 05:56:59 crc kubenswrapper[4809]: I1206 05:56:59.683027 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 06 05:57:00 crc kubenswrapper[4809]: I1206 05:57:00.436810 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 06 05:57:06 crc kubenswrapper[4809]: I1206 05:57:06.455653 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 06 05:57:07 crc kubenswrapper[4809]: I1206 05:57:07.081410 4809 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 06 05:57:07 crc kubenswrapper[4809]: I1206 05:57:07.081662 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://7eb71090d0b795b8d65e3853cd4fae206875ec874c6e70eddaaf8b431b7889fa" gracePeriod=5 Dec 06 05:57:12 crc kubenswrapper[4809]: I1206 05:57:12.643340 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 06 05:57:12 crc kubenswrapper[4809]: I1206 05:57:12.644773 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 05:57:12 crc kubenswrapper[4809]: I1206 05:57:12.721735 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 06 05:57:12 crc kubenswrapper[4809]: I1206 05:57:12.721849 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 06 05:57:12 crc kubenswrapper[4809]: I1206 05:57:12.721873 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 06 05:57:12 crc kubenswrapper[4809]: I1206 05:57:12.721857 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 05:57:12 crc kubenswrapper[4809]: I1206 05:57:12.721914 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 06 05:57:12 crc kubenswrapper[4809]: I1206 05:57:12.721959 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 05:57:12 crc kubenswrapper[4809]: I1206 05:57:12.721974 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 06 05:57:12 crc kubenswrapper[4809]: I1206 05:57:12.722001 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 05:57:12 crc kubenswrapper[4809]: I1206 05:57:12.722001 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 05:57:12 crc kubenswrapper[4809]: I1206 05:57:12.722184 4809 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 06 05:57:12 crc kubenswrapper[4809]: I1206 05:57:12.722199 4809 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 06 05:57:12 crc kubenswrapper[4809]: I1206 05:57:12.722211 4809 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 06 05:57:12 crc kubenswrapper[4809]: I1206 05:57:12.722224 4809 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 06 05:57:12 crc kubenswrapper[4809]: I1206 05:57:12.730996 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 05:57:12 crc kubenswrapper[4809]: I1206 05:57:12.766684 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 06 05:57:12 crc kubenswrapper[4809]: I1206 05:57:12.766748 4809 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="7eb71090d0b795b8d65e3853cd4fae206875ec874c6e70eddaaf8b431b7889fa" exitCode=137 Dec 06 05:57:12 crc kubenswrapper[4809]: I1206 05:57:12.766810 4809 scope.go:117] "RemoveContainer" containerID="7eb71090d0b795b8d65e3853cd4fae206875ec874c6e70eddaaf8b431b7889fa" Dec 06 05:57:12 crc kubenswrapper[4809]: I1206 05:57:12.766839 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 05:57:12 crc kubenswrapper[4809]: I1206 05:57:12.789611 4809 scope.go:117] "RemoveContainer" containerID="7eb71090d0b795b8d65e3853cd4fae206875ec874c6e70eddaaf8b431b7889fa" Dec 06 05:57:12 crc kubenswrapper[4809]: E1206 05:57:12.790199 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7eb71090d0b795b8d65e3853cd4fae206875ec874c6e70eddaaf8b431b7889fa\": container with ID starting with 7eb71090d0b795b8d65e3853cd4fae206875ec874c6e70eddaaf8b431b7889fa not found: ID does not exist" containerID="7eb71090d0b795b8d65e3853cd4fae206875ec874c6e70eddaaf8b431b7889fa" Dec 06 05:57:12 crc kubenswrapper[4809]: I1206 05:57:12.790246 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7eb71090d0b795b8d65e3853cd4fae206875ec874c6e70eddaaf8b431b7889fa"} err="failed to get container status \"7eb71090d0b795b8d65e3853cd4fae206875ec874c6e70eddaaf8b431b7889fa\": rpc error: code = NotFound desc = could not find container \"7eb71090d0b795b8d65e3853cd4fae206875ec874c6e70eddaaf8b431b7889fa\": container with ID starting with 7eb71090d0b795b8d65e3853cd4fae206875ec874c6e70eddaaf8b431b7889fa not found: ID does not exist" Dec 06 05:57:12 crc kubenswrapper[4809]: I1206 05:57:12.823084 4809 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 06 05:57:13 crc kubenswrapper[4809]: I1206 05:57:13.396147 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 06 05:57:13 crc kubenswrapper[4809]: I1206 05:57:13.396858 4809 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Dec 06 05:57:13 crc kubenswrapper[4809]: I1206 05:57:13.407172 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 06 05:57:13 crc kubenswrapper[4809]: I1206 05:57:13.407633 4809 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="f20ec6b1-9565-48d4-b4d7-4bb2b1d72a66" Dec 06 05:57:13 crc kubenswrapper[4809]: I1206 05:57:13.411066 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 06 05:57:13 crc kubenswrapper[4809]: I1206 05:57:13.411091 4809 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="f20ec6b1-9565-48d4-b4d7-4bb2b1d72a66" Dec 06 05:57:25 crc kubenswrapper[4809]: I1206 05:57:25.493219 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ggvcl"] Dec 06 05:57:25 crc kubenswrapper[4809]: I1206 05:57:25.493983 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-ggvcl" podUID="49fef68f-df3c-4bb7-80cf-f0f651797762" containerName="controller-manager" containerID="cri-o://b4f1ca19e4f9be63b5be9830ba9302245bb78083537d87f172a89ffc807173d9" gracePeriod=30 Dec 06 05:57:25 crc kubenswrapper[4809]: I1206 05:57:25.606620 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-nff6m"] Dec 06 05:57:25 crc kubenswrapper[4809]: I1206 05:57:25.606837 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nff6m" podUID="45f3dc48-d71b-47f8-8567-90df740103b5" containerName="route-controller-manager" containerID="cri-o://4507cc1b69f02c2fd4dcfe1dbeee2ae4957f6de7013d22737bc865859b71ae48" gracePeriod=30 Dec 06 05:57:25 crc kubenswrapper[4809]: I1206 05:57:25.838713 4809 generic.go:334] "Generic (PLEG): container finished" podID="49fef68f-df3c-4bb7-80cf-f0f651797762" containerID="b4f1ca19e4f9be63b5be9830ba9302245bb78083537d87f172a89ffc807173d9" exitCode=0 Dec 06 05:57:25 crc kubenswrapper[4809]: I1206 05:57:25.839033 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-ggvcl" event={"ID":"49fef68f-df3c-4bb7-80cf-f0f651797762","Type":"ContainerDied","Data":"b4f1ca19e4f9be63b5be9830ba9302245bb78083537d87f172a89ffc807173d9"} Dec 06 05:57:25 crc kubenswrapper[4809]: I1206 05:57:25.839064 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-ggvcl" event={"ID":"49fef68f-df3c-4bb7-80cf-f0f651797762","Type":"ContainerDied","Data":"f8ea221c8a98918db5b341707f6a71678c5e760cbe478ec79c6b59ecba3680fb"} Dec 06 05:57:25 crc kubenswrapper[4809]: I1206 05:57:25.839076 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f8ea221c8a98918db5b341707f6a71678c5e760cbe478ec79c6b59ecba3680fb" Dec 06 05:57:25 crc kubenswrapper[4809]: I1206 05:57:25.840355 4809 generic.go:334] "Generic (PLEG): container finished" podID="45f3dc48-d71b-47f8-8567-90df740103b5" containerID="4507cc1b69f02c2fd4dcfe1dbeee2ae4957f6de7013d22737bc865859b71ae48" exitCode=0 Dec 06 05:57:25 crc kubenswrapper[4809]: I1206 05:57:25.840394 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nff6m" event={"ID":"45f3dc48-d71b-47f8-8567-90df740103b5","Type":"ContainerDied","Data":"4507cc1b69f02c2fd4dcfe1dbeee2ae4957f6de7013d22737bc865859b71ae48"} Dec 06 05:57:25 crc kubenswrapper[4809]: I1206 05:57:25.860762 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-ggvcl" Dec 06 05:57:25 crc kubenswrapper[4809]: I1206 05:57:25.908897 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49fef68f-df3c-4bb7-80cf-f0f651797762-config\") pod \"49fef68f-df3c-4bb7-80cf-f0f651797762\" (UID: \"49fef68f-df3c-4bb7-80cf-f0f651797762\") " Dec 06 05:57:25 crc kubenswrapper[4809]: I1206 05:57:25.908984 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/49fef68f-df3c-4bb7-80cf-f0f651797762-client-ca\") pod \"49fef68f-df3c-4bb7-80cf-f0f651797762\" (UID: \"49fef68f-df3c-4bb7-80cf-f0f651797762\") " Dec 06 05:57:25 crc kubenswrapper[4809]: I1206 05:57:25.909046 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zb9vt\" (UniqueName: \"kubernetes.io/projected/49fef68f-df3c-4bb7-80cf-f0f651797762-kube-api-access-zb9vt\") pod \"49fef68f-df3c-4bb7-80cf-f0f651797762\" (UID: \"49fef68f-df3c-4bb7-80cf-f0f651797762\") " Dec 06 05:57:25 crc kubenswrapper[4809]: I1206 05:57:25.909065 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/49fef68f-df3c-4bb7-80cf-f0f651797762-proxy-ca-bundles\") pod \"49fef68f-df3c-4bb7-80cf-f0f651797762\" (UID: \"49fef68f-df3c-4bb7-80cf-f0f651797762\") " Dec 06 05:57:25 crc kubenswrapper[4809]: I1206 05:57:25.909097 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/49fef68f-df3c-4bb7-80cf-f0f651797762-serving-cert\") pod \"49fef68f-df3c-4bb7-80cf-f0f651797762\" (UID: \"49fef68f-df3c-4bb7-80cf-f0f651797762\") " Dec 06 05:57:25 crc kubenswrapper[4809]: I1206 05:57:25.910649 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49fef68f-df3c-4bb7-80cf-f0f651797762-client-ca" (OuterVolumeSpecName: "client-ca") pod "49fef68f-df3c-4bb7-80cf-f0f651797762" (UID: "49fef68f-df3c-4bb7-80cf-f0f651797762"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:57:25 crc kubenswrapper[4809]: I1206 05:57:25.912121 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49fef68f-df3c-4bb7-80cf-f0f651797762-config" (OuterVolumeSpecName: "config") pod "49fef68f-df3c-4bb7-80cf-f0f651797762" (UID: "49fef68f-df3c-4bb7-80cf-f0f651797762"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:57:25 crc kubenswrapper[4809]: I1206 05:57:25.914687 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49fef68f-df3c-4bb7-80cf-f0f651797762-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "49fef68f-df3c-4bb7-80cf-f0f651797762" (UID: "49fef68f-df3c-4bb7-80cf-f0f651797762"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:57:25 crc kubenswrapper[4809]: I1206 05:57:25.928059 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49fef68f-df3c-4bb7-80cf-f0f651797762-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "49fef68f-df3c-4bb7-80cf-f0f651797762" (UID: "49fef68f-df3c-4bb7-80cf-f0f651797762"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:57:25 crc kubenswrapper[4809]: I1206 05:57:25.930337 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49fef68f-df3c-4bb7-80cf-f0f651797762-kube-api-access-zb9vt" (OuterVolumeSpecName: "kube-api-access-zb9vt") pod "49fef68f-df3c-4bb7-80cf-f0f651797762" (UID: "49fef68f-df3c-4bb7-80cf-f0f651797762"). InnerVolumeSpecName "kube-api-access-zb9vt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:57:25 crc kubenswrapper[4809]: I1206 05:57:25.943327 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nff6m" Dec 06 05:57:26 crc kubenswrapper[4809]: I1206 05:57:26.010172 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zb9vt\" (UniqueName: \"kubernetes.io/projected/49fef68f-df3c-4bb7-80cf-f0f651797762-kube-api-access-zb9vt\") on node \"crc\" DevicePath \"\"" Dec 06 05:57:26 crc kubenswrapper[4809]: I1206 05:57:26.010222 4809 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/49fef68f-df3c-4bb7-80cf-f0f651797762-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 06 05:57:26 crc kubenswrapper[4809]: I1206 05:57:26.010236 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/49fef68f-df3c-4bb7-80cf-f0f651797762-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:57:26 crc kubenswrapper[4809]: I1206 05:57:26.010255 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49fef68f-df3c-4bb7-80cf-f0f651797762-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:57:26 crc kubenswrapper[4809]: I1206 05:57:26.010271 4809 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/49fef68f-df3c-4bb7-80cf-f0f651797762-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 05:57:26 crc kubenswrapper[4809]: I1206 05:57:26.110684 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/45f3dc48-d71b-47f8-8567-90df740103b5-client-ca\") pod \"45f3dc48-d71b-47f8-8567-90df740103b5\" (UID: \"45f3dc48-d71b-47f8-8567-90df740103b5\") " Dec 06 05:57:26 crc kubenswrapper[4809]: I1206 05:57:26.110757 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/45f3dc48-d71b-47f8-8567-90df740103b5-serving-cert\") pod \"45f3dc48-d71b-47f8-8567-90df740103b5\" (UID: \"45f3dc48-d71b-47f8-8567-90df740103b5\") " Dec 06 05:57:26 crc kubenswrapper[4809]: I1206 05:57:26.110807 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45f3dc48-d71b-47f8-8567-90df740103b5-config\") pod \"45f3dc48-d71b-47f8-8567-90df740103b5\" (UID: \"45f3dc48-d71b-47f8-8567-90df740103b5\") " Dec 06 05:57:26 crc kubenswrapper[4809]: I1206 05:57:26.110854 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fhld\" (UniqueName: \"kubernetes.io/projected/45f3dc48-d71b-47f8-8567-90df740103b5-kube-api-access-7fhld\") pod \"45f3dc48-d71b-47f8-8567-90df740103b5\" (UID: \"45f3dc48-d71b-47f8-8567-90df740103b5\") " Dec 06 05:57:26 crc kubenswrapper[4809]: I1206 05:57:26.112438 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45f3dc48-d71b-47f8-8567-90df740103b5-client-ca" (OuterVolumeSpecName: "client-ca") pod "45f3dc48-d71b-47f8-8567-90df740103b5" (UID: "45f3dc48-d71b-47f8-8567-90df740103b5"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:57:26 crc kubenswrapper[4809]: I1206 05:57:26.112494 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45f3dc48-d71b-47f8-8567-90df740103b5-config" (OuterVolumeSpecName: "config") pod "45f3dc48-d71b-47f8-8567-90df740103b5" (UID: "45f3dc48-d71b-47f8-8567-90df740103b5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:57:26 crc kubenswrapper[4809]: I1206 05:57:26.114679 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45f3dc48-d71b-47f8-8567-90df740103b5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "45f3dc48-d71b-47f8-8567-90df740103b5" (UID: "45f3dc48-d71b-47f8-8567-90df740103b5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:57:26 crc kubenswrapper[4809]: I1206 05:57:26.114733 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45f3dc48-d71b-47f8-8567-90df740103b5-kube-api-access-7fhld" (OuterVolumeSpecName: "kube-api-access-7fhld") pod "45f3dc48-d71b-47f8-8567-90df740103b5" (UID: "45f3dc48-d71b-47f8-8567-90df740103b5"). InnerVolumeSpecName "kube-api-access-7fhld". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:57:26 crc kubenswrapper[4809]: I1206 05:57:26.190126 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 06 05:57:26 crc kubenswrapper[4809]: I1206 05:57:26.212225 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45f3dc48-d71b-47f8-8567-90df740103b5-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:57:26 crc kubenswrapper[4809]: I1206 05:57:26.212271 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fhld\" (UniqueName: \"kubernetes.io/projected/45f3dc48-d71b-47f8-8567-90df740103b5-kube-api-access-7fhld\") on node \"crc\" DevicePath \"\"" Dec 06 05:57:26 crc kubenswrapper[4809]: I1206 05:57:26.212282 4809 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/45f3dc48-d71b-47f8-8567-90df740103b5-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 05:57:26 crc kubenswrapper[4809]: I1206 05:57:26.212292 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/45f3dc48-d71b-47f8-8567-90df740103b5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:57:26 crc kubenswrapper[4809]: I1206 05:57:26.849603 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-ggvcl" Dec 06 05:57:26 crc kubenswrapper[4809]: I1206 05:57:26.849637 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nff6m" event={"ID":"45f3dc48-d71b-47f8-8567-90df740103b5","Type":"ContainerDied","Data":"9b88cf79838ed43cf246dba0624b73c0b37eea8cf22f6b2c637b1214303990d4"} Dec 06 05:57:26 crc kubenswrapper[4809]: I1206 05:57:26.849721 4809 scope.go:117] "RemoveContainer" containerID="4507cc1b69f02c2fd4dcfe1dbeee2ae4957f6de7013d22737bc865859b71ae48" Dec 06 05:57:26 crc kubenswrapper[4809]: I1206 05:57:26.849609 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nff6m" Dec 06 05:57:26 crc kubenswrapper[4809]: I1206 05:57:26.894170 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-nff6m"] Dec 06 05:57:26 crc kubenswrapper[4809]: I1206 05:57:26.901422 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-nff6m"] Dec 06 05:57:26 crc kubenswrapper[4809]: I1206 05:57:26.905823 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ggvcl"] Dec 06 05:57:26 crc kubenswrapper[4809]: I1206 05:57:26.910356 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ggvcl"] Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.075428 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-774f84cbfb-85q6w"] Dec 06 05:57:27 crc kubenswrapper[4809]: E1206 05:57:27.075897 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45f3dc48-d71b-47f8-8567-90df740103b5" containerName="route-controller-manager" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.075921 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="45f3dc48-d71b-47f8-8567-90df740103b5" containerName="route-controller-manager" Dec 06 05:57:27 crc kubenswrapper[4809]: E1206 05:57:27.075957 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49fef68f-df3c-4bb7-80cf-f0f651797762" containerName="controller-manager" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.075963 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="49fef68f-df3c-4bb7-80cf-f0f651797762" containerName="controller-manager" Dec 06 05:57:27 crc kubenswrapper[4809]: E1206 05:57:27.075999 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.076009 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.076131 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="45f3dc48-d71b-47f8-8567-90df740103b5" containerName="route-controller-manager" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.076153 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="49fef68f-df3c-4bb7-80cf-f0f651797762" containerName="controller-manager" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.076162 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.076743 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6799c5f44c-blghh"] Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.077395 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6799c5f44c-blghh" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.078531 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-774f84cbfb-85q6w" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.080839 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.081855 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.082042 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.084283 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.084615 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.084893 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.085077 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.085200 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.086196 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.086599 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.087120 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.088001 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.088078 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-774f84cbfb-85q6w"] Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.091411 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.093366 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6799c5f44c-blghh"] Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.226482 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33ff7af6-6f4f-42a9-91d9-7536827ccee0-config\") pod \"route-controller-manager-774f84cbfb-85q6w\" (UID: \"33ff7af6-6f4f-42a9-91d9-7536827ccee0\") " pod="openshift-route-controller-manager/route-controller-manager-774f84cbfb-85q6w" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.226559 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258-proxy-ca-bundles\") pod \"controller-manager-6799c5f44c-blghh\" (UID: \"dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258\") " pod="openshift-controller-manager/controller-manager-6799c5f44c-blghh" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.226579 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/33ff7af6-6f4f-42a9-91d9-7536827ccee0-serving-cert\") pod \"route-controller-manager-774f84cbfb-85q6w\" (UID: \"33ff7af6-6f4f-42a9-91d9-7536827ccee0\") " pod="openshift-route-controller-manager/route-controller-manager-774f84cbfb-85q6w" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.226627 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258-config\") pod \"controller-manager-6799c5f44c-blghh\" (UID: \"dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258\") " pod="openshift-controller-manager/controller-manager-6799c5f44c-blghh" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.227124 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/33ff7af6-6f4f-42a9-91d9-7536827ccee0-client-ca\") pod \"route-controller-manager-774f84cbfb-85q6w\" (UID: \"33ff7af6-6f4f-42a9-91d9-7536827ccee0\") " pod="openshift-route-controller-manager/route-controller-manager-774f84cbfb-85q6w" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.227196 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8ss2\" (UniqueName: \"kubernetes.io/projected/dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258-kube-api-access-h8ss2\") pod \"controller-manager-6799c5f44c-blghh\" (UID: \"dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258\") " pod="openshift-controller-manager/controller-manager-6799c5f44c-blghh" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.227256 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258-client-ca\") pod \"controller-manager-6799c5f44c-blghh\" (UID: \"dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258\") " pod="openshift-controller-manager/controller-manager-6799c5f44c-blghh" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.227343 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258-serving-cert\") pod \"controller-manager-6799c5f44c-blghh\" (UID: \"dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258\") " pod="openshift-controller-manager/controller-manager-6799c5f44c-blghh" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.227407 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5tsw\" (UniqueName: \"kubernetes.io/projected/33ff7af6-6f4f-42a9-91d9-7536827ccee0-kube-api-access-b5tsw\") pod \"route-controller-manager-774f84cbfb-85q6w\" (UID: \"33ff7af6-6f4f-42a9-91d9-7536827ccee0\") " pod="openshift-route-controller-manager/route-controller-manager-774f84cbfb-85q6w" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.331770 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33ff7af6-6f4f-42a9-91d9-7536827ccee0-config\") pod \"route-controller-manager-774f84cbfb-85q6w\" (UID: \"33ff7af6-6f4f-42a9-91d9-7536827ccee0\") " pod="openshift-route-controller-manager/route-controller-manager-774f84cbfb-85q6w" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.331885 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258-proxy-ca-bundles\") pod \"controller-manager-6799c5f44c-blghh\" (UID: \"dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258\") " pod="openshift-controller-manager/controller-manager-6799c5f44c-blghh" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.331976 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/33ff7af6-6f4f-42a9-91d9-7536827ccee0-serving-cert\") pod \"route-controller-manager-774f84cbfb-85q6w\" (UID: \"33ff7af6-6f4f-42a9-91d9-7536827ccee0\") " pod="openshift-route-controller-manager/route-controller-manager-774f84cbfb-85q6w" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.332118 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258-config\") pod \"controller-manager-6799c5f44c-blghh\" (UID: \"dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258\") " pod="openshift-controller-manager/controller-manager-6799c5f44c-blghh" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.332210 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/33ff7af6-6f4f-42a9-91d9-7536827ccee0-client-ca\") pod \"route-controller-manager-774f84cbfb-85q6w\" (UID: \"33ff7af6-6f4f-42a9-91d9-7536827ccee0\") " pod="openshift-route-controller-manager/route-controller-manager-774f84cbfb-85q6w" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.332262 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8ss2\" (UniqueName: \"kubernetes.io/projected/dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258-kube-api-access-h8ss2\") pod \"controller-manager-6799c5f44c-blghh\" (UID: \"dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258\") " pod="openshift-controller-manager/controller-manager-6799c5f44c-blghh" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.332301 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258-client-ca\") pod \"controller-manager-6799c5f44c-blghh\" (UID: \"dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258\") " pod="openshift-controller-manager/controller-manager-6799c5f44c-blghh" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.332345 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258-serving-cert\") pod \"controller-manager-6799c5f44c-blghh\" (UID: \"dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258\") " pod="openshift-controller-manager/controller-manager-6799c5f44c-blghh" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.332386 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5tsw\" (UniqueName: \"kubernetes.io/projected/33ff7af6-6f4f-42a9-91d9-7536827ccee0-kube-api-access-b5tsw\") pod \"route-controller-manager-774f84cbfb-85q6w\" (UID: \"33ff7af6-6f4f-42a9-91d9-7536827ccee0\") " pod="openshift-route-controller-manager/route-controller-manager-774f84cbfb-85q6w" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.333279 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258-proxy-ca-bundles\") pod \"controller-manager-6799c5f44c-blghh\" (UID: \"dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258\") " pod="openshift-controller-manager/controller-manager-6799c5f44c-blghh" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.333709 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/33ff7af6-6f4f-42a9-91d9-7536827ccee0-client-ca\") pod \"route-controller-manager-774f84cbfb-85q6w\" (UID: \"33ff7af6-6f4f-42a9-91d9-7536827ccee0\") " pod="openshift-route-controller-manager/route-controller-manager-774f84cbfb-85q6w" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.333739 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258-client-ca\") pod \"controller-manager-6799c5f44c-blghh\" (UID: \"dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258\") " pod="openshift-controller-manager/controller-manager-6799c5f44c-blghh" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.334747 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33ff7af6-6f4f-42a9-91d9-7536827ccee0-config\") pod \"route-controller-manager-774f84cbfb-85q6w\" (UID: \"33ff7af6-6f4f-42a9-91d9-7536827ccee0\") " pod="openshift-route-controller-manager/route-controller-manager-774f84cbfb-85q6w" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.335294 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258-config\") pod \"controller-manager-6799c5f44c-blghh\" (UID: \"dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258\") " pod="openshift-controller-manager/controller-manager-6799c5f44c-blghh" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.338863 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258-serving-cert\") pod \"controller-manager-6799c5f44c-blghh\" (UID: \"dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258\") " pod="openshift-controller-manager/controller-manager-6799c5f44c-blghh" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.349631 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/33ff7af6-6f4f-42a9-91d9-7536827ccee0-serving-cert\") pod \"route-controller-manager-774f84cbfb-85q6w\" (UID: \"33ff7af6-6f4f-42a9-91d9-7536827ccee0\") " pod="openshift-route-controller-manager/route-controller-manager-774f84cbfb-85q6w" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.353711 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5tsw\" (UniqueName: \"kubernetes.io/projected/33ff7af6-6f4f-42a9-91d9-7536827ccee0-kube-api-access-b5tsw\") pod \"route-controller-manager-774f84cbfb-85q6w\" (UID: \"33ff7af6-6f4f-42a9-91d9-7536827ccee0\") " pod="openshift-route-controller-manager/route-controller-manager-774f84cbfb-85q6w" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.358762 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8ss2\" (UniqueName: \"kubernetes.io/projected/dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258-kube-api-access-h8ss2\") pod \"controller-manager-6799c5f44c-blghh\" (UID: \"dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258\") " pod="openshift-controller-manager/controller-manager-6799c5f44c-blghh" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.398357 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45f3dc48-d71b-47f8-8567-90df740103b5" path="/var/lib/kubelet/pods/45f3dc48-d71b-47f8-8567-90df740103b5/volumes" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.399212 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49fef68f-df3c-4bb7-80cf-f0f651797762" path="/var/lib/kubelet/pods/49fef68f-df3c-4bb7-80cf-f0f651797762/volumes" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.401551 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6799c5f44c-blghh" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.415786 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-774f84cbfb-85q6w" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.604397 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6799c5f44c-blghh"] Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.856896 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6799c5f44c-blghh" event={"ID":"dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258","Type":"ContainerStarted","Data":"c616d5099c71763c7b9d107be801d2048f5b7f1850cdbb2ae8e492d980291200"} Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.856954 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6799c5f44c-blghh" event={"ID":"dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258","Type":"ContainerStarted","Data":"77e1b72ff89ca1442ee0049ac40be1a2d6af2a346153bbd6eea5f2468687fedd"} Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.857245 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6799c5f44c-blghh" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.858512 4809 patch_prober.go:28] interesting pod/controller-manager-6799c5f44c-blghh container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.58:8443/healthz\": dial tcp 10.217.0.58:8443: connect: connection refused" start-of-body= Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.858596 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-6799c5f44c-blghh" podUID="dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.58:8443/healthz\": dial tcp 10.217.0.58:8443: connect: connection refused" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.879031 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6799c5f44c-blghh" podStartSLOduration=2.879008681 podStartE2EDuration="2.879008681s" podCreationTimestamp="2025-12-06 05:57:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:57:27.875141954 +0000 UTC m=+372.764124906" watchObservedRunningTime="2025-12-06 05:57:27.879008681 +0000 UTC m=+372.767991623" Dec 06 05:57:27 crc kubenswrapper[4809]: I1206 05:57:27.886929 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-774f84cbfb-85q6w"] Dec 06 05:57:27 crc kubenswrapper[4809]: W1206 05:57:27.889665 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33ff7af6_6f4f_42a9_91d9_7536827ccee0.slice/crio-dde2b53a42a43b1c144042aecf1fa99dcc967e30528db9c4ebc24d2339eeeead WatchSource:0}: Error finding container dde2b53a42a43b1c144042aecf1fa99dcc967e30528db9c4ebc24d2339eeeead: Status 404 returned error can't find the container with id dde2b53a42a43b1c144042aecf1fa99dcc967e30528db9c4ebc24d2339eeeead Dec 06 05:57:28 crc kubenswrapper[4809]: I1206 05:57:28.863852 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-774f84cbfb-85q6w" event={"ID":"33ff7af6-6f4f-42a9-91d9-7536827ccee0","Type":"ContainerStarted","Data":"877819f51490d6a063ee3e89ac166c7061f3e12b438672a96c5eff0f97196a88"} Dec 06 05:57:28 crc kubenswrapper[4809]: I1206 05:57:28.864192 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-774f84cbfb-85q6w" event={"ID":"33ff7af6-6f4f-42a9-91d9-7536827ccee0","Type":"ContainerStarted","Data":"dde2b53a42a43b1c144042aecf1fa99dcc967e30528db9c4ebc24d2339eeeead"} Dec 06 05:57:28 crc kubenswrapper[4809]: I1206 05:57:28.869035 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6799c5f44c-blghh" Dec 06 05:57:28 crc kubenswrapper[4809]: I1206 05:57:28.900487 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-774f84cbfb-85q6w" podStartSLOduration=3.900467053 podStartE2EDuration="3.900467053s" podCreationTimestamp="2025-12-06 05:57:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:57:28.885158562 +0000 UTC m=+373.774141504" watchObservedRunningTime="2025-12-06 05:57:28.900467053 +0000 UTC m=+373.789449995" Dec 06 05:57:29 crc kubenswrapper[4809]: I1206 05:57:29.870549 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-774f84cbfb-85q6w" Dec 06 05:57:29 crc kubenswrapper[4809]: I1206 05:57:29.874086 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-774f84cbfb-85q6w" Dec 06 05:57:34 crc kubenswrapper[4809]: I1206 05:57:34.496307 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 05:57:34 crc kubenswrapper[4809]: I1206 05:57:34.496910 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 05:57:43 crc kubenswrapper[4809]: I1206 05:57:43.203687 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dvwjq"] Dec 06 05:57:43 crc kubenswrapper[4809]: I1206 05:57:43.204541 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dvwjq" podUID="a83cbcc6-62ac-4814-9fce-eb7ea465bcbf" containerName="registry-server" containerID="cri-o://50137c90f10698b2438f0408ede009051a0a7d8ecb4527e96df10128ea127b02" gracePeriod=2 Dec 06 05:57:43 crc kubenswrapper[4809]: I1206 05:57:43.804913 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hfgk4"] Dec 06 05:57:43 crc kubenswrapper[4809]: I1206 05:57:43.806188 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hfgk4" podUID="e356b822-4e99-42f8-99ba-7aea197f3d95" containerName="registry-server" containerID="cri-o://284f02413d4d6e0a2fe38ddb39686def4715cf410535ea504dc82e53c186ee64" gracePeriod=2 Dec 06 05:57:43 crc kubenswrapper[4809]: I1206 05:57:43.962472 4809 generic.go:334] "Generic (PLEG): container finished" podID="e356b822-4e99-42f8-99ba-7aea197f3d95" containerID="284f02413d4d6e0a2fe38ddb39686def4715cf410535ea504dc82e53c186ee64" exitCode=0 Dec 06 05:57:43 crc kubenswrapper[4809]: I1206 05:57:43.962550 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hfgk4" event={"ID":"e356b822-4e99-42f8-99ba-7aea197f3d95","Type":"ContainerDied","Data":"284f02413d4d6e0a2fe38ddb39686def4715cf410535ea504dc82e53c186ee64"} Dec 06 05:57:43 crc kubenswrapper[4809]: I1206 05:57:43.964049 4809 generic.go:334] "Generic (PLEG): container finished" podID="a83cbcc6-62ac-4814-9fce-eb7ea465bcbf" containerID="50137c90f10698b2438f0408ede009051a0a7d8ecb4527e96df10128ea127b02" exitCode=0 Dec 06 05:57:43 crc kubenswrapper[4809]: I1206 05:57:43.964081 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dvwjq" event={"ID":"a83cbcc6-62ac-4814-9fce-eb7ea465bcbf","Type":"ContainerDied","Data":"50137c90f10698b2438f0408ede009051a0a7d8ecb4527e96df10128ea127b02"} Dec 06 05:57:44 crc kubenswrapper[4809]: I1206 05:57:44.138116 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dvwjq" Dec 06 05:57:44 crc kubenswrapper[4809]: I1206 05:57:44.224765 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hfgk4" Dec 06 05:57:44 crc kubenswrapper[4809]: I1206 05:57:44.256060 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a83cbcc6-62ac-4814-9fce-eb7ea465bcbf-catalog-content\") pod \"a83cbcc6-62ac-4814-9fce-eb7ea465bcbf\" (UID: \"a83cbcc6-62ac-4814-9fce-eb7ea465bcbf\") " Dec 06 05:57:44 crc kubenswrapper[4809]: I1206 05:57:44.256180 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a83cbcc6-62ac-4814-9fce-eb7ea465bcbf-utilities\") pod \"a83cbcc6-62ac-4814-9fce-eb7ea465bcbf\" (UID: \"a83cbcc6-62ac-4814-9fce-eb7ea465bcbf\") " Dec 06 05:57:44 crc kubenswrapper[4809]: I1206 05:57:44.256251 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9b9l\" (UniqueName: \"kubernetes.io/projected/a83cbcc6-62ac-4814-9fce-eb7ea465bcbf-kube-api-access-j9b9l\") pod \"a83cbcc6-62ac-4814-9fce-eb7ea465bcbf\" (UID: \"a83cbcc6-62ac-4814-9fce-eb7ea465bcbf\") " Dec 06 05:57:44 crc kubenswrapper[4809]: I1206 05:57:44.257120 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a83cbcc6-62ac-4814-9fce-eb7ea465bcbf-utilities" (OuterVolumeSpecName: "utilities") pod "a83cbcc6-62ac-4814-9fce-eb7ea465bcbf" (UID: "a83cbcc6-62ac-4814-9fce-eb7ea465bcbf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:57:44 crc kubenswrapper[4809]: I1206 05:57:44.257575 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a83cbcc6-62ac-4814-9fce-eb7ea465bcbf-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 05:57:44 crc kubenswrapper[4809]: I1206 05:57:44.260924 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a83cbcc6-62ac-4814-9fce-eb7ea465bcbf-kube-api-access-j9b9l" (OuterVolumeSpecName: "kube-api-access-j9b9l") pod "a83cbcc6-62ac-4814-9fce-eb7ea465bcbf" (UID: "a83cbcc6-62ac-4814-9fce-eb7ea465bcbf"). InnerVolumeSpecName "kube-api-access-j9b9l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:57:44 crc kubenswrapper[4809]: I1206 05:57:44.304257 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a83cbcc6-62ac-4814-9fce-eb7ea465bcbf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a83cbcc6-62ac-4814-9fce-eb7ea465bcbf" (UID: "a83cbcc6-62ac-4814-9fce-eb7ea465bcbf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:57:44 crc kubenswrapper[4809]: I1206 05:57:44.358323 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5kmb\" (UniqueName: \"kubernetes.io/projected/e356b822-4e99-42f8-99ba-7aea197f3d95-kube-api-access-z5kmb\") pod \"e356b822-4e99-42f8-99ba-7aea197f3d95\" (UID: \"e356b822-4e99-42f8-99ba-7aea197f3d95\") " Dec 06 05:57:44 crc kubenswrapper[4809]: I1206 05:57:44.358658 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e356b822-4e99-42f8-99ba-7aea197f3d95-catalog-content\") pod \"e356b822-4e99-42f8-99ba-7aea197f3d95\" (UID: \"e356b822-4e99-42f8-99ba-7aea197f3d95\") " Dec 06 05:57:44 crc kubenswrapper[4809]: I1206 05:57:44.358816 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e356b822-4e99-42f8-99ba-7aea197f3d95-utilities\") pod \"e356b822-4e99-42f8-99ba-7aea197f3d95\" (UID: \"e356b822-4e99-42f8-99ba-7aea197f3d95\") " Dec 06 05:57:44 crc kubenswrapper[4809]: I1206 05:57:44.359145 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9b9l\" (UniqueName: \"kubernetes.io/projected/a83cbcc6-62ac-4814-9fce-eb7ea465bcbf-kube-api-access-j9b9l\") on node \"crc\" DevicePath \"\"" Dec 06 05:57:44 crc kubenswrapper[4809]: I1206 05:57:44.359164 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a83cbcc6-62ac-4814-9fce-eb7ea465bcbf-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 05:57:44 crc kubenswrapper[4809]: I1206 05:57:44.360015 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e356b822-4e99-42f8-99ba-7aea197f3d95-utilities" (OuterVolumeSpecName: "utilities") pod "e356b822-4e99-42f8-99ba-7aea197f3d95" (UID: "e356b822-4e99-42f8-99ba-7aea197f3d95"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:57:44 crc kubenswrapper[4809]: I1206 05:57:44.368237 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e356b822-4e99-42f8-99ba-7aea197f3d95-kube-api-access-z5kmb" (OuterVolumeSpecName: "kube-api-access-z5kmb") pod "e356b822-4e99-42f8-99ba-7aea197f3d95" (UID: "e356b822-4e99-42f8-99ba-7aea197f3d95"). InnerVolumeSpecName "kube-api-access-z5kmb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:57:44 crc kubenswrapper[4809]: I1206 05:57:44.460585 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e356b822-4e99-42f8-99ba-7aea197f3d95-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 05:57:44 crc kubenswrapper[4809]: I1206 05:57:44.460617 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z5kmb\" (UniqueName: \"kubernetes.io/projected/e356b822-4e99-42f8-99ba-7aea197f3d95-kube-api-access-z5kmb\") on node \"crc\" DevicePath \"\"" Dec 06 05:57:44 crc kubenswrapper[4809]: I1206 05:57:44.482448 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e356b822-4e99-42f8-99ba-7aea197f3d95-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e356b822-4e99-42f8-99ba-7aea197f3d95" (UID: "e356b822-4e99-42f8-99ba-7aea197f3d95"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:57:44 crc kubenswrapper[4809]: I1206 05:57:44.564370 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e356b822-4e99-42f8-99ba-7aea197f3d95-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 05:57:44 crc kubenswrapper[4809]: I1206 05:57:44.973367 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hfgk4" event={"ID":"e356b822-4e99-42f8-99ba-7aea197f3d95","Type":"ContainerDied","Data":"3ad5e7413596dd464ef04ea4692a871bf193a5d69854708021f197793efb26c7"} Dec 06 05:57:44 crc kubenswrapper[4809]: I1206 05:57:44.973642 4809 scope.go:117] "RemoveContainer" containerID="284f02413d4d6e0a2fe38ddb39686def4715cf410535ea504dc82e53c186ee64" Dec 06 05:57:44 crc kubenswrapper[4809]: I1206 05:57:44.973816 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hfgk4" Dec 06 05:57:44 crc kubenswrapper[4809]: I1206 05:57:44.978432 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dvwjq" event={"ID":"a83cbcc6-62ac-4814-9fce-eb7ea465bcbf","Type":"ContainerDied","Data":"c93dd1919eb659fb476b1a42a1897b415fbd0e41fcbbf08b63d2bc95644f037f"} Dec 06 05:57:44 crc kubenswrapper[4809]: I1206 05:57:44.978555 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dvwjq" Dec 06 05:57:44 crc kubenswrapper[4809]: I1206 05:57:44.992245 4809 scope.go:117] "RemoveContainer" containerID="d9effd39b1fa4f05d488767fd74860ce039f4a882d6439ccb3930ab59a749496" Dec 06 05:57:45 crc kubenswrapper[4809]: I1206 05:57:45.006172 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hfgk4"] Dec 06 05:57:45 crc kubenswrapper[4809]: I1206 05:57:45.009772 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hfgk4"] Dec 06 05:57:45 crc kubenswrapper[4809]: I1206 05:57:45.023233 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dvwjq"] Dec 06 05:57:45 crc kubenswrapper[4809]: I1206 05:57:45.027324 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dvwjq"] Dec 06 05:57:45 crc kubenswrapper[4809]: I1206 05:57:45.034332 4809 scope.go:117] "RemoveContainer" containerID="59b3a77acd59b69589a0f06f5fa141f0a8538dc3c3422d7932a033031e3e5d80" Dec 06 05:57:45 crc kubenswrapper[4809]: I1206 05:57:45.050406 4809 scope.go:117] "RemoveContainer" containerID="50137c90f10698b2438f0408ede009051a0a7d8ecb4527e96df10128ea127b02" Dec 06 05:57:45 crc kubenswrapper[4809]: I1206 05:57:45.075242 4809 scope.go:117] "RemoveContainer" containerID="d9cb75f3a92a568740557049d70f009040650c11c5fe919255949f3f9d01b275" Dec 06 05:57:45 crc kubenswrapper[4809]: I1206 05:57:45.094854 4809 scope.go:117] "RemoveContainer" containerID="6c8f901f4b4a3252b4273b5561ce69cc87f1ebacbc0de5a67bd1e9a705eff52f" Dec 06 05:57:45 crc kubenswrapper[4809]: I1206 05:57:45.402921 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a83cbcc6-62ac-4814-9fce-eb7ea465bcbf" path="/var/lib/kubelet/pods/a83cbcc6-62ac-4814-9fce-eb7ea465bcbf/volumes" Dec 06 05:57:45 crc kubenswrapper[4809]: I1206 05:57:45.405075 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e356b822-4e99-42f8-99ba-7aea197f3d95" path="/var/lib/kubelet/pods/e356b822-4e99-42f8-99ba-7aea197f3d95/volumes" Dec 06 05:57:45 crc kubenswrapper[4809]: I1206 05:57:45.507154 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-774f84cbfb-85q6w"] Dec 06 05:57:45 crc kubenswrapper[4809]: I1206 05:57:45.507449 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-774f84cbfb-85q6w" podUID="33ff7af6-6f4f-42a9-91d9-7536827ccee0" containerName="route-controller-manager" containerID="cri-o://877819f51490d6a063ee3e89ac166c7061f3e12b438672a96c5eff0f97196a88" gracePeriod=30 Dec 06 05:57:46 crc kubenswrapper[4809]: I1206 05:57:46.204011 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9zm48"] Dec 06 05:57:46 crc kubenswrapper[4809]: I1206 05:57:46.204824 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9zm48" podUID="c822a806-352f-4b4f-8549-051ff20c4ca2" containerName="registry-server" containerID="cri-o://ee55f33e692c658c851dc535b2c6cf2b3c68342211988109544307f25874ea9e" gracePeriod=2 Dec 06 05:57:46 crc kubenswrapper[4809]: I1206 05:57:46.995129 4809 generic.go:334] "Generic (PLEG): container finished" podID="c822a806-352f-4b4f-8549-051ff20c4ca2" containerID="ee55f33e692c658c851dc535b2c6cf2b3c68342211988109544307f25874ea9e" exitCode=0 Dec 06 05:57:46 crc kubenswrapper[4809]: I1206 05:57:46.995460 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9zm48" event={"ID":"c822a806-352f-4b4f-8549-051ff20c4ca2","Type":"ContainerDied","Data":"ee55f33e692c658c851dc535b2c6cf2b3c68342211988109544307f25874ea9e"} Dec 06 05:57:46 crc kubenswrapper[4809]: I1206 05:57:46.997306 4809 generic.go:334] "Generic (PLEG): container finished" podID="33ff7af6-6f4f-42a9-91d9-7536827ccee0" containerID="877819f51490d6a063ee3e89ac166c7061f3e12b438672a96c5eff0f97196a88" exitCode=0 Dec 06 05:57:46 crc kubenswrapper[4809]: I1206 05:57:46.997331 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-774f84cbfb-85q6w" event={"ID":"33ff7af6-6f4f-42a9-91d9-7536827ccee0","Type":"ContainerDied","Data":"877819f51490d6a063ee3e89ac166c7061f3e12b438672a96c5eff0f97196a88"} Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.060950 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-774f84cbfb-85q6w" Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.084768 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6cd65596c4-9z8ww"] Dec 06 05:57:47 crc kubenswrapper[4809]: E1206 05:57:47.084998 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a83cbcc6-62ac-4814-9fce-eb7ea465bcbf" containerName="extract-utilities" Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.085011 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a83cbcc6-62ac-4814-9fce-eb7ea465bcbf" containerName="extract-utilities" Dec 06 05:57:47 crc kubenswrapper[4809]: E1206 05:57:47.085022 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e356b822-4e99-42f8-99ba-7aea197f3d95" containerName="extract-utilities" Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.085028 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e356b822-4e99-42f8-99ba-7aea197f3d95" containerName="extract-utilities" Dec 06 05:57:47 crc kubenswrapper[4809]: E1206 05:57:47.085036 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e356b822-4e99-42f8-99ba-7aea197f3d95" containerName="extract-content" Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.085042 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e356b822-4e99-42f8-99ba-7aea197f3d95" containerName="extract-content" Dec 06 05:57:47 crc kubenswrapper[4809]: E1206 05:57:47.085052 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a83cbcc6-62ac-4814-9fce-eb7ea465bcbf" containerName="registry-server" Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.085057 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a83cbcc6-62ac-4814-9fce-eb7ea465bcbf" containerName="registry-server" Dec 06 05:57:47 crc kubenswrapper[4809]: E1206 05:57:47.085066 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33ff7af6-6f4f-42a9-91d9-7536827ccee0" containerName="route-controller-manager" Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.085071 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="33ff7af6-6f4f-42a9-91d9-7536827ccee0" containerName="route-controller-manager" Dec 06 05:57:47 crc kubenswrapper[4809]: E1206 05:57:47.085085 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e356b822-4e99-42f8-99ba-7aea197f3d95" containerName="registry-server" Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.085091 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e356b822-4e99-42f8-99ba-7aea197f3d95" containerName="registry-server" Dec 06 05:57:47 crc kubenswrapper[4809]: E1206 05:57:47.085099 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a83cbcc6-62ac-4814-9fce-eb7ea465bcbf" containerName="extract-content" Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.085105 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a83cbcc6-62ac-4814-9fce-eb7ea465bcbf" containerName="extract-content" Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.085200 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="e356b822-4e99-42f8-99ba-7aea197f3d95" containerName="registry-server" Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.085220 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="a83cbcc6-62ac-4814-9fce-eb7ea465bcbf" containerName="registry-server" Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.085237 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="33ff7af6-6f4f-42a9-91d9-7536827ccee0" containerName="route-controller-manager" Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.085594 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6cd65596c4-9z8ww" Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.100126 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6cd65596c4-9z8ww"] Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.192385 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33ff7af6-6f4f-42a9-91d9-7536827ccee0-config\") pod \"33ff7af6-6f4f-42a9-91d9-7536827ccee0\" (UID: \"33ff7af6-6f4f-42a9-91d9-7536827ccee0\") " Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.192486 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/33ff7af6-6f4f-42a9-91d9-7536827ccee0-client-ca\") pod \"33ff7af6-6f4f-42a9-91d9-7536827ccee0\" (UID: \"33ff7af6-6f4f-42a9-91d9-7536827ccee0\") " Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.192819 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/33ff7af6-6f4f-42a9-91d9-7536827ccee0-serving-cert\") pod \"33ff7af6-6f4f-42a9-91d9-7536827ccee0\" (UID: \"33ff7af6-6f4f-42a9-91d9-7536827ccee0\") " Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.192900 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5tsw\" (UniqueName: \"kubernetes.io/projected/33ff7af6-6f4f-42a9-91d9-7536827ccee0-kube-api-access-b5tsw\") pod \"33ff7af6-6f4f-42a9-91d9-7536827ccee0\" (UID: \"33ff7af6-6f4f-42a9-91d9-7536827ccee0\") " Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.193190 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8e47b3dc-260e-46a6-9ce2-2e4f4a8a5977-client-ca\") pod \"route-controller-manager-6cd65596c4-9z8ww\" (UID: \"8e47b3dc-260e-46a6-9ce2-2e4f4a8a5977\") " pod="openshift-route-controller-manager/route-controller-manager-6cd65596c4-9z8ww" Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.193229 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e47b3dc-260e-46a6-9ce2-2e4f4a8a5977-serving-cert\") pod \"route-controller-manager-6cd65596c4-9z8ww\" (UID: \"8e47b3dc-260e-46a6-9ce2-2e4f4a8a5977\") " pod="openshift-route-controller-manager/route-controller-manager-6cd65596c4-9z8ww" Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.193261 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e47b3dc-260e-46a6-9ce2-2e4f4a8a5977-config\") pod \"route-controller-manager-6cd65596c4-9z8ww\" (UID: \"8e47b3dc-260e-46a6-9ce2-2e4f4a8a5977\") " pod="openshift-route-controller-manager/route-controller-manager-6cd65596c4-9z8ww" Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.193305 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmhqk\" (UniqueName: \"kubernetes.io/projected/8e47b3dc-260e-46a6-9ce2-2e4f4a8a5977-kube-api-access-pmhqk\") pod \"route-controller-manager-6cd65596c4-9z8ww\" (UID: \"8e47b3dc-260e-46a6-9ce2-2e4f4a8a5977\") " pod="openshift-route-controller-manager/route-controller-manager-6cd65596c4-9z8ww" Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.193541 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33ff7af6-6f4f-42a9-91d9-7536827ccee0-client-ca" (OuterVolumeSpecName: "client-ca") pod "33ff7af6-6f4f-42a9-91d9-7536827ccee0" (UID: "33ff7af6-6f4f-42a9-91d9-7536827ccee0"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.193554 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33ff7af6-6f4f-42a9-91d9-7536827ccee0-config" (OuterVolumeSpecName: "config") pod "33ff7af6-6f4f-42a9-91d9-7536827ccee0" (UID: "33ff7af6-6f4f-42a9-91d9-7536827ccee0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.198473 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33ff7af6-6f4f-42a9-91d9-7536827ccee0-kube-api-access-b5tsw" (OuterVolumeSpecName: "kube-api-access-b5tsw") pod "33ff7af6-6f4f-42a9-91d9-7536827ccee0" (UID: "33ff7af6-6f4f-42a9-91d9-7536827ccee0"). InnerVolumeSpecName "kube-api-access-b5tsw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.199687 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33ff7af6-6f4f-42a9-91d9-7536827ccee0-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "33ff7af6-6f4f-42a9-91d9-7536827ccee0" (UID: "33ff7af6-6f4f-42a9-91d9-7536827ccee0"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.294968 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8e47b3dc-260e-46a6-9ce2-2e4f4a8a5977-client-ca\") pod \"route-controller-manager-6cd65596c4-9z8ww\" (UID: \"8e47b3dc-260e-46a6-9ce2-2e4f4a8a5977\") " pod="openshift-route-controller-manager/route-controller-manager-6cd65596c4-9z8ww" Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.295049 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e47b3dc-260e-46a6-9ce2-2e4f4a8a5977-serving-cert\") pod \"route-controller-manager-6cd65596c4-9z8ww\" (UID: \"8e47b3dc-260e-46a6-9ce2-2e4f4a8a5977\") " pod="openshift-route-controller-manager/route-controller-manager-6cd65596c4-9z8ww" Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.295080 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e47b3dc-260e-46a6-9ce2-2e4f4a8a5977-config\") pod \"route-controller-manager-6cd65596c4-9z8ww\" (UID: \"8e47b3dc-260e-46a6-9ce2-2e4f4a8a5977\") " pod="openshift-route-controller-manager/route-controller-manager-6cd65596c4-9z8ww" Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.295122 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmhqk\" (UniqueName: \"kubernetes.io/projected/8e47b3dc-260e-46a6-9ce2-2e4f4a8a5977-kube-api-access-pmhqk\") pod \"route-controller-manager-6cd65596c4-9z8ww\" (UID: \"8e47b3dc-260e-46a6-9ce2-2e4f4a8a5977\") " pod="openshift-route-controller-manager/route-controller-manager-6cd65596c4-9z8ww" Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.295164 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33ff7af6-6f4f-42a9-91d9-7536827ccee0-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.295180 4809 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/33ff7af6-6f4f-42a9-91d9-7536827ccee0-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.295192 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/33ff7af6-6f4f-42a9-91d9-7536827ccee0-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.295203 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5tsw\" (UniqueName: \"kubernetes.io/projected/33ff7af6-6f4f-42a9-91d9-7536827ccee0-kube-api-access-b5tsw\") on node \"crc\" DevicePath \"\"" Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.297231 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e47b3dc-260e-46a6-9ce2-2e4f4a8a5977-config\") pod \"route-controller-manager-6cd65596c4-9z8ww\" (UID: \"8e47b3dc-260e-46a6-9ce2-2e4f4a8a5977\") " pod="openshift-route-controller-manager/route-controller-manager-6cd65596c4-9z8ww" Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.307174 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8e47b3dc-260e-46a6-9ce2-2e4f4a8a5977-client-ca\") pod \"route-controller-manager-6cd65596c4-9z8ww\" (UID: \"8e47b3dc-260e-46a6-9ce2-2e4f4a8a5977\") " pod="openshift-route-controller-manager/route-controller-manager-6cd65596c4-9z8ww" Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.307887 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e47b3dc-260e-46a6-9ce2-2e4f4a8a5977-serving-cert\") pod \"route-controller-manager-6cd65596c4-9z8ww\" (UID: \"8e47b3dc-260e-46a6-9ce2-2e4f4a8a5977\") " pod="openshift-route-controller-manager/route-controller-manager-6cd65596c4-9z8ww" Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.310818 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmhqk\" (UniqueName: \"kubernetes.io/projected/8e47b3dc-260e-46a6-9ce2-2e4f4a8a5977-kube-api-access-pmhqk\") pod \"route-controller-manager-6cd65596c4-9z8ww\" (UID: \"8e47b3dc-260e-46a6-9ce2-2e4f4a8a5977\") " pod="openshift-route-controller-manager/route-controller-manager-6cd65596c4-9z8ww" Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.400803 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6cd65596c4-9z8ww" Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.503529 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9zm48" Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.598123 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c822a806-352f-4b4f-8549-051ff20c4ca2-utilities\") pod \"c822a806-352f-4b4f-8549-051ff20c4ca2\" (UID: \"c822a806-352f-4b4f-8549-051ff20c4ca2\") " Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.598170 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7f6b\" (UniqueName: \"kubernetes.io/projected/c822a806-352f-4b4f-8549-051ff20c4ca2-kube-api-access-p7f6b\") pod \"c822a806-352f-4b4f-8549-051ff20c4ca2\" (UID: \"c822a806-352f-4b4f-8549-051ff20c4ca2\") " Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.598305 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c822a806-352f-4b4f-8549-051ff20c4ca2-catalog-content\") pod \"c822a806-352f-4b4f-8549-051ff20c4ca2\" (UID: \"c822a806-352f-4b4f-8549-051ff20c4ca2\") " Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.599128 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c822a806-352f-4b4f-8549-051ff20c4ca2-utilities" (OuterVolumeSpecName: "utilities") pod "c822a806-352f-4b4f-8549-051ff20c4ca2" (UID: "c822a806-352f-4b4f-8549-051ff20c4ca2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.603648 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c822a806-352f-4b4f-8549-051ff20c4ca2-kube-api-access-p7f6b" (OuterVolumeSpecName: "kube-api-access-p7f6b") pod "c822a806-352f-4b4f-8549-051ff20c4ca2" (UID: "c822a806-352f-4b4f-8549-051ff20c4ca2"). InnerVolumeSpecName "kube-api-access-p7f6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.655135 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c822a806-352f-4b4f-8549-051ff20c4ca2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c822a806-352f-4b4f-8549-051ff20c4ca2" (UID: "c822a806-352f-4b4f-8549-051ff20c4ca2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.700121 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c822a806-352f-4b4f-8549-051ff20c4ca2-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.700169 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7f6b\" (UniqueName: \"kubernetes.io/projected/c822a806-352f-4b4f-8549-051ff20c4ca2-kube-api-access-p7f6b\") on node \"crc\" DevicePath \"\"" Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.700192 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c822a806-352f-4b4f-8549-051ff20c4ca2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 05:57:47 crc kubenswrapper[4809]: I1206 05:57:47.809100 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6cd65596c4-9z8ww"] Dec 06 05:57:47 crc kubenswrapper[4809]: W1206 05:57:47.810798 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8e47b3dc_260e_46a6_9ce2_2e4f4a8a5977.slice/crio-74a368d30a77b6f082a982d96695a2afef4e52fd96e62abe91c179b663ee210f WatchSource:0}: Error finding container 74a368d30a77b6f082a982d96695a2afef4e52fd96e62abe91c179b663ee210f: Status 404 returned error can't find the container with id 74a368d30a77b6f082a982d96695a2afef4e52fd96e62abe91c179b663ee210f Dec 06 05:57:48 crc kubenswrapper[4809]: I1206 05:57:48.007599 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-774f84cbfb-85q6w" event={"ID":"33ff7af6-6f4f-42a9-91d9-7536827ccee0","Type":"ContainerDied","Data":"dde2b53a42a43b1c144042aecf1fa99dcc967e30528db9c4ebc24d2339eeeead"} Dec 06 05:57:48 crc kubenswrapper[4809]: I1206 05:57:48.007657 4809 scope.go:117] "RemoveContainer" containerID="877819f51490d6a063ee3e89ac166c7061f3e12b438672a96c5eff0f97196a88" Dec 06 05:57:48 crc kubenswrapper[4809]: I1206 05:57:48.007607 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-774f84cbfb-85q6w" Dec 06 05:57:48 crc kubenswrapper[4809]: I1206 05:57:48.010948 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6cd65596c4-9z8ww" event={"ID":"8e47b3dc-260e-46a6-9ce2-2e4f4a8a5977","Type":"ContainerStarted","Data":"74a368d30a77b6f082a982d96695a2afef4e52fd96e62abe91c179b663ee210f"} Dec 06 05:57:48 crc kubenswrapper[4809]: I1206 05:57:48.027782 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9zm48" Dec 06 05:57:48 crc kubenswrapper[4809]: I1206 05:57:48.027753 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9zm48" event={"ID":"c822a806-352f-4b4f-8549-051ff20c4ca2","Type":"ContainerDied","Data":"2f5eb5c5202712e416c3fec6986590204363bf55169b18391542dc759ee946f8"} Dec 06 05:57:48 crc kubenswrapper[4809]: I1206 05:57:48.035359 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-774f84cbfb-85q6w"] Dec 06 05:57:48 crc kubenswrapper[4809]: I1206 05:57:48.038157 4809 scope.go:117] "RemoveContainer" containerID="ee55f33e692c658c851dc535b2c6cf2b3c68342211988109544307f25874ea9e" Dec 06 05:57:48 crc kubenswrapper[4809]: I1206 05:57:48.042975 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-774f84cbfb-85q6w"] Dec 06 05:57:48 crc kubenswrapper[4809]: I1206 05:57:48.052766 4809 scope.go:117] "RemoveContainer" containerID="f9d032773c1b7216ba187e731f469d9970a1224cf425291d216f7e609ce79973" Dec 06 05:57:48 crc kubenswrapper[4809]: I1206 05:57:48.063756 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9zm48"] Dec 06 05:57:48 crc kubenswrapper[4809]: I1206 05:57:48.068058 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9zm48"] Dec 06 05:57:48 crc kubenswrapper[4809]: I1206 05:57:48.091369 4809 scope.go:117] "RemoveContainer" containerID="3ba2cbb88fc0c5d32e76c8b1b832077c55033f367ed4a3e955e7122340be0961" Dec 06 05:57:49 crc kubenswrapper[4809]: I1206 05:57:49.037813 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6cd65596c4-9z8ww" event={"ID":"8e47b3dc-260e-46a6-9ce2-2e4f4a8a5977","Type":"ContainerStarted","Data":"e8bf0abdd861a983256a02efb092adaf0e0507936b90854d4d6e3af607e05d32"} Dec 06 05:57:49 crc kubenswrapper[4809]: I1206 05:57:49.038123 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6cd65596c4-9z8ww" Dec 06 05:57:49 crc kubenswrapper[4809]: I1206 05:57:49.044479 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6cd65596c4-9z8ww" Dec 06 05:57:49 crc kubenswrapper[4809]: I1206 05:57:49.055149 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6cd65596c4-9z8ww" podStartSLOduration=4.05512816 podStartE2EDuration="4.05512816s" podCreationTimestamp="2025-12-06 05:57:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:57:49.054098783 +0000 UTC m=+393.943081725" watchObservedRunningTime="2025-12-06 05:57:49.05512816 +0000 UTC m=+393.944111132" Dec 06 05:57:49 crc kubenswrapper[4809]: I1206 05:57:49.396186 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33ff7af6-6f4f-42a9-91d9-7536827ccee0" path="/var/lib/kubelet/pods/33ff7af6-6f4f-42a9-91d9-7536827ccee0/volumes" Dec 06 05:57:49 crc kubenswrapper[4809]: I1206 05:57:49.396755 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c822a806-352f-4b4f-8549-051ff20c4ca2" path="/var/lib/kubelet/pods/c822a806-352f-4b4f-8549-051ff20c4ca2/volumes" Dec 06 05:58:04 crc kubenswrapper[4809]: I1206 05:58:04.497104 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 05:58:04 crc kubenswrapper[4809]: I1206 05:58:04.497826 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 05:58:05 crc kubenswrapper[4809]: I1206 05:58:05.505880 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6799c5f44c-blghh"] Dec 06 05:58:05 crc kubenswrapper[4809]: I1206 05:58:05.506539 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6799c5f44c-blghh" podUID="dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258" containerName="controller-manager" containerID="cri-o://c616d5099c71763c7b9d107be801d2048f5b7f1850cdbb2ae8e492d980291200" gracePeriod=30 Dec 06 05:58:06 crc kubenswrapper[4809]: I1206 05:58:06.143789 4809 generic.go:334] "Generic (PLEG): container finished" podID="dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258" containerID="c616d5099c71763c7b9d107be801d2048f5b7f1850cdbb2ae8e492d980291200" exitCode=0 Dec 06 05:58:06 crc kubenswrapper[4809]: I1206 05:58:06.143836 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6799c5f44c-blghh" event={"ID":"dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258","Type":"ContainerDied","Data":"c616d5099c71763c7b9d107be801d2048f5b7f1850cdbb2ae8e492d980291200"} Dec 06 05:58:06 crc kubenswrapper[4809]: I1206 05:58:06.431168 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6799c5f44c-blghh" Dec 06 05:58:06 crc kubenswrapper[4809]: I1206 05:58:06.555204 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258-proxy-ca-bundles\") pod \"dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258\" (UID: \"dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258\") " Dec 06 05:58:06 crc kubenswrapper[4809]: I1206 05:58:06.555265 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258-client-ca\") pod \"dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258\" (UID: \"dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258\") " Dec 06 05:58:06 crc kubenswrapper[4809]: I1206 05:58:06.555324 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258-serving-cert\") pod \"dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258\" (UID: \"dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258\") " Dec 06 05:58:06 crc kubenswrapper[4809]: I1206 05:58:06.555387 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258-config\") pod \"dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258\" (UID: \"dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258\") " Dec 06 05:58:06 crc kubenswrapper[4809]: I1206 05:58:06.555439 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h8ss2\" (UniqueName: \"kubernetes.io/projected/dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258-kube-api-access-h8ss2\") pod \"dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258\" (UID: \"dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258\") " Dec 06 05:58:06 crc kubenswrapper[4809]: I1206 05:58:06.556098 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258" (UID: "dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:58:06 crc kubenswrapper[4809]: I1206 05:58:06.556342 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258-client-ca" (OuterVolumeSpecName: "client-ca") pod "dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258" (UID: "dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:58:06 crc kubenswrapper[4809]: I1206 05:58:06.556550 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258-config" (OuterVolumeSpecName: "config") pod "dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258" (UID: "dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:58:06 crc kubenswrapper[4809]: I1206 05:58:06.560177 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258" (UID: "dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:58:06 crc kubenswrapper[4809]: I1206 05:58:06.562297 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258-kube-api-access-h8ss2" (OuterVolumeSpecName: "kube-api-access-h8ss2") pod "dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258" (UID: "dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258"). InnerVolumeSpecName "kube-api-access-h8ss2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:58:06 crc kubenswrapper[4809]: I1206 05:58:06.658149 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h8ss2\" (UniqueName: \"kubernetes.io/projected/dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258-kube-api-access-h8ss2\") on node \"crc\" DevicePath \"\"" Dec 06 05:58:06 crc kubenswrapper[4809]: I1206 05:58:06.658227 4809 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 06 05:58:06 crc kubenswrapper[4809]: I1206 05:58:06.658256 4809 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 05:58:06 crc kubenswrapper[4809]: I1206 05:58:06.658281 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:58:06 crc kubenswrapper[4809]: I1206 05:58:06.658305 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:58:07 crc kubenswrapper[4809]: I1206 05:58:07.091622 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7f967c4b75-bhj49"] Dec 06 05:58:07 crc kubenswrapper[4809]: E1206 05:58:07.091819 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c822a806-352f-4b4f-8549-051ff20c4ca2" containerName="registry-server" Dec 06 05:58:07 crc kubenswrapper[4809]: I1206 05:58:07.091830 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c822a806-352f-4b4f-8549-051ff20c4ca2" containerName="registry-server" Dec 06 05:58:07 crc kubenswrapper[4809]: E1206 05:58:07.091842 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c822a806-352f-4b4f-8549-051ff20c4ca2" containerName="extract-utilities" Dec 06 05:58:07 crc kubenswrapper[4809]: I1206 05:58:07.091848 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c822a806-352f-4b4f-8549-051ff20c4ca2" containerName="extract-utilities" Dec 06 05:58:07 crc kubenswrapper[4809]: E1206 05:58:07.091856 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258" containerName="controller-manager" Dec 06 05:58:07 crc kubenswrapper[4809]: I1206 05:58:07.091862 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258" containerName="controller-manager" Dec 06 05:58:07 crc kubenswrapper[4809]: E1206 05:58:07.091871 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c822a806-352f-4b4f-8549-051ff20c4ca2" containerName="extract-content" Dec 06 05:58:07 crc kubenswrapper[4809]: I1206 05:58:07.091877 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c822a806-352f-4b4f-8549-051ff20c4ca2" containerName="extract-content" Dec 06 05:58:07 crc kubenswrapper[4809]: I1206 05:58:07.091980 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258" containerName="controller-manager" Dec 06 05:58:07 crc kubenswrapper[4809]: I1206 05:58:07.091993 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="c822a806-352f-4b4f-8549-051ff20c4ca2" containerName="registry-server" Dec 06 05:58:07 crc kubenswrapper[4809]: I1206 05:58:07.092329 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7f967c4b75-bhj49" Dec 06 05:58:07 crc kubenswrapper[4809]: I1206 05:58:07.102826 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7f967c4b75-bhj49"] Dec 06 05:58:07 crc kubenswrapper[4809]: I1206 05:58:07.150017 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6799c5f44c-blghh" event={"ID":"dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258","Type":"ContainerDied","Data":"77e1b72ff89ca1442ee0049ac40be1a2d6af2a346153bbd6eea5f2468687fedd"} Dec 06 05:58:07 crc kubenswrapper[4809]: I1206 05:58:07.150078 4809 scope.go:117] "RemoveContainer" containerID="c616d5099c71763c7b9d107be801d2048f5b7f1850cdbb2ae8e492d980291200" Dec 06 05:58:07 crc kubenswrapper[4809]: I1206 05:58:07.150084 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6799c5f44c-blghh" Dec 06 05:58:07 crc kubenswrapper[4809]: I1206 05:58:07.175120 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6799c5f44c-blghh"] Dec 06 05:58:07 crc kubenswrapper[4809]: I1206 05:58:07.177756 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6799c5f44c-blghh"] Dec 06 05:58:07 crc kubenswrapper[4809]: I1206 05:58:07.265618 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8fa68b06-fd0e-4a09-97ee-35d77cae7b80-serving-cert\") pod \"controller-manager-7f967c4b75-bhj49\" (UID: \"8fa68b06-fd0e-4a09-97ee-35d77cae7b80\") " pod="openshift-controller-manager/controller-manager-7f967c4b75-bhj49" Dec 06 05:58:07 crc kubenswrapper[4809]: I1206 05:58:07.265670 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fa68b06-fd0e-4a09-97ee-35d77cae7b80-config\") pod \"controller-manager-7f967c4b75-bhj49\" (UID: \"8fa68b06-fd0e-4a09-97ee-35d77cae7b80\") " pod="openshift-controller-manager/controller-manager-7f967c4b75-bhj49" Dec 06 05:58:07 crc kubenswrapper[4809]: I1206 05:58:07.265735 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8fa68b06-fd0e-4a09-97ee-35d77cae7b80-client-ca\") pod \"controller-manager-7f967c4b75-bhj49\" (UID: \"8fa68b06-fd0e-4a09-97ee-35d77cae7b80\") " pod="openshift-controller-manager/controller-manager-7f967c4b75-bhj49" Dec 06 05:58:07 crc kubenswrapper[4809]: I1206 05:58:07.265848 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8fa68b06-fd0e-4a09-97ee-35d77cae7b80-proxy-ca-bundles\") pod \"controller-manager-7f967c4b75-bhj49\" (UID: \"8fa68b06-fd0e-4a09-97ee-35d77cae7b80\") " pod="openshift-controller-manager/controller-manager-7f967c4b75-bhj49" Dec 06 05:58:07 crc kubenswrapper[4809]: I1206 05:58:07.265914 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86t8x\" (UniqueName: \"kubernetes.io/projected/8fa68b06-fd0e-4a09-97ee-35d77cae7b80-kube-api-access-86t8x\") pod \"controller-manager-7f967c4b75-bhj49\" (UID: \"8fa68b06-fd0e-4a09-97ee-35d77cae7b80\") " pod="openshift-controller-manager/controller-manager-7f967c4b75-bhj49" Dec 06 05:58:07 crc kubenswrapper[4809]: I1206 05:58:07.367532 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fa68b06-fd0e-4a09-97ee-35d77cae7b80-config\") pod \"controller-manager-7f967c4b75-bhj49\" (UID: \"8fa68b06-fd0e-4a09-97ee-35d77cae7b80\") " pod="openshift-controller-manager/controller-manager-7f967c4b75-bhj49" Dec 06 05:58:07 crc kubenswrapper[4809]: I1206 05:58:07.367596 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8fa68b06-fd0e-4a09-97ee-35d77cae7b80-client-ca\") pod \"controller-manager-7f967c4b75-bhj49\" (UID: \"8fa68b06-fd0e-4a09-97ee-35d77cae7b80\") " pod="openshift-controller-manager/controller-manager-7f967c4b75-bhj49" Dec 06 05:58:07 crc kubenswrapper[4809]: I1206 05:58:07.367635 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8fa68b06-fd0e-4a09-97ee-35d77cae7b80-proxy-ca-bundles\") pod \"controller-manager-7f967c4b75-bhj49\" (UID: \"8fa68b06-fd0e-4a09-97ee-35d77cae7b80\") " pod="openshift-controller-manager/controller-manager-7f967c4b75-bhj49" Dec 06 05:58:07 crc kubenswrapper[4809]: I1206 05:58:07.367663 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86t8x\" (UniqueName: \"kubernetes.io/projected/8fa68b06-fd0e-4a09-97ee-35d77cae7b80-kube-api-access-86t8x\") pod \"controller-manager-7f967c4b75-bhj49\" (UID: \"8fa68b06-fd0e-4a09-97ee-35d77cae7b80\") " pod="openshift-controller-manager/controller-manager-7f967c4b75-bhj49" Dec 06 05:58:07 crc kubenswrapper[4809]: I1206 05:58:07.367705 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8fa68b06-fd0e-4a09-97ee-35d77cae7b80-serving-cert\") pod \"controller-manager-7f967c4b75-bhj49\" (UID: \"8fa68b06-fd0e-4a09-97ee-35d77cae7b80\") " pod="openshift-controller-manager/controller-manager-7f967c4b75-bhj49" Dec 06 05:58:07 crc kubenswrapper[4809]: I1206 05:58:07.369041 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8fa68b06-fd0e-4a09-97ee-35d77cae7b80-client-ca\") pod \"controller-manager-7f967c4b75-bhj49\" (UID: \"8fa68b06-fd0e-4a09-97ee-35d77cae7b80\") " pod="openshift-controller-manager/controller-manager-7f967c4b75-bhj49" Dec 06 05:58:07 crc kubenswrapper[4809]: I1206 05:58:07.369349 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8fa68b06-fd0e-4a09-97ee-35d77cae7b80-proxy-ca-bundles\") pod \"controller-manager-7f967c4b75-bhj49\" (UID: \"8fa68b06-fd0e-4a09-97ee-35d77cae7b80\") " pod="openshift-controller-manager/controller-manager-7f967c4b75-bhj49" Dec 06 05:58:07 crc kubenswrapper[4809]: I1206 05:58:07.369714 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fa68b06-fd0e-4a09-97ee-35d77cae7b80-config\") pod \"controller-manager-7f967c4b75-bhj49\" (UID: \"8fa68b06-fd0e-4a09-97ee-35d77cae7b80\") " pod="openshift-controller-manager/controller-manager-7f967c4b75-bhj49" Dec 06 05:58:07 crc kubenswrapper[4809]: I1206 05:58:07.371927 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8fa68b06-fd0e-4a09-97ee-35d77cae7b80-serving-cert\") pod \"controller-manager-7f967c4b75-bhj49\" (UID: \"8fa68b06-fd0e-4a09-97ee-35d77cae7b80\") " pod="openshift-controller-manager/controller-manager-7f967c4b75-bhj49" Dec 06 05:58:07 crc kubenswrapper[4809]: I1206 05:58:07.384163 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86t8x\" (UniqueName: \"kubernetes.io/projected/8fa68b06-fd0e-4a09-97ee-35d77cae7b80-kube-api-access-86t8x\") pod \"controller-manager-7f967c4b75-bhj49\" (UID: \"8fa68b06-fd0e-4a09-97ee-35d77cae7b80\") " pod="openshift-controller-manager/controller-manager-7f967c4b75-bhj49" Dec 06 05:58:07 crc kubenswrapper[4809]: I1206 05:58:07.397442 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258" path="/var/lib/kubelet/pods/dd2fd840-1e8a-44c8-8d70-7c9eb6a3a258/volumes" Dec 06 05:58:07 crc kubenswrapper[4809]: I1206 05:58:07.408234 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7f967c4b75-bhj49" Dec 06 05:58:07 crc kubenswrapper[4809]: I1206 05:58:07.619787 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7f967c4b75-bhj49"] Dec 06 05:58:08 crc kubenswrapper[4809]: I1206 05:58:08.157282 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7f967c4b75-bhj49" event={"ID":"8fa68b06-fd0e-4a09-97ee-35d77cae7b80","Type":"ContainerStarted","Data":"c87608a53694d9bbb470df23c18080f6be45b67b57b4ae51a3ed1d85dfa06494"} Dec 06 05:58:08 crc kubenswrapper[4809]: I1206 05:58:08.157537 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7f967c4b75-bhj49" event={"ID":"8fa68b06-fd0e-4a09-97ee-35d77cae7b80","Type":"ContainerStarted","Data":"135a3d32eaf2d90f0ca32f6442de9a41633c3370d5a141a8358af2851b62c758"} Dec 06 05:58:08 crc kubenswrapper[4809]: I1206 05:58:08.157560 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7f967c4b75-bhj49" Dec 06 05:58:08 crc kubenswrapper[4809]: I1206 05:58:08.162148 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7f967c4b75-bhj49" Dec 06 05:58:08 crc kubenswrapper[4809]: I1206 05:58:08.176453 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7f967c4b75-bhj49" podStartSLOduration=3.176426887 podStartE2EDuration="3.176426887s" podCreationTimestamp="2025-12-06 05:58:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:58:08.175517991 +0000 UTC m=+413.064500953" watchObservedRunningTime="2025-12-06 05:58:08.176426887 +0000 UTC m=+413.065409839" Dec 06 05:58:13 crc kubenswrapper[4809]: I1206 05:58:13.535480 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-dgspm"] Dec 06 05:58:13 crc kubenswrapper[4809]: I1206 05:58:13.536700 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-dgspm" Dec 06 05:58:13 crc kubenswrapper[4809]: I1206 05:58:13.556733 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-dgspm"] Dec 06 05:58:13 crc kubenswrapper[4809]: I1206 05:58:13.638073 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/26c71fe2-c0e4-405d-a6c6-2ce933013cba-registry-tls\") pod \"image-registry-66df7c8f76-dgspm\" (UID: \"26c71fe2-c0e4-405d-a6c6-2ce933013cba\") " pod="openshift-image-registry/image-registry-66df7c8f76-dgspm" Dec 06 05:58:13 crc kubenswrapper[4809]: I1206 05:58:13.638121 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/26c71fe2-c0e4-405d-a6c6-2ce933013cba-ca-trust-extracted\") pod \"image-registry-66df7c8f76-dgspm\" (UID: \"26c71fe2-c0e4-405d-a6c6-2ce933013cba\") " pod="openshift-image-registry/image-registry-66df7c8f76-dgspm" Dec 06 05:58:13 crc kubenswrapper[4809]: I1206 05:58:13.638169 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-dgspm\" (UID: \"26c71fe2-c0e4-405d-a6c6-2ce933013cba\") " pod="openshift-image-registry/image-registry-66df7c8f76-dgspm" Dec 06 05:58:13 crc kubenswrapper[4809]: I1206 05:58:13.638216 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/26c71fe2-c0e4-405d-a6c6-2ce933013cba-bound-sa-token\") pod \"image-registry-66df7c8f76-dgspm\" (UID: \"26c71fe2-c0e4-405d-a6c6-2ce933013cba\") " pod="openshift-image-registry/image-registry-66df7c8f76-dgspm" Dec 06 05:58:13 crc kubenswrapper[4809]: I1206 05:58:13.638262 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/26c71fe2-c0e4-405d-a6c6-2ce933013cba-registry-certificates\") pod \"image-registry-66df7c8f76-dgspm\" (UID: \"26c71fe2-c0e4-405d-a6c6-2ce933013cba\") " pod="openshift-image-registry/image-registry-66df7c8f76-dgspm" Dec 06 05:58:13 crc kubenswrapper[4809]: I1206 05:58:13.638295 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/26c71fe2-c0e4-405d-a6c6-2ce933013cba-trusted-ca\") pod \"image-registry-66df7c8f76-dgspm\" (UID: \"26c71fe2-c0e4-405d-a6c6-2ce933013cba\") " pod="openshift-image-registry/image-registry-66df7c8f76-dgspm" Dec 06 05:58:13 crc kubenswrapper[4809]: I1206 05:58:13.638316 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/26c71fe2-c0e4-405d-a6c6-2ce933013cba-installation-pull-secrets\") pod \"image-registry-66df7c8f76-dgspm\" (UID: \"26c71fe2-c0e4-405d-a6c6-2ce933013cba\") " pod="openshift-image-registry/image-registry-66df7c8f76-dgspm" Dec 06 05:58:13 crc kubenswrapper[4809]: I1206 05:58:13.638343 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2ns8\" (UniqueName: \"kubernetes.io/projected/26c71fe2-c0e4-405d-a6c6-2ce933013cba-kube-api-access-z2ns8\") pod \"image-registry-66df7c8f76-dgspm\" (UID: \"26c71fe2-c0e4-405d-a6c6-2ce933013cba\") " pod="openshift-image-registry/image-registry-66df7c8f76-dgspm" Dec 06 05:58:13 crc kubenswrapper[4809]: I1206 05:58:13.664325 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-dgspm\" (UID: \"26c71fe2-c0e4-405d-a6c6-2ce933013cba\") " pod="openshift-image-registry/image-registry-66df7c8f76-dgspm" Dec 06 05:58:13 crc kubenswrapper[4809]: I1206 05:58:13.739057 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/26c71fe2-c0e4-405d-a6c6-2ce933013cba-registry-certificates\") pod \"image-registry-66df7c8f76-dgspm\" (UID: \"26c71fe2-c0e4-405d-a6c6-2ce933013cba\") " pod="openshift-image-registry/image-registry-66df7c8f76-dgspm" Dec 06 05:58:13 crc kubenswrapper[4809]: I1206 05:58:13.739117 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/26c71fe2-c0e4-405d-a6c6-2ce933013cba-trusted-ca\") pod \"image-registry-66df7c8f76-dgspm\" (UID: \"26c71fe2-c0e4-405d-a6c6-2ce933013cba\") " pod="openshift-image-registry/image-registry-66df7c8f76-dgspm" Dec 06 05:58:13 crc kubenswrapper[4809]: I1206 05:58:13.739138 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/26c71fe2-c0e4-405d-a6c6-2ce933013cba-installation-pull-secrets\") pod \"image-registry-66df7c8f76-dgspm\" (UID: \"26c71fe2-c0e4-405d-a6c6-2ce933013cba\") " pod="openshift-image-registry/image-registry-66df7c8f76-dgspm" Dec 06 05:58:13 crc kubenswrapper[4809]: I1206 05:58:13.739157 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2ns8\" (UniqueName: \"kubernetes.io/projected/26c71fe2-c0e4-405d-a6c6-2ce933013cba-kube-api-access-z2ns8\") pod \"image-registry-66df7c8f76-dgspm\" (UID: \"26c71fe2-c0e4-405d-a6c6-2ce933013cba\") " pod="openshift-image-registry/image-registry-66df7c8f76-dgspm" Dec 06 05:58:13 crc kubenswrapper[4809]: I1206 05:58:13.739201 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/26c71fe2-c0e4-405d-a6c6-2ce933013cba-registry-tls\") pod \"image-registry-66df7c8f76-dgspm\" (UID: \"26c71fe2-c0e4-405d-a6c6-2ce933013cba\") " pod="openshift-image-registry/image-registry-66df7c8f76-dgspm" Dec 06 05:58:13 crc kubenswrapper[4809]: I1206 05:58:13.739220 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/26c71fe2-c0e4-405d-a6c6-2ce933013cba-ca-trust-extracted\") pod \"image-registry-66df7c8f76-dgspm\" (UID: \"26c71fe2-c0e4-405d-a6c6-2ce933013cba\") " pod="openshift-image-registry/image-registry-66df7c8f76-dgspm" Dec 06 05:58:13 crc kubenswrapper[4809]: I1206 05:58:13.739265 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/26c71fe2-c0e4-405d-a6c6-2ce933013cba-bound-sa-token\") pod \"image-registry-66df7c8f76-dgspm\" (UID: \"26c71fe2-c0e4-405d-a6c6-2ce933013cba\") " pod="openshift-image-registry/image-registry-66df7c8f76-dgspm" Dec 06 05:58:13 crc kubenswrapper[4809]: I1206 05:58:13.739764 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/26c71fe2-c0e4-405d-a6c6-2ce933013cba-ca-trust-extracted\") pod \"image-registry-66df7c8f76-dgspm\" (UID: \"26c71fe2-c0e4-405d-a6c6-2ce933013cba\") " pod="openshift-image-registry/image-registry-66df7c8f76-dgspm" Dec 06 05:58:13 crc kubenswrapper[4809]: I1206 05:58:13.740379 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/26c71fe2-c0e4-405d-a6c6-2ce933013cba-registry-certificates\") pod \"image-registry-66df7c8f76-dgspm\" (UID: \"26c71fe2-c0e4-405d-a6c6-2ce933013cba\") " pod="openshift-image-registry/image-registry-66df7c8f76-dgspm" Dec 06 05:58:13 crc kubenswrapper[4809]: I1206 05:58:13.740802 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/26c71fe2-c0e4-405d-a6c6-2ce933013cba-trusted-ca\") pod \"image-registry-66df7c8f76-dgspm\" (UID: \"26c71fe2-c0e4-405d-a6c6-2ce933013cba\") " pod="openshift-image-registry/image-registry-66df7c8f76-dgspm" Dec 06 05:58:13 crc kubenswrapper[4809]: I1206 05:58:13.748733 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/26c71fe2-c0e4-405d-a6c6-2ce933013cba-registry-tls\") pod \"image-registry-66df7c8f76-dgspm\" (UID: \"26c71fe2-c0e4-405d-a6c6-2ce933013cba\") " pod="openshift-image-registry/image-registry-66df7c8f76-dgspm" Dec 06 05:58:13 crc kubenswrapper[4809]: I1206 05:58:13.748734 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/26c71fe2-c0e4-405d-a6c6-2ce933013cba-installation-pull-secrets\") pod \"image-registry-66df7c8f76-dgspm\" (UID: \"26c71fe2-c0e4-405d-a6c6-2ce933013cba\") " pod="openshift-image-registry/image-registry-66df7c8f76-dgspm" Dec 06 05:58:13 crc kubenswrapper[4809]: I1206 05:58:13.757648 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2ns8\" (UniqueName: \"kubernetes.io/projected/26c71fe2-c0e4-405d-a6c6-2ce933013cba-kube-api-access-z2ns8\") pod \"image-registry-66df7c8f76-dgspm\" (UID: \"26c71fe2-c0e4-405d-a6c6-2ce933013cba\") " pod="openshift-image-registry/image-registry-66df7c8f76-dgspm" Dec 06 05:58:13 crc kubenswrapper[4809]: I1206 05:58:13.761527 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/26c71fe2-c0e4-405d-a6c6-2ce933013cba-bound-sa-token\") pod \"image-registry-66df7c8f76-dgspm\" (UID: \"26c71fe2-c0e4-405d-a6c6-2ce933013cba\") " pod="openshift-image-registry/image-registry-66df7c8f76-dgspm" Dec 06 05:58:13 crc kubenswrapper[4809]: I1206 05:58:13.854684 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-dgspm" Dec 06 05:58:14 crc kubenswrapper[4809]: I1206 05:58:14.293511 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-dgspm"] Dec 06 05:58:15 crc kubenswrapper[4809]: I1206 05:58:15.195178 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-dgspm" event={"ID":"26c71fe2-c0e4-405d-a6c6-2ce933013cba","Type":"ContainerStarted","Data":"612a3fe42eee0688abbbffeaabcc33ed0e406a5112e17d3dfd8260179c087593"} Dec 06 05:58:15 crc kubenswrapper[4809]: I1206 05:58:15.195582 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-dgspm" Dec 06 05:58:15 crc kubenswrapper[4809]: I1206 05:58:15.195601 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-dgspm" event={"ID":"26c71fe2-c0e4-405d-a6c6-2ce933013cba","Type":"ContainerStarted","Data":"5dea4b6f49ddd6332fb38d7df6e4079b332203d9db7d99314147a1ad1fced321"} Dec 06 05:58:15 crc kubenswrapper[4809]: I1206 05:58:15.230880 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-dgspm" podStartSLOduration=2.230846189 podStartE2EDuration="2.230846189s" podCreationTimestamp="2025-12-06 05:58:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:58:15.226354417 +0000 UTC m=+420.115337369" watchObservedRunningTime="2025-12-06 05:58:15.230846189 +0000 UTC m=+420.119829171" Dec 06 05:58:32 crc kubenswrapper[4809]: I1206 05:58:32.137577 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zk7tm"] Dec 06 05:58:32 crc kubenswrapper[4809]: I1206 05:58:32.138209 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zk7tm" podUID="5e67c2b8-8bd1-466c-98d8-e93d411162ba" containerName="registry-server" containerID="cri-o://20b718263143dd67f4af46e1f388040c1db9cc32b952a8cc6f35f0c6f8039868" gracePeriod=30 Dec 06 05:58:32 crc kubenswrapper[4809]: I1206 05:58:32.145971 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-24qxt"] Dec 06 05:58:32 crc kubenswrapper[4809]: I1206 05:58:32.146198 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-24qxt" podUID="893020b2-fa30-4371-86f9-a585517839ce" containerName="registry-server" containerID="cri-o://6ad127ae57fe37bc2ab09e2d03a74dd79af68385bbfee5069273b6c01f037ccf" gracePeriod=30 Dec 06 05:58:32 crc kubenswrapper[4809]: I1206 05:58:32.156895 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-slt9h"] Dec 06 05:58:32 crc kubenswrapper[4809]: I1206 05:58:32.157147 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-slt9h" podUID="ee96792c-e168-4c58-a477-cd88c14da612" containerName="marketplace-operator" containerID="cri-o://349bce7f833411f0b73b5cfe568b8093ff256296c433415448588ec0b9d28efc" gracePeriod=30 Dec 06 05:58:32 crc kubenswrapper[4809]: E1206 05:58:32.168105 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="20b718263143dd67f4af46e1f388040c1db9cc32b952a8cc6f35f0c6f8039868" cmd=["grpc_health_probe","-addr=:50051"] Dec 06 05:58:32 crc kubenswrapper[4809]: E1206 05:58:32.173091 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="20b718263143dd67f4af46e1f388040c1db9cc32b952a8cc6f35f0c6f8039868" cmd=["grpc_health_probe","-addr=:50051"] Dec 06 05:58:32 crc kubenswrapper[4809]: I1206 05:58:32.176069 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wj2h7"] Dec 06 05:58:32 crc kubenswrapper[4809]: I1206 05:58:32.176315 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wj2h7" podUID="2c4b2f56-0b03-41ae-af23-6192b8bd3d34" containerName="registry-server" containerID="cri-o://69983aab74d142a4d79a6b927b484e1faff5ba399abafd3c6e39c4f4c1805346" gracePeriod=30 Dec 06 05:58:32 crc kubenswrapper[4809]: E1206 05:58:32.176417 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="20b718263143dd67f4af46e1f388040c1db9cc32b952a8cc6f35f0c6f8039868" cmd=["grpc_health_probe","-addr=:50051"] Dec 06 05:58:32 crc kubenswrapper[4809]: E1206 05:58:32.176500 4809 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-marketplace/certified-operators-zk7tm" podUID="5e67c2b8-8bd1-466c-98d8-e93d411162ba" containerName="registry-server" Dec 06 05:58:32 crc kubenswrapper[4809]: I1206 05:58:32.188069 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-r2lg6"] Dec 06 05:58:32 crc kubenswrapper[4809]: I1206 05:58:32.189021 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-r2lg6" Dec 06 05:58:32 crc kubenswrapper[4809]: I1206 05:58:32.191876 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lpb28"] Dec 06 05:58:32 crc kubenswrapper[4809]: I1206 05:58:32.192351 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lpb28" podUID="40a90e94-16ee-42ba-beff-5103476432c1" containerName="registry-server" containerID="cri-o://12e34109bddd8e50e1a581763ffb23958cb523a827d40e9a9dd070c15da134dc" gracePeriod=30 Dec 06 05:58:32 crc kubenswrapper[4809]: I1206 05:58:32.203724 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-r2lg6"] Dec 06 05:58:32 crc kubenswrapper[4809]: I1206 05:58:32.280429 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzbq5\" (UniqueName: \"kubernetes.io/projected/286c8de4-0ea9-43ad-bb4f-970319e0f4d3-kube-api-access-jzbq5\") pod \"marketplace-operator-79b997595-r2lg6\" (UID: \"286c8de4-0ea9-43ad-bb4f-970319e0f4d3\") " pod="openshift-marketplace/marketplace-operator-79b997595-r2lg6" Dec 06 05:58:32 crc kubenswrapper[4809]: I1206 05:58:32.280509 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/286c8de4-0ea9-43ad-bb4f-970319e0f4d3-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-r2lg6\" (UID: \"286c8de4-0ea9-43ad-bb4f-970319e0f4d3\") " pod="openshift-marketplace/marketplace-operator-79b997595-r2lg6" Dec 06 05:58:32 crc kubenswrapper[4809]: I1206 05:58:32.280543 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/286c8de4-0ea9-43ad-bb4f-970319e0f4d3-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-r2lg6\" (UID: \"286c8de4-0ea9-43ad-bb4f-970319e0f4d3\") " pod="openshift-marketplace/marketplace-operator-79b997595-r2lg6" Dec 06 05:58:32 crc kubenswrapper[4809]: I1206 05:58:32.382062 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzbq5\" (UniqueName: \"kubernetes.io/projected/286c8de4-0ea9-43ad-bb4f-970319e0f4d3-kube-api-access-jzbq5\") pod \"marketplace-operator-79b997595-r2lg6\" (UID: \"286c8de4-0ea9-43ad-bb4f-970319e0f4d3\") " pod="openshift-marketplace/marketplace-operator-79b997595-r2lg6" Dec 06 05:58:32 crc kubenswrapper[4809]: I1206 05:58:32.382127 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/286c8de4-0ea9-43ad-bb4f-970319e0f4d3-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-r2lg6\" (UID: \"286c8de4-0ea9-43ad-bb4f-970319e0f4d3\") " pod="openshift-marketplace/marketplace-operator-79b997595-r2lg6" Dec 06 05:58:32 crc kubenswrapper[4809]: I1206 05:58:32.382148 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/286c8de4-0ea9-43ad-bb4f-970319e0f4d3-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-r2lg6\" (UID: \"286c8de4-0ea9-43ad-bb4f-970319e0f4d3\") " pod="openshift-marketplace/marketplace-operator-79b997595-r2lg6" Dec 06 05:58:32 crc kubenswrapper[4809]: I1206 05:58:32.383519 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/286c8de4-0ea9-43ad-bb4f-970319e0f4d3-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-r2lg6\" (UID: \"286c8de4-0ea9-43ad-bb4f-970319e0f4d3\") " pod="openshift-marketplace/marketplace-operator-79b997595-r2lg6" Dec 06 05:58:32 crc kubenswrapper[4809]: I1206 05:58:32.388248 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/286c8de4-0ea9-43ad-bb4f-970319e0f4d3-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-r2lg6\" (UID: \"286c8de4-0ea9-43ad-bb4f-970319e0f4d3\") " pod="openshift-marketplace/marketplace-operator-79b997595-r2lg6" Dec 06 05:58:32 crc kubenswrapper[4809]: I1206 05:58:32.398008 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzbq5\" (UniqueName: \"kubernetes.io/projected/286c8de4-0ea9-43ad-bb4f-970319e0f4d3-kube-api-access-jzbq5\") pod \"marketplace-operator-79b997595-r2lg6\" (UID: \"286c8de4-0ea9-43ad-bb4f-970319e0f4d3\") " pod="openshift-marketplace/marketplace-operator-79b997595-r2lg6" Dec 06 05:58:32 crc kubenswrapper[4809]: I1206 05:58:32.507646 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-r2lg6" Dec 06 05:58:32 crc kubenswrapper[4809]: I1206 05:58:32.908019 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-r2lg6"] Dec 06 05:58:32 crc kubenswrapper[4809]: W1206 05:58:32.938876 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod286c8de4_0ea9_43ad_bb4f_970319e0f4d3.slice/crio-1b2c5f112f428b081b9cef5265ab866a914e7565c1f9007b847bd96e407b3d9d WatchSource:0}: Error finding container 1b2c5f112f428b081b9cef5265ab866a914e7565c1f9007b847bd96e407b3d9d: Status 404 returned error can't find the container with id 1b2c5f112f428b081b9cef5265ab866a914e7565c1f9007b847bd96e407b3d9d Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.220152 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wj2h7" Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.292130 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-24qxt" event={"ID":"893020b2-fa30-4371-86f9-a585517839ce","Type":"ContainerDied","Data":"6ad127ae57fe37bc2ab09e2d03a74dd79af68385bbfee5069273b6c01f037ccf"} Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.292020 4809 generic.go:334] "Generic (PLEG): container finished" podID="893020b2-fa30-4371-86f9-a585517839ce" containerID="6ad127ae57fe37bc2ab09e2d03a74dd79af68385bbfee5069273b6c01f037ccf" exitCode=0 Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.294617 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gglkq\" (UniqueName: \"kubernetes.io/projected/2c4b2f56-0b03-41ae-af23-6192b8bd3d34-kube-api-access-gglkq\") pod \"2c4b2f56-0b03-41ae-af23-6192b8bd3d34\" (UID: \"2c4b2f56-0b03-41ae-af23-6192b8bd3d34\") " Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.294681 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c4b2f56-0b03-41ae-af23-6192b8bd3d34-utilities\") pod \"2c4b2f56-0b03-41ae-af23-6192b8bd3d34\" (UID: \"2c4b2f56-0b03-41ae-af23-6192b8bd3d34\") " Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.294738 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c4b2f56-0b03-41ae-af23-6192b8bd3d34-catalog-content\") pod \"2c4b2f56-0b03-41ae-af23-6192b8bd3d34\" (UID: \"2c4b2f56-0b03-41ae-af23-6192b8bd3d34\") " Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.295623 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c4b2f56-0b03-41ae-af23-6192b8bd3d34-utilities" (OuterVolumeSpecName: "utilities") pod "2c4b2f56-0b03-41ae-af23-6192b8bd3d34" (UID: "2c4b2f56-0b03-41ae-af23-6192b8bd3d34"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.305987 4809 generic.go:334] "Generic (PLEG): container finished" podID="5e67c2b8-8bd1-466c-98d8-e93d411162ba" containerID="20b718263143dd67f4af46e1f388040c1db9cc32b952a8cc6f35f0c6f8039868" exitCode=0 Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.306097 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zk7tm" event={"ID":"5e67c2b8-8bd1-466c-98d8-e93d411162ba","Type":"ContainerDied","Data":"20b718263143dd67f4af46e1f388040c1db9cc32b952a8cc6f35f0c6f8039868"} Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.311839 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c4b2f56-0b03-41ae-af23-6192b8bd3d34-kube-api-access-gglkq" (OuterVolumeSpecName: "kube-api-access-gglkq") pod "2c4b2f56-0b03-41ae-af23-6192b8bd3d34" (UID: "2c4b2f56-0b03-41ae-af23-6192b8bd3d34"). InnerVolumeSpecName "kube-api-access-gglkq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.315317 4809 generic.go:334] "Generic (PLEG): container finished" podID="40a90e94-16ee-42ba-beff-5103476432c1" containerID="12e34109bddd8e50e1a581763ffb23958cb523a827d40e9a9dd070c15da134dc" exitCode=0 Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.315405 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lpb28" event={"ID":"40a90e94-16ee-42ba-beff-5103476432c1","Type":"ContainerDied","Data":"12e34109bddd8e50e1a581763ffb23958cb523a827d40e9a9dd070c15da134dc"} Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.317797 4809 generic.go:334] "Generic (PLEG): container finished" podID="ee96792c-e168-4c58-a477-cd88c14da612" containerID="349bce7f833411f0b73b5cfe568b8093ff256296c433415448588ec0b9d28efc" exitCode=0 Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.318160 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-slt9h" event={"ID":"ee96792c-e168-4c58-a477-cd88c14da612","Type":"ContainerDied","Data":"349bce7f833411f0b73b5cfe568b8093ff256296c433415448588ec0b9d28efc"} Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.323104 4809 generic.go:334] "Generic (PLEG): container finished" podID="2c4b2f56-0b03-41ae-af23-6192b8bd3d34" containerID="69983aab74d142a4d79a6b927b484e1faff5ba399abafd3c6e39c4f4c1805346" exitCode=0 Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.323201 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wj2h7" event={"ID":"2c4b2f56-0b03-41ae-af23-6192b8bd3d34","Type":"ContainerDied","Data":"69983aab74d142a4d79a6b927b484e1faff5ba399abafd3c6e39c4f4c1805346"} Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.323231 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wj2h7" event={"ID":"2c4b2f56-0b03-41ae-af23-6192b8bd3d34","Type":"ContainerDied","Data":"e8b5b9437be705b12df2db7d1f3d605c98cd02d84ae03d3cbdd5a044275dbd78"} Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.323249 4809 scope.go:117] "RemoveContainer" containerID="69983aab74d142a4d79a6b927b484e1faff5ba399abafd3c6e39c4f4c1805346" Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.323393 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wj2h7" Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.337819 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c4b2f56-0b03-41ae-af23-6192b8bd3d34-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2c4b2f56-0b03-41ae-af23-6192b8bd3d34" (UID: "2c4b2f56-0b03-41ae-af23-6192b8bd3d34"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.356291 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-r2lg6" event={"ID":"286c8de4-0ea9-43ad-bb4f-970319e0f4d3","Type":"ContainerStarted","Data":"2d014b49e0d244c39b905b46ea027251fa65dd9a14a898db42f83a7103e4412b"} Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.357575 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-r2lg6" event={"ID":"286c8de4-0ea9-43ad-bb4f-970319e0f4d3","Type":"ContainerStarted","Data":"1b2c5f112f428b081b9cef5265ab866a914e7565c1f9007b847bd96e407b3d9d"} Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.359130 4809 scope.go:117] "RemoveContainer" containerID="fbf084ae1a873d7f4aa9bb837f4ea622d95d00fc7f6da3f8aecf04c74bf90da9" Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.375500 4809 scope.go:117] "RemoveContainer" containerID="53c987bae54781114f4630739e892e70b687516ae3183ac01861392512f94375" Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.394872 4809 scope.go:117] "RemoveContainer" containerID="69983aab74d142a4d79a6b927b484e1faff5ba399abafd3c6e39c4f4c1805346" Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.395601 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gglkq\" (UniqueName: \"kubernetes.io/projected/2c4b2f56-0b03-41ae-af23-6192b8bd3d34-kube-api-access-gglkq\") on node \"crc\" DevicePath \"\"" Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.395624 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c4b2f56-0b03-41ae-af23-6192b8bd3d34-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.395636 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c4b2f56-0b03-41ae-af23-6192b8bd3d34-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 05:58:33 crc kubenswrapper[4809]: E1206 05:58:33.403716 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69983aab74d142a4d79a6b927b484e1faff5ba399abafd3c6e39c4f4c1805346\": container with ID starting with 69983aab74d142a4d79a6b927b484e1faff5ba399abafd3c6e39c4f4c1805346 not found: ID does not exist" containerID="69983aab74d142a4d79a6b927b484e1faff5ba399abafd3c6e39c4f4c1805346" Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.403753 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69983aab74d142a4d79a6b927b484e1faff5ba399abafd3c6e39c4f4c1805346"} err="failed to get container status \"69983aab74d142a4d79a6b927b484e1faff5ba399abafd3c6e39c4f4c1805346\": rpc error: code = NotFound desc = could not find container \"69983aab74d142a4d79a6b927b484e1faff5ba399abafd3c6e39c4f4c1805346\": container with ID starting with 69983aab74d142a4d79a6b927b484e1faff5ba399abafd3c6e39c4f4c1805346 not found: ID does not exist" Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.403783 4809 scope.go:117] "RemoveContainer" containerID="fbf084ae1a873d7f4aa9bb837f4ea622d95d00fc7f6da3f8aecf04c74bf90da9" Dec 06 05:58:33 crc kubenswrapper[4809]: E1206 05:58:33.415800 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fbf084ae1a873d7f4aa9bb837f4ea622d95d00fc7f6da3f8aecf04c74bf90da9\": container with ID starting with fbf084ae1a873d7f4aa9bb837f4ea622d95d00fc7f6da3f8aecf04c74bf90da9 not found: ID does not exist" containerID="fbf084ae1a873d7f4aa9bb837f4ea622d95d00fc7f6da3f8aecf04c74bf90da9" Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.415839 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fbf084ae1a873d7f4aa9bb837f4ea622d95d00fc7f6da3f8aecf04c74bf90da9"} err="failed to get container status \"fbf084ae1a873d7f4aa9bb837f4ea622d95d00fc7f6da3f8aecf04c74bf90da9\": rpc error: code = NotFound desc = could not find container \"fbf084ae1a873d7f4aa9bb837f4ea622d95d00fc7f6da3f8aecf04c74bf90da9\": container with ID starting with fbf084ae1a873d7f4aa9bb837f4ea622d95d00fc7f6da3f8aecf04c74bf90da9 not found: ID does not exist" Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.415871 4809 scope.go:117] "RemoveContainer" containerID="53c987bae54781114f4630739e892e70b687516ae3183ac01861392512f94375" Dec 06 05:58:33 crc kubenswrapper[4809]: E1206 05:58:33.416576 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53c987bae54781114f4630739e892e70b687516ae3183ac01861392512f94375\": container with ID starting with 53c987bae54781114f4630739e892e70b687516ae3183ac01861392512f94375 not found: ID does not exist" containerID="53c987bae54781114f4630739e892e70b687516ae3183ac01861392512f94375" Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.416603 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53c987bae54781114f4630739e892e70b687516ae3183ac01861392512f94375"} err="failed to get container status \"53c987bae54781114f4630739e892e70b687516ae3183ac01861392512f94375\": rpc error: code = NotFound desc = could not find container \"53c987bae54781114f4630739e892e70b687516ae3183ac01861392512f94375\": container with ID starting with 53c987bae54781114f4630739e892e70b687516ae3183ac01861392512f94375 not found: ID does not exist" Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.463888 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-24qxt" Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.482220 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zk7tm" Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.494782 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-slt9h" Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.504839 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lpb28" Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.597324 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ee96792c-e168-4c58-a477-cd88c14da612-marketplace-trusted-ca\") pod \"ee96792c-e168-4c58-a477-cd88c14da612\" (UID: \"ee96792c-e168-4c58-a477-cd88c14da612\") " Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.597387 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qwbq2\" (UniqueName: \"kubernetes.io/projected/5e67c2b8-8bd1-466c-98d8-e93d411162ba-kube-api-access-qwbq2\") pod \"5e67c2b8-8bd1-466c-98d8-e93d411162ba\" (UID: \"5e67c2b8-8bd1-466c-98d8-e93d411162ba\") " Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.597412 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c5q7r\" (UniqueName: \"kubernetes.io/projected/ee96792c-e168-4c58-a477-cd88c14da612-kube-api-access-c5q7r\") pod \"ee96792c-e168-4c58-a477-cd88c14da612\" (UID: \"ee96792c-e168-4c58-a477-cd88c14da612\") " Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.597480 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ee96792c-e168-4c58-a477-cd88c14da612-marketplace-operator-metrics\") pod \"ee96792c-e168-4c58-a477-cd88c14da612\" (UID: \"ee96792c-e168-4c58-a477-cd88c14da612\") " Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.598242 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e67c2b8-8bd1-466c-98d8-e93d411162ba-utilities\") pod \"5e67c2b8-8bd1-466c-98d8-e93d411162ba\" (UID: \"5e67c2b8-8bd1-466c-98d8-e93d411162ba\") " Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.598279 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dssmc\" (UniqueName: \"kubernetes.io/projected/40a90e94-16ee-42ba-beff-5103476432c1-kube-api-access-dssmc\") pod \"40a90e94-16ee-42ba-beff-5103476432c1\" (UID: \"40a90e94-16ee-42ba-beff-5103476432c1\") " Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.598317 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ld72t\" (UniqueName: \"kubernetes.io/projected/893020b2-fa30-4371-86f9-a585517839ce-kube-api-access-ld72t\") pod \"893020b2-fa30-4371-86f9-a585517839ce\" (UID: \"893020b2-fa30-4371-86f9-a585517839ce\") " Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.598344 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e67c2b8-8bd1-466c-98d8-e93d411162ba-catalog-content\") pod \"5e67c2b8-8bd1-466c-98d8-e93d411162ba\" (UID: \"5e67c2b8-8bd1-466c-98d8-e93d411162ba\") " Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.598380 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/893020b2-fa30-4371-86f9-a585517839ce-utilities\") pod \"893020b2-fa30-4371-86f9-a585517839ce\" (UID: \"893020b2-fa30-4371-86f9-a585517839ce\") " Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.598434 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40a90e94-16ee-42ba-beff-5103476432c1-utilities\") pod \"40a90e94-16ee-42ba-beff-5103476432c1\" (UID: \"40a90e94-16ee-42ba-beff-5103476432c1\") " Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.598474 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/893020b2-fa30-4371-86f9-a585517839ce-catalog-content\") pod \"893020b2-fa30-4371-86f9-a585517839ce\" (UID: \"893020b2-fa30-4371-86f9-a585517839ce\") " Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.598503 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40a90e94-16ee-42ba-beff-5103476432c1-catalog-content\") pod \"40a90e94-16ee-42ba-beff-5103476432c1\" (UID: \"40a90e94-16ee-42ba-beff-5103476432c1\") " Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.598773 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee96792c-e168-4c58-a477-cd88c14da612-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "ee96792c-e168-4c58-a477-cd88c14da612" (UID: "ee96792c-e168-4c58-a477-cd88c14da612"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.598960 4809 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ee96792c-e168-4c58-a477-cd88c14da612-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.599761 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40a90e94-16ee-42ba-beff-5103476432c1-utilities" (OuterVolumeSpecName: "utilities") pod "40a90e94-16ee-42ba-beff-5103476432c1" (UID: "40a90e94-16ee-42ba-beff-5103476432c1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.599769 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e67c2b8-8bd1-466c-98d8-e93d411162ba-utilities" (OuterVolumeSpecName: "utilities") pod "5e67c2b8-8bd1-466c-98d8-e93d411162ba" (UID: "5e67c2b8-8bd1-466c-98d8-e93d411162ba"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.599851 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/893020b2-fa30-4371-86f9-a585517839ce-utilities" (OuterVolumeSpecName: "utilities") pod "893020b2-fa30-4371-86f9-a585517839ce" (UID: "893020b2-fa30-4371-86f9-a585517839ce"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.600904 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e67c2b8-8bd1-466c-98d8-e93d411162ba-kube-api-access-qwbq2" (OuterVolumeSpecName: "kube-api-access-qwbq2") pod "5e67c2b8-8bd1-466c-98d8-e93d411162ba" (UID: "5e67c2b8-8bd1-466c-98d8-e93d411162ba"). InnerVolumeSpecName "kube-api-access-qwbq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.601189 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee96792c-e168-4c58-a477-cd88c14da612-kube-api-access-c5q7r" (OuterVolumeSpecName: "kube-api-access-c5q7r") pod "ee96792c-e168-4c58-a477-cd88c14da612" (UID: "ee96792c-e168-4c58-a477-cd88c14da612"). InnerVolumeSpecName "kube-api-access-c5q7r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.601472 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee96792c-e168-4c58-a477-cd88c14da612-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "ee96792c-e168-4c58-a477-cd88c14da612" (UID: "ee96792c-e168-4c58-a477-cd88c14da612"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.602548 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/893020b2-fa30-4371-86f9-a585517839ce-kube-api-access-ld72t" (OuterVolumeSpecName: "kube-api-access-ld72t") pod "893020b2-fa30-4371-86f9-a585517839ce" (UID: "893020b2-fa30-4371-86f9-a585517839ce"). InnerVolumeSpecName "kube-api-access-ld72t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.603102 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40a90e94-16ee-42ba-beff-5103476432c1-kube-api-access-dssmc" (OuterVolumeSpecName: "kube-api-access-dssmc") pod "40a90e94-16ee-42ba-beff-5103476432c1" (UID: "40a90e94-16ee-42ba-beff-5103476432c1"). InnerVolumeSpecName "kube-api-access-dssmc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.639994 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wj2h7"] Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.644276 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wj2h7"] Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.659543 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e67c2b8-8bd1-466c-98d8-e93d411162ba-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5e67c2b8-8bd1-466c-98d8-e93d411162ba" (UID: "5e67c2b8-8bd1-466c-98d8-e93d411162ba"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.662252 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/893020b2-fa30-4371-86f9-a585517839ce-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "893020b2-fa30-4371-86f9-a585517839ce" (UID: "893020b2-fa30-4371-86f9-a585517839ce"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.701517 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c5q7r\" (UniqueName: \"kubernetes.io/projected/ee96792c-e168-4c58-a477-cd88c14da612-kube-api-access-c5q7r\") on node \"crc\" DevicePath \"\"" Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.701561 4809 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ee96792c-e168-4c58-a477-cd88c14da612-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.701572 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e67c2b8-8bd1-466c-98d8-e93d411162ba-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.701583 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dssmc\" (UniqueName: \"kubernetes.io/projected/40a90e94-16ee-42ba-beff-5103476432c1-kube-api-access-dssmc\") on node \"crc\" DevicePath \"\"" Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.701591 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e67c2b8-8bd1-466c-98d8-e93d411162ba-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.701615 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ld72t\" (UniqueName: \"kubernetes.io/projected/893020b2-fa30-4371-86f9-a585517839ce-kube-api-access-ld72t\") on node \"crc\" DevicePath \"\"" Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.701625 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/893020b2-fa30-4371-86f9-a585517839ce-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.701633 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40a90e94-16ee-42ba-beff-5103476432c1-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.701641 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/893020b2-fa30-4371-86f9-a585517839ce-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.701648 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qwbq2\" (UniqueName: \"kubernetes.io/projected/5e67c2b8-8bd1-466c-98d8-e93d411162ba-kube-api-access-qwbq2\") on node \"crc\" DevicePath \"\"" Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.734368 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40a90e94-16ee-42ba-beff-5103476432c1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "40a90e94-16ee-42ba-beff-5103476432c1" (UID: "40a90e94-16ee-42ba-beff-5103476432c1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.802467 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40a90e94-16ee-42ba-beff-5103476432c1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.860149 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-dgspm" Dec 06 05:58:33 crc kubenswrapper[4809]: I1206 05:58:33.914945 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kwzrc"] Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.357556 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zjvtm"] Dec 06 05:58:34 crc kubenswrapper[4809]: E1206 05:58:34.357866 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="893020b2-fa30-4371-86f9-a585517839ce" containerName="extract-content" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.357885 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="893020b2-fa30-4371-86f9-a585517839ce" containerName="extract-content" Dec 06 05:58:34 crc kubenswrapper[4809]: E1206 05:58:34.357906 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c4b2f56-0b03-41ae-af23-6192b8bd3d34" containerName="extract-content" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.357919 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c4b2f56-0b03-41ae-af23-6192b8bd3d34" containerName="extract-content" Dec 06 05:58:34 crc kubenswrapper[4809]: E1206 05:58:34.357972 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40a90e94-16ee-42ba-beff-5103476432c1" containerName="registry-server" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.357987 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="40a90e94-16ee-42ba-beff-5103476432c1" containerName="registry-server" Dec 06 05:58:34 crc kubenswrapper[4809]: E1206 05:58:34.358004 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="893020b2-fa30-4371-86f9-a585517839ce" containerName="extract-utilities" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.358016 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="893020b2-fa30-4371-86f9-a585517839ce" containerName="extract-utilities" Dec 06 05:58:34 crc kubenswrapper[4809]: E1206 05:58:34.358034 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e67c2b8-8bd1-466c-98d8-e93d411162ba" containerName="registry-server" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.358046 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e67c2b8-8bd1-466c-98d8-e93d411162ba" containerName="registry-server" Dec 06 05:58:34 crc kubenswrapper[4809]: E1206 05:58:34.358065 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e67c2b8-8bd1-466c-98d8-e93d411162ba" containerName="extract-utilities" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.358077 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e67c2b8-8bd1-466c-98d8-e93d411162ba" containerName="extract-utilities" Dec 06 05:58:34 crc kubenswrapper[4809]: E1206 05:58:34.358097 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c4b2f56-0b03-41ae-af23-6192b8bd3d34" containerName="registry-server" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.358109 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c4b2f56-0b03-41ae-af23-6192b8bd3d34" containerName="registry-server" Dec 06 05:58:34 crc kubenswrapper[4809]: E1206 05:58:34.358127 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="893020b2-fa30-4371-86f9-a585517839ce" containerName="registry-server" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.358139 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="893020b2-fa30-4371-86f9-a585517839ce" containerName="registry-server" Dec 06 05:58:34 crc kubenswrapper[4809]: E1206 05:58:34.358159 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40a90e94-16ee-42ba-beff-5103476432c1" containerName="extract-content" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.358171 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="40a90e94-16ee-42ba-beff-5103476432c1" containerName="extract-content" Dec 06 05:58:34 crc kubenswrapper[4809]: E1206 05:58:34.358191 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e67c2b8-8bd1-466c-98d8-e93d411162ba" containerName="extract-content" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.358207 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e67c2b8-8bd1-466c-98d8-e93d411162ba" containerName="extract-content" Dec 06 05:58:34 crc kubenswrapper[4809]: E1206 05:58:34.358230 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40a90e94-16ee-42ba-beff-5103476432c1" containerName="extract-utilities" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.358245 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="40a90e94-16ee-42ba-beff-5103476432c1" containerName="extract-utilities" Dec 06 05:58:34 crc kubenswrapper[4809]: E1206 05:58:34.358268 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee96792c-e168-4c58-a477-cd88c14da612" containerName="marketplace-operator" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.358283 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee96792c-e168-4c58-a477-cd88c14da612" containerName="marketplace-operator" Dec 06 05:58:34 crc kubenswrapper[4809]: E1206 05:58:34.358308 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c4b2f56-0b03-41ae-af23-6192b8bd3d34" containerName="extract-utilities" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.358320 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c4b2f56-0b03-41ae-af23-6192b8bd3d34" containerName="extract-utilities" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.358479 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee96792c-e168-4c58-a477-cd88c14da612" containerName="marketplace-operator" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.358496 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="893020b2-fa30-4371-86f9-a585517839ce" containerName="registry-server" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.358513 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e67c2b8-8bd1-466c-98d8-e93d411162ba" containerName="registry-server" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.358532 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="40a90e94-16ee-42ba-beff-5103476432c1" containerName="registry-server" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.358547 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c4b2f56-0b03-41ae-af23-6192b8bd3d34" containerName="registry-server" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.359791 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zjvtm" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.362355 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.363631 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-slt9h" event={"ID":"ee96792c-e168-4c58-a477-cd88c14da612","Type":"ContainerDied","Data":"65d6d119611bca3ae30cc5f8c6cd1480d76ea44c6ad1509ceea897b0e9bbd8b0"} Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.363753 4809 scope.go:117] "RemoveContainer" containerID="349bce7f833411f0b73b5cfe568b8093ff256296c433415448588ec0b9d28efc" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.363902 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-slt9h" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.369379 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-24qxt" event={"ID":"893020b2-fa30-4371-86f9-a585517839ce","Type":"ContainerDied","Data":"7dbcb9c7799bcc1e1ab2ed512b6573baac17ee3f42d68c69e661de5457fd6989"} Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.369450 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-24qxt" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.369680 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zjvtm"] Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.385409 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zk7tm" event={"ID":"5e67c2b8-8bd1-466c-98d8-e93d411162ba","Type":"ContainerDied","Data":"1ae5116ddd876d822729ec2c212a93c39ac1adac9d1ae8936821138f3c6fa2a1"} Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.385524 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zk7tm" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.388714 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lpb28" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.390122 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lpb28" event={"ID":"40a90e94-16ee-42ba-beff-5103476432c1","Type":"ContainerDied","Data":"3417db87bbbc2d4bd8805f87c74161a886bb760bf9cf47491c3c7ffa124afa32"} Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.390185 4809 scope.go:117] "RemoveContainer" containerID="6ad127ae57fe37bc2ab09e2d03a74dd79af68385bbfee5069273b6c01f037ccf" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.390320 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-r2lg6" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.394872 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-r2lg6" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.411750 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0714d252-8cd5-4a99-9050-8e383ad64885-catalog-content\") pod \"redhat-marketplace-zjvtm\" (UID: \"0714d252-8cd5-4a99-9050-8e383ad64885\") " pod="openshift-marketplace/redhat-marketplace-zjvtm" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.411848 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0714d252-8cd5-4a99-9050-8e383ad64885-utilities\") pod \"redhat-marketplace-zjvtm\" (UID: \"0714d252-8cd5-4a99-9050-8e383ad64885\") " pod="openshift-marketplace/redhat-marketplace-zjvtm" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.412079 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cccmm\" (UniqueName: \"kubernetes.io/projected/0714d252-8cd5-4a99-9050-8e383ad64885-kube-api-access-cccmm\") pod \"redhat-marketplace-zjvtm\" (UID: \"0714d252-8cd5-4a99-9050-8e383ad64885\") " pod="openshift-marketplace/redhat-marketplace-zjvtm" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.413856 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-r2lg6" podStartSLOduration=2.413841787 podStartE2EDuration="2.413841787s" podCreationTimestamp="2025-12-06 05:58:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:58:34.410881546 +0000 UTC m=+439.299864488" watchObservedRunningTime="2025-12-06 05:58:34.413841787 +0000 UTC m=+439.302824729" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.416547 4809 scope.go:117] "RemoveContainer" containerID="8cc7549375e24b0c91ac812b6424f92854ecac7d3424bdee33188d237ec03f3c" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.449078 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-24qxt"] Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.449816 4809 scope.go:117] "RemoveContainer" containerID="2ad2d97bb0902f7353810c067603e67f7dc4a0c5339e0a9f536eb8729e07d753" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.476114 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-24qxt"] Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.483172 4809 scope.go:117] "RemoveContainer" containerID="20b718263143dd67f4af46e1f388040c1db9cc32b952a8cc6f35f0c6f8039868" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.488759 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zk7tm"] Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.497134 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.497170 4809 scope.go:117] "RemoveContainer" containerID="ca19224db7d03a2501accfdaa1cf9b66092ad9c0d0ea63def51a482eff77fc61" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.497187 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.497232 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-npms2" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.498265 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2442b241343358fa541b07551125d7c63b82fcdcd98be18afedabbdc8da517a8"} pod="openshift-machine-config-operator/machine-config-daemon-npms2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.498348 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" containerID="cri-o://2442b241343358fa541b07551125d7c63b82fcdcd98be18afedabbdc8da517a8" gracePeriod=600 Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.515454 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0714d252-8cd5-4a99-9050-8e383ad64885-catalog-content\") pod \"redhat-marketplace-zjvtm\" (UID: \"0714d252-8cd5-4a99-9050-8e383ad64885\") " pod="openshift-marketplace/redhat-marketplace-zjvtm" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.515780 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0714d252-8cd5-4a99-9050-8e383ad64885-utilities\") pod \"redhat-marketplace-zjvtm\" (UID: \"0714d252-8cd5-4a99-9050-8e383ad64885\") " pod="openshift-marketplace/redhat-marketplace-zjvtm" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.515912 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cccmm\" (UniqueName: \"kubernetes.io/projected/0714d252-8cd5-4a99-9050-8e383ad64885-kube-api-access-cccmm\") pod \"redhat-marketplace-zjvtm\" (UID: \"0714d252-8cd5-4a99-9050-8e383ad64885\") " pod="openshift-marketplace/redhat-marketplace-zjvtm" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.519594 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0714d252-8cd5-4a99-9050-8e383ad64885-utilities\") pod \"redhat-marketplace-zjvtm\" (UID: \"0714d252-8cd5-4a99-9050-8e383ad64885\") " pod="openshift-marketplace/redhat-marketplace-zjvtm" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.521621 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0714d252-8cd5-4a99-9050-8e383ad64885-catalog-content\") pod \"redhat-marketplace-zjvtm\" (UID: \"0714d252-8cd5-4a99-9050-8e383ad64885\") " pod="openshift-marketplace/redhat-marketplace-zjvtm" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.521875 4809 scope.go:117] "RemoveContainer" containerID="c672463a6b1697c333512809b2c102eaf5fc994ca7c2b0bd7c2e9ffd6436fac0" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.527450 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zk7tm"] Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.531311 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-slt9h"] Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.538829 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-slt9h"] Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.541115 4809 scope.go:117] "RemoveContainer" containerID="12e34109bddd8e50e1a581763ffb23958cb523a827d40e9a9dd070c15da134dc" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.541723 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cccmm\" (UniqueName: \"kubernetes.io/projected/0714d252-8cd5-4a99-9050-8e383ad64885-kube-api-access-cccmm\") pod \"redhat-marketplace-zjvtm\" (UID: \"0714d252-8cd5-4a99-9050-8e383ad64885\") " pod="openshift-marketplace/redhat-marketplace-zjvtm" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.542767 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lpb28"] Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.545894 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lpb28"] Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.566002 4809 scope.go:117] "RemoveContainer" containerID="29dd3f6f4e238c2aaea07bfb17de884e652c4fc6d66a928725123e15777f54c0" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.581287 4809 scope.go:117] "RemoveContainer" containerID="fb25f0d646c14b2a4d5d1288817d466f004acaaa4d3fc0aa90dae543dd397b6b" Dec 06 05:58:34 crc kubenswrapper[4809]: I1206 05:58:34.691631 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zjvtm" Dec 06 05:58:35 crc kubenswrapper[4809]: I1206 05:58:35.104050 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zjvtm"] Dec 06 05:58:35 crc kubenswrapper[4809]: I1206 05:58:35.395484 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c4b2f56-0b03-41ae-af23-6192b8bd3d34" path="/var/lib/kubelet/pods/2c4b2f56-0b03-41ae-af23-6192b8bd3d34/volumes" Dec 06 05:58:35 crc kubenswrapper[4809]: I1206 05:58:35.396647 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40a90e94-16ee-42ba-beff-5103476432c1" path="/var/lib/kubelet/pods/40a90e94-16ee-42ba-beff-5103476432c1/volumes" Dec 06 05:58:35 crc kubenswrapper[4809]: I1206 05:58:35.398378 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e67c2b8-8bd1-466c-98d8-e93d411162ba" path="/var/lib/kubelet/pods/5e67c2b8-8bd1-466c-98d8-e93d411162ba/volumes" Dec 06 05:58:35 crc kubenswrapper[4809]: I1206 05:58:35.398784 4809 generic.go:334] "Generic (PLEG): container finished" podID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerID="2442b241343358fa541b07551125d7c63b82fcdcd98be18afedabbdc8da517a8" exitCode=0 Dec 06 05:58:35 crc kubenswrapper[4809]: I1206 05:58:35.399816 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="893020b2-fa30-4371-86f9-a585517839ce" path="/var/lib/kubelet/pods/893020b2-fa30-4371-86f9-a585517839ce/volumes" Dec 06 05:58:35 crc kubenswrapper[4809]: I1206 05:58:35.400634 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee96792c-e168-4c58-a477-cd88c14da612" path="/var/lib/kubelet/pods/ee96792c-e168-4c58-a477-cd88c14da612/volumes" Dec 06 05:58:35 crc kubenswrapper[4809]: I1206 05:58:35.401096 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zjvtm" event={"ID":"0714d252-8cd5-4a99-9050-8e383ad64885","Type":"ContainerStarted","Data":"fc2c7cbcd98380c8b9622fcdaad6ea925f9a2bb0b008f4c480c7250890daa390"} Dec 06 05:58:35 crc kubenswrapper[4809]: I1206 05:58:35.401130 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" event={"ID":"cbf4cf62-024e-4703-a8b8-9aecda9cd26a","Type":"ContainerDied","Data":"2442b241343358fa541b07551125d7c63b82fcdcd98be18afedabbdc8da517a8"} Dec 06 05:58:35 crc kubenswrapper[4809]: I1206 05:58:35.401172 4809 scope.go:117] "RemoveContainer" containerID="930788ec9b4d97fdc39353a93111b84707037ce445f1930aa7737a20a44e1d37" Dec 06 05:58:36 crc kubenswrapper[4809]: I1206 05:58:36.558361 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4p5rc"] Dec 06 05:58:36 crc kubenswrapper[4809]: I1206 05:58:36.559638 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4p5rc" Dec 06 05:58:36 crc kubenswrapper[4809]: I1206 05:58:36.562078 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 06 05:58:36 crc kubenswrapper[4809]: I1206 05:58:36.582571 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4p5rc"] Dec 06 05:58:36 crc kubenswrapper[4809]: I1206 05:58:36.648763 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4qlt\" (UniqueName: \"kubernetes.io/projected/6fa58271-edf3-4d3d-8782-83a959a691fe-kube-api-access-m4qlt\") pod \"community-operators-4p5rc\" (UID: \"6fa58271-edf3-4d3d-8782-83a959a691fe\") " pod="openshift-marketplace/community-operators-4p5rc" Dec 06 05:58:36 crc kubenswrapper[4809]: I1206 05:58:36.648831 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fa58271-edf3-4d3d-8782-83a959a691fe-catalog-content\") pod \"community-operators-4p5rc\" (UID: \"6fa58271-edf3-4d3d-8782-83a959a691fe\") " pod="openshift-marketplace/community-operators-4p5rc" Dec 06 05:58:36 crc kubenswrapper[4809]: I1206 05:58:36.648857 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fa58271-edf3-4d3d-8782-83a959a691fe-utilities\") pod \"community-operators-4p5rc\" (UID: \"6fa58271-edf3-4d3d-8782-83a959a691fe\") " pod="openshift-marketplace/community-operators-4p5rc" Dec 06 05:58:36 crc kubenswrapper[4809]: I1206 05:58:36.749817 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fa58271-edf3-4d3d-8782-83a959a691fe-utilities\") pod \"community-operators-4p5rc\" (UID: \"6fa58271-edf3-4d3d-8782-83a959a691fe\") " pod="openshift-marketplace/community-operators-4p5rc" Dec 06 05:58:36 crc kubenswrapper[4809]: I1206 05:58:36.750220 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4qlt\" (UniqueName: \"kubernetes.io/projected/6fa58271-edf3-4d3d-8782-83a959a691fe-kube-api-access-m4qlt\") pod \"community-operators-4p5rc\" (UID: \"6fa58271-edf3-4d3d-8782-83a959a691fe\") " pod="openshift-marketplace/community-operators-4p5rc" Dec 06 05:58:36 crc kubenswrapper[4809]: I1206 05:58:36.750383 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fa58271-edf3-4d3d-8782-83a959a691fe-catalog-content\") pod \"community-operators-4p5rc\" (UID: \"6fa58271-edf3-4d3d-8782-83a959a691fe\") " pod="openshift-marketplace/community-operators-4p5rc" Dec 06 05:58:36 crc kubenswrapper[4809]: I1206 05:58:36.750824 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fa58271-edf3-4d3d-8782-83a959a691fe-catalog-content\") pod \"community-operators-4p5rc\" (UID: \"6fa58271-edf3-4d3d-8782-83a959a691fe\") " pod="openshift-marketplace/community-operators-4p5rc" Dec 06 05:58:36 crc kubenswrapper[4809]: I1206 05:58:36.751330 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fa58271-edf3-4d3d-8782-83a959a691fe-utilities\") pod \"community-operators-4p5rc\" (UID: \"6fa58271-edf3-4d3d-8782-83a959a691fe\") " pod="openshift-marketplace/community-operators-4p5rc" Dec 06 05:58:36 crc kubenswrapper[4809]: I1206 05:58:36.756578 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nqpwx"] Dec 06 05:58:36 crc kubenswrapper[4809]: I1206 05:58:36.757730 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nqpwx" Dec 06 05:58:36 crc kubenswrapper[4809]: I1206 05:58:36.760439 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 06 05:58:36 crc kubenswrapper[4809]: I1206 05:58:36.774317 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4qlt\" (UniqueName: \"kubernetes.io/projected/6fa58271-edf3-4d3d-8782-83a959a691fe-kube-api-access-m4qlt\") pod \"community-operators-4p5rc\" (UID: \"6fa58271-edf3-4d3d-8782-83a959a691fe\") " pod="openshift-marketplace/community-operators-4p5rc" Dec 06 05:58:36 crc kubenswrapper[4809]: I1206 05:58:36.781586 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nqpwx"] Dec 06 05:58:36 crc kubenswrapper[4809]: I1206 05:58:36.852086 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/408b531c-9f2a-4ee5-ad65-db67b2313be3-utilities\") pod \"certified-operators-nqpwx\" (UID: \"408b531c-9f2a-4ee5-ad65-db67b2313be3\") " pod="openshift-marketplace/certified-operators-nqpwx" Dec 06 05:58:36 crc kubenswrapper[4809]: I1206 05:58:36.852201 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/408b531c-9f2a-4ee5-ad65-db67b2313be3-catalog-content\") pod \"certified-operators-nqpwx\" (UID: \"408b531c-9f2a-4ee5-ad65-db67b2313be3\") " pod="openshift-marketplace/certified-operators-nqpwx" Dec 06 05:58:36 crc kubenswrapper[4809]: I1206 05:58:36.852247 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfwjd\" (UniqueName: \"kubernetes.io/projected/408b531c-9f2a-4ee5-ad65-db67b2313be3-kube-api-access-mfwjd\") pod \"certified-operators-nqpwx\" (UID: \"408b531c-9f2a-4ee5-ad65-db67b2313be3\") " pod="openshift-marketplace/certified-operators-nqpwx" Dec 06 05:58:36 crc kubenswrapper[4809]: I1206 05:58:36.884831 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4p5rc" Dec 06 05:58:36 crc kubenswrapper[4809]: I1206 05:58:36.953964 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/408b531c-9f2a-4ee5-ad65-db67b2313be3-catalog-content\") pod \"certified-operators-nqpwx\" (UID: \"408b531c-9f2a-4ee5-ad65-db67b2313be3\") " pod="openshift-marketplace/certified-operators-nqpwx" Dec 06 05:58:36 crc kubenswrapper[4809]: I1206 05:58:36.954017 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfwjd\" (UniqueName: \"kubernetes.io/projected/408b531c-9f2a-4ee5-ad65-db67b2313be3-kube-api-access-mfwjd\") pod \"certified-operators-nqpwx\" (UID: \"408b531c-9f2a-4ee5-ad65-db67b2313be3\") " pod="openshift-marketplace/certified-operators-nqpwx" Dec 06 05:58:36 crc kubenswrapper[4809]: I1206 05:58:36.954052 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/408b531c-9f2a-4ee5-ad65-db67b2313be3-utilities\") pod \"certified-operators-nqpwx\" (UID: \"408b531c-9f2a-4ee5-ad65-db67b2313be3\") " pod="openshift-marketplace/certified-operators-nqpwx" Dec 06 05:58:36 crc kubenswrapper[4809]: I1206 05:58:36.954552 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/408b531c-9f2a-4ee5-ad65-db67b2313be3-utilities\") pod \"certified-operators-nqpwx\" (UID: \"408b531c-9f2a-4ee5-ad65-db67b2313be3\") " pod="openshift-marketplace/certified-operators-nqpwx" Dec 06 05:58:36 crc kubenswrapper[4809]: I1206 05:58:36.954724 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/408b531c-9f2a-4ee5-ad65-db67b2313be3-catalog-content\") pod \"certified-operators-nqpwx\" (UID: \"408b531c-9f2a-4ee5-ad65-db67b2313be3\") " pod="openshift-marketplace/certified-operators-nqpwx" Dec 06 05:58:36 crc kubenswrapper[4809]: I1206 05:58:36.973469 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfwjd\" (UniqueName: \"kubernetes.io/projected/408b531c-9f2a-4ee5-ad65-db67b2313be3-kube-api-access-mfwjd\") pod \"certified-operators-nqpwx\" (UID: \"408b531c-9f2a-4ee5-ad65-db67b2313be3\") " pod="openshift-marketplace/certified-operators-nqpwx" Dec 06 05:58:37 crc kubenswrapper[4809]: I1206 05:58:37.108961 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nqpwx" Dec 06 05:58:37 crc kubenswrapper[4809]: I1206 05:58:37.285840 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4p5rc"] Dec 06 05:58:37 crc kubenswrapper[4809]: W1206 05:58:37.287236 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6fa58271_edf3_4d3d_8782_83a959a691fe.slice/crio-c7521299070a87cd972f9e2a72cc51c65cf3bb93f7a26f49c7051daddcdfca45 WatchSource:0}: Error finding container c7521299070a87cd972f9e2a72cc51c65cf3bb93f7a26f49c7051daddcdfca45: Status 404 returned error can't find the container with id c7521299070a87cd972f9e2a72cc51c65cf3bb93f7a26f49c7051daddcdfca45 Dec 06 05:58:37 crc kubenswrapper[4809]: I1206 05:58:37.413884 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4p5rc" event={"ID":"6fa58271-edf3-4d3d-8782-83a959a691fe","Type":"ContainerStarted","Data":"c7521299070a87cd972f9e2a72cc51c65cf3bb93f7a26f49c7051daddcdfca45"} Dec 06 05:58:37 crc kubenswrapper[4809]: I1206 05:58:37.511213 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nqpwx"] Dec 06 05:58:38 crc kubenswrapper[4809]: I1206 05:58:38.420362 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nqpwx" event={"ID":"408b531c-9f2a-4ee5-ad65-db67b2313be3","Type":"ContainerStarted","Data":"5454d58efc07a5daf9f0a8143bae6525ef2cbe14613dcae5f446155fa3d92185"} Dec 06 05:58:38 crc kubenswrapper[4809]: I1206 05:58:38.952459 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-czfkp"] Dec 06 05:58:38 crc kubenswrapper[4809]: I1206 05:58:38.953749 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-czfkp" Dec 06 05:58:38 crc kubenswrapper[4809]: I1206 05:58:38.955191 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 06 05:58:38 crc kubenswrapper[4809]: I1206 05:58:38.969218 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-czfkp"] Dec 06 05:58:39 crc kubenswrapper[4809]: I1206 05:58:39.080399 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fd2rx\" (UniqueName: \"kubernetes.io/projected/1c5ff94a-d9b0-41f2-9d47-47f67e0a9895-kube-api-access-fd2rx\") pod \"redhat-operators-czfkp\" (UID: \"1c5ff94a-d9b0-41f2-9d47-47f67e0a9895\") " pod="openshift-marketplace/redhat-operators-czfkp" Dec 06 05:58:39 crc kubenswrapper[4809]: I1206 05:58:39.080996 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c5ff94a-d9b0-41f2-9d47-47f67e0a9895-catalog-content\") pod \"redhat-operators-czfkp\" (UID: \"1c5ff94a-d9b0-41f2-9d47-47f67e0a9895\") " pod="openshift-marketplace/redhat-operators-czfkp" Dec 06 05:58:39 crc kubenswrapper[4809]: I1206 05:58:39.081130 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c5ff94a-d9b0-41f2-9d47-47f67e0a9895-utilities\") pod \"redhat-operators-czfkp\" (UID: \"1c5ff94a-d9b0-41f2-9d47-47f67e0a9895\") " pod="openshift-marketplace/redhat-operators-czfkp" Dec 06 05:58:39 crc kubenswrapper[4809]: I1206 05:58:39.182404 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fd2rx\" (UniqueName: \"kubernetes.io/projected/1c5ff94a-d9b0-41f2-9d47-47f67e0a9895-kube-api-access-fd2rx\") pod \"redhat-operators-czfkp\" (UID: \"1c5ff94a-d9b0-41f2-9d47-47f67e0a9895\") " pod="openshift-marketplace/redhat-operators-czfkp" Dec 06 05:58:39 crc kubenswrapper[4809]: I1206 05:58:39.182455 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c5ff94a-d9b0-41f2-9d47-47f67e0a9895-catalog-content\") pod \"redhat-operators-czfkp\" (UID: \"1c5ff94a-d9b0-41f2-9d47-47f67e0a9895\") " pod="openshift-marketplace/redhat-operators-czfkp" Dec 06 05:58:39 crc kubenswrapper[4809]: I1206 05:58:39.182559 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c5ff94a-d9b0-41f2-9d47-47f67e0a9895-utilities\") pod \"redhat-operators-czfkp\" (UID: \"1c5ff94a-d9b0-41f2-9d47-47f67e0a9895\") " pod="openshift-marketplace/redhat-operators-czfkp" Dec 06 05:58:39 crc kubenswrapper[4809]: I1206 05:58:39.182954 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c5ff94a-d9b0-41f2-9d47-47f67e0a9895-utilities\") pod \"redhat-operators-czfkp\" (UID: \"1c5ff94a-d9b0-41f2-9d47-47f67e0a9895\") " pod="openshift-marketplace/redhat-operators-czfkp" Dec 06 05:58:39 crc kubenswrapper[4809]: I1206 05:58:39.182958 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c5ff94a-d9b0-41f2-9d47-47f67e0a9895-catalog-content\") pod \"redhat-operators-czfkp\" (UID: \"1c5ff94a-d9b0-41f2-9d47-47f67e0a9895\") " pod="openshift-marketplace/redhat-operators-czfkp" Dec 06 05:58:39 crc kubenswrapper[4809]: I1206 05:58:39.202467 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fd2rx\" (UniqueName: \"kubernetes.io/projected/1c5ff94a-d9b0-41f2-9d47-47f67e0a9895-kube-api-access-fd2rx\") pod \"redhat-operators-czfkp\" (UID: \"1c5ff94a-d9b0-41f2-9d47-47f67e0a9895\") " pod="openshift-marketplace/redhat-operators-czfkp" Dec 06 05:58:39 crc kubenswrapper[4809]: I1206 05:58:39.275345 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-czfkp" Dec 06 05:58:39 crc kubenswrapper[4809]: I1206 05:58:39.427375 4809 generic.go:334] "Generic (PLEG): container finished" podID="0714d252-8cd5-4a99-9050-8e383ad64885" containerID="d2e7f6aa12dc7a2d7716347f1c80b5b86b0084387ae9989943884c530ff6a3ab" exitCode=0 Dec 06 05:58:39 crc kubenswrapper[4809]: I1206 05:58:39.427463 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zjvtm" event={"ID":"0714d252-8cd5-4a99-9050-8e383ad64885","Type":"ContainerDied","Data":"d2e7f6aa12dc7a2d7716347f1c80b5b86b0084387ae9989943884c530ff6a3ab"} Dec 06 05:58:39 crc kubenswrapper[4809]: I1206 05:58:39.447116 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" event={"ID":"cbf4cf62-024e-4703-a8b8-9aecda9cd26a","Type":"ContainerStarted","Data":"e6b027954a4e3b35818e27fb0ec5b9b81b52eb65e6a41e8dc2459f7e54dc0c30"} Dec 06 05:58:39 crc kubenswrapper[4809]: I1206 05:58:39.451858 4809 generic.go:334] "Generic (PLEG): container finished" podID="6fa58271-edf3-4d3d-8782-83a959a691fe" containerID="d0928cbb94dfa71eb6848cc8cf6441a184a1dbec2a82e18c94293ad8554754e0" exitCode=0 Dec 06 05:58:39 crc kubenswrapper[4809]: I1206 05:58:39.451991 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4p5rc" event={"ID":"6fa58271-edf3-4d3d-8782-83a959a691fe","Type":"ContainerDied","Data":"d0928cbb94dfa71eb6848cc8cf6441a184a1dbec2a82e18c94293ad8554754e0"} Dec 06 05:58:39 crc kubenswrapper[4809]: I1206 05:58:39.453747 4809 generic.go:334] "Generic (PLEG): container finished" podID="408b531c-9f2a-4ee5-ad65-db67b2313be3" containerID="75b48adf1bbc883d4431b2fb24d5f7a914b874e879fd815d8b5be4e58d07f543" exitCode=0 Dec 06 05:58:39 crc kubenswrapper[4809]: I1206 05:58:39.453780 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nqpwx" event={"ID":"408b531c-9f2a-4ee5-ad65-db67b2313be3","Type":"ContainerDied","Data":"75b48adf1bbc883d4431b2fb24d5f7a914b874e879fd815d8b5be4e58d07f543"} Dec 06 05:58:39 crc kubenswrapper[4809]: I1206 05:58:39.635967 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-czfkp"] Dec 06 05:58:39 crc kubenswrapper[4809]: W1206 05:58:39.641119 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1c5ff94a_d9b0_41f2_9d47_47f67e0a9895.slice/crio-ddc2a449ad1dee2157a30485aeba4955a68f42d40d245808b2a5584596f98a58 WatchSource:0}: Error finding container ddc2a449ad1dee2157a30485aeba4955a68f42d40d245808b2a5584596f98a58: Status 404 returned error can't find the container with id ddc2a449ad1dee2157a30485aeba4955a68f42d40d245808b2a5584596f98a58 Dec 06 05:58:40 crc kubenswrapper[4809]: I1206 05:58:40.460032 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-czfkp" event={"ID":"1c5ff94a-d9b0-41f2-9d47-47f67e0a9895","Type":"ContainerStarted","Data":"a0fe1c50875b979d4ccb581be96c00b189381f87c4513b1ac48f5a4ef8fdbb0e"} Dec 06 05:58:40 crc kubenswrapper[4809]: I1206 05:58:40.460429 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-czfkp" event={"ID":"1c5ff94a-d9b0-41f2-9d47-47f67e0a9895","Type":"ContainerStarted","Data":"ddc2a449ad1dee2157a30485aeba4955a68f42d40d245808b2a5584596f98a58"} Dec 06 05:58:41 crc kubenswrapper[4809]: I1206 05:58:41.475406 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zjvtm" event={"ID":"0714d252-8cd5-4a99-9050-8e383ad64885","Type":"ContainerStarted","Data":"14d1a7fc2da16a98fe5e3704385165ca76fe17f2310e932dbafb9a7747281f63"} Dec 06 05:58:41 crc kubenswrapper[4809]: I1206 05:58:41.477632 4809 generic.go:334] "Generic (PLEG): container finished" podID="1c5ff94a-d9b0-41f2-9d47-47f67e0a9895" containerID="a0fe1c50875b979d4ccb581be96c00b189381f87c4513b1ac48f5a4ef8fdbb0e" exitCode=0 Dec 06 05:58:41 crc kubenswrapper[4809]: I1206 05:58:41.477712 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-czfkp" event={"ID":"1c5ff94a-d9b0-41f2-9d47-47f67e0a9895","Type":"ContainerDied","Data":"a0fe1c50875b979d4ccb581be96c00b189381f87c4513b1ac48f5a4ef8fdbb0e"} Dec 06 05:58:41 crc kubenswrapper[4809]: I1206 05:58:41.509534 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4p5rc" event={"ID":"6fa58271-edf3-4d3d-8782-83a959a691fe","Type":"ContainerStarted","Data":"73e1ce115ef05e486697731b5849964166674d52c24b8da30ad9101d42eef954"} Dec 06 05:58:41 crc kubenswrapper[4809]: I1206 05:58:41.516441 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nqpwx" event={"ID":"408b531c-9f2a-4ee5-ad65-db67b2313be3","Type":"ContainerStarted","Data":"6ca7aa64bc2c58ec1b69e7c5d978093965aec9fe915257139b861089f9f7400e"} Dec 06 05:58:42 crc kubenswrapper[4809]: I1206 05:58:42.524546 4809 generic.go:334] "Generic (PLEG): container finished" podID="6fa58271-edf3-4d3d-8782-83a959a691fe" containerID="73e1ce115ef05e486697731b5849964166674d52c24b8da30ad9101d42eef954" exitCode=0 Dec 06 05:58:42 crc kubenswrapper[4809]: I1206 05:58:42.524639 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4p5rc" event={"ID":"6fa58271-edf3-4d3d-8782-83a959a691fe","Type":"ContainerDied","Data":"73e1ce115ef05e486697731b5849964166674d52c24b8da30ad9101d42eef954"} Dec 06 05:58:42 crc kubenswrapper[4809]: I1206 05:58:42.527538 4809 generic.go:334] "Generic (PLEG): container finished" podID="408b531c-9f2a-4ee5-ad65-db67b2313be3" containerID="6ca7aa64bc2c58ec1b69e7c5d978093965aec9fe915257139b861089f9f7400e" exitCode=0 Dec 06 05:58:42 crc kubenswrapper[4809]: I1206 05:58:42.527797 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nqpwx" event={"ID":"408b531c-9f2a-4ee5-ad65-db67b2313be3","Type":"ContainerDied","Data":"6ca7aa64bc2c58ec1b69e7c5d978093965aec9fe915257139b861089f9f7400e"} Dec 06 05:58:42 crc kubenswrapper[4809]: I1206 05:58:42.529599 4809 generic.go:334] "Generic (PLEG): container finished" podID="0714d252-8cd5-4a99-9050-8e383ad64885" containerID="14d1a7fc2da16a98fe5e3704385165ca76fe17f2310e932dbafb9a7747281f63" exitCode=0 Dec 06 05:58:42 crc kubenswrapper[4809]: I1206 05:58:42.529693 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zjvtm" event={"ID":"0714d252-8cd5-4a99-9050-8e383ad64885","Type":"ContainerDied","Data":"14d1a7fc2da16a98fe5e3704385165ca76fe17f2310e932dbafb9a7747281f63"} Dec 06 05:58:42 crc kubenswrapper[4809]: I1206 05:58:42.535731 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-czfkp" event={"ID":"1c5ff94a-d9b0-41f2-9d47-47f67e0a9895","Type":"ContainerStarted","Data":"d455533e19e283b699b92795bb10d7bf43ec89c326cf4dc18c89da94be7fb16b"} Dec 06 05:58:43 crc kubenswrapper[4809]: I1206 05:58:43.551072 4809 generic.go:334] "Generic (PLEG): container finished" podID="1c5ff94a-d9b0-41f2-9d47-47f67e0a9895" containerID="d455533e19e283b699b92795bb10d7bf43ec89c326cf4dc18c89da94be7fb16b" exitCode=0 Dec 06 05:58:43 crc kubenswrapper[4809]: I1206 05:58:43.551168 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-czfkp" event={"ID":"1c5ff94a-d9b0-41f2-9d47-47f67e0a9895","Type":"ContainerDied","Data":"d455533e19e283b699b92795bb10d7bf43ec89c326cf4dc18c89da94be7fb16b"} Dec 06 05:58:43 crc kubenswrapper[4809]: I1206 05:58:43.579505 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zjvtm" event={"ID":"0714d252-8cd5-4a99-9050-8e383ad64885","Type":"ContainerStarted","Data":"c850a60f524244d4ee3b28ed6779948436f8129597b1c419c5d4cef41ead795e"} Dec 06 05:58:43 crc kubenswrapper[4809]: I1206 05:58:43.599319 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4p5rc" podStartSLOduration=3.6824097609999997 podStartE2EDuration="7.599300836s" podCreationTimestamp="2025-12-06 05:58:36 +0000 UTC" firstStartedPulling="2025-12-06 05:58:39.454121793 +0000 UTC m=+444.343104735" lastFinishedPulling="2025-12-06 05:58:43.371012848 +0000 UTC m=+448.259995810" observedRunningTime="2025-12-06 05:58:43.596588792 +0000 UTC m=+448.485571744" watchObservedRunningTime="2025-12-06 05:58:43.599300836 +0000 UTC m=+448.488283788" Dec 06 05:58:43 crc kubenswrapper[4809]: I1206 05:58:43.634688 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zjvtm" podStartSLOduration=5.895203234 podStartE2EDuration="9.634665454s" podCreationTimestamp="2025-12-06 05:58:34 +0000 UTC" firstStartedPulling="2025-12-06 05:58:39.43133694 +0000 UTC m=+444.320319882" lastFinishedPulling="2025-12-06 05:58:43.17079917 +0000 UTC m=+448.059782102" observedRunningTime="2025-12-06 05:58:43.624994639 +0000 UTC m=+448.513977581" watchObservedRunningTime="2025-12-06 05:58:43.634665454 +0000 UTC m=+448.523648396" Dec 06 05:58:44 crc kubenswrapper[4809]: I1206 05:58:44.588783 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nqpwx" event={"ID":"408b531c-9f2a-4ee5-ad65-db67b2313be3","Type":"ContainerStarted","Data":"51544848c3a38a592a69b21684ed46ca5343ac8d86a66bb376364a4fe40f0e3c"} Dec 06 05:58:44 crc kubenswrapper[4809]: I1206 05:58:44.591564 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-czfkp" event={"ID":"1c5ff94a-d9b0-41f2-9d47-47f67e0a9895","Type":"ContainerStarted","Data":"963d399cb0d94a64494c49b796cce0168a9b2195c1e7354a61917b0e014825db"} Dec 06 05:58:44 crc kubenswrapper[4809]: I1206 05:58:44.594451 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4p5rc" event={"ID":"6fa58271-edf3-4d3d-8782-83a959a691fe","Type":"ContainerStarted","Data":"779b773e8f19051505ec32af881e6187c8ad42f783c8123b2f9f9191193699dc"} Dec 06 05:58:44 crc kubenswrapper[4809]: I1206 05:58:44.619378 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nqpwx" podStartSLOduration=4.7354322060000005 podStartE2EDuration="8.619356389s" podCreationTimestamp="2025-12-06 05:58:36 +0000 UTC" firstStartedPulling="2025-12-06 05:58:39.45506844 +0000 UTC m=+444.344051382" lastFinishedPulling="2025-12-06 05:58:43.338992623 +0000 UTC m=+448.227975565" observedRunningTime="2025-12-06 05:58:43.644436391 +0000 UTC m=+448.533419343" watchObservedRunningTime="2025-12-06 05:58:44.619356389 +0000 UTC m=+449.508339331" Dec 06 05:58:44 crc kubenswrapper[4809]: I1206 05:58:44.691813 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zjvtm" Dec 06 05:58:44 crc kubenswrapper[4809]: I1206 05:58:44.691867 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zjvtm" Dec 06 05:58:44 crc kubenswrapper[4809]: I1206 05:58:44.739614 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zjvtm" Dec 06 05:58:44 crc kubenswrapper[4809]: I1206 05:58:44.769558 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-czfkp" podStartSLOduration=4.043800439 podStartE2EDuration="6.769534719s" podCreationTimestamp="2025-12-06 05:58:38 +0000 UTC" firstStartedPulling="2025-12-06 05:58:41.480398662 +0000 UTC m=+446.369381604" lastFinishedPulling="2025-12-06 05:58:44.206132942 +0000 UTC m=+449.095115884" observedRunningTime="2025-12-06 05:58:44.621782596 +0000 UTC m=+449.510765548" watchObservedRunningTime="2025-12-06 05:58:44.769534719 +0000 UTC m=+449.658517661" Dec 06 05:58:46 crc kubenswrapper[4809]: I1206 05:58:46.885963 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4p5rc" Dec 06 05:58:46 crc kubenswrapper[4809]: I1206 05:58:46.886340 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4p5rc" Dec 06 05:58:46 crc kubenswrapper[4809]: I1206 05:58:46.937916 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4p5rc" Dec 06 05:58:47 crc kubenswrapper[4809]: I1206 05:58:47.109374 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nqpwx" Dec 06 05:58:47 crc kubenswrapper[4809]: I1206 05:58:47.109758 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nqpwx" Dec 06 05:58:47 crc kubenswrapper[4809]: I1206 05:58:47.161412 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nqpwx" Dec 06 05:58:49 crc kubenswrapper[4809]: I1206 05:58:49.275633 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-czfkp" Dec 06 05:58:49 crc kubenswrapper[4809]: I1206 05:58:49.276168 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-czfkp" Dec 06 05:58:50 crc kubenswrapper[4809]: I1206 05:58:50.321195 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-czfkp" podUID="1c5ff94a-d9b0-41f2-9d47-47f67e0a9895" containerName="registry-server" probeResult="failure" output=< Dec 06 05:58:50 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Dec 06 05:58:50 crc kubenswrapper[4809]: > Dec 06 05:58:54 crc kubenswrapper[4809]: I1206 05:58:54.742613 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zjvtm" Dec 06 05:58:56 crc kubenswrapper[4809]: I1206 05:58:56.928592 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4p5rc" Dec 06 05:58:57 crc kubenswrapper[4809]: I1206 05:58:57.162890 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nqpwx" Dec 06 05:58:58 crc kubenswrapper[4809]: I1206 05:58:58.956576 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" podUID="c8703a5f-60fb-40f5-be0c-d1025852c05a" containerName="registry" containerID="cri-o://7b7947c95ae0c2b087e5964c2a4f311424f6e49c17d8601b507e5c18e01992c9" gracePeriod=30 Dec 06 05:58:59 crc kubenswrapper[4809]: I1206 05:58:59.323699 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-czfkp" Dec 06 05:58:59 crc kubenswrapper[4809]: I1206 05:58:59.368336 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-czfkp" Dec 06 05:59:02 crc kubenswrapper[4809]: I1206 05:59:02.694248 4809 generic.go:334] "Generic (PLEG): container finished" podID="c8703a5f-60fb-40f5-be0c-d1025852c05a" containerID="7b7947c95ae0c2b087e5964c2a4f311424f6e49c17d8601b507e5c18e01992c9" exitCode=0 Dec 06 05:59:02 crc kubenswrapper[4809]: I1206 05:59:02.694339 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" event={"ID":"c8703a5f-60fb-40f5-be0c-d1025852c05a","Type":"ContainerDied","Data":"7b7947c95ae0c2b087e5964c2a4f311424f6e49c17d8601b507e5c18e01992c9"} Dec 06 05:59:03 crc kubenswrapper[4809]: I1206 05:59:03.214138 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:59:03 crc kubenswrapper[4809]: I1206 05:59:03.342350 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c8703a5f-60fb-40f5-be0c-d1025852c05a-installation-pull-secrets\") pod \"c8703a5f-60fb-40f5-be0c-d1025852c05a\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " Dec 06 05:59:03 crc kubenswrapper[4809]: I1206 05:59:03.342721 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c8703a5f-60fb-40f5-be0c-d1025852c05a-registry-certificates\") pod \"c8703a5f-60fb-40f5-be0c-d1025852c05a\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " Dec 06 05:59:03 crc kubenswrapper[4809]: I1206 05:59:03.342826 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"c8703a5f-60fb-40f5-be0c-d1025852c05a\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " Dec 06 05:59:03 crc kubenswrapper[4809]: I1206 05:59:03.342891 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c8703a5f-60fb-40f5-be0c-d1025852c05a-ca-trust-extracted\") pod \"c8703a5f-60fb-40f5-be0c-d1025852c05a\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " Dec 06 05:59:03 crc kubenswrapper[4809]: I1206 05:59:03.342970 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c8703a5f-60fb-40f5-be0c-d1025852c05a-trusted-ca\") pod \"c8703a5f-60fb-40f5-be0c-d1025852c05a\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " Dec 06 05:59:03 crc kubenswrapper[4809]: I1206 05:59:03.342997 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55jls\" (UniqueName: \"kubernetes.io/projected/c8703a5f-60fb-40f5-be0c-d1025852c05a-kube-api-access-55jls\") pod \"c8703a5f-60fb-40f5-be0c-d1025852c05a\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " Dec 06 05:59:03 crc kubenswrapper[4809]: I1206 05:59:03.343022 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c8703a5f-60fb-40f5-be0c-d1025852c05a-bound-sa-token\") pod \"c8703a5f-60fb-40f5-be0c-d1025852c05a\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " Dec 06 05:59:03 crc kubenswrapper[4809]: I1206 05:59:03.343068 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c8703a5f-60fb-40f5-be0c-d1025852c05a-registry-tls\") pod \"c8703a5f-60fb-40f5-be0c-d1025852c05a\" (UID: \"c8703a5f-60fb-40f5-be0c-d1025852c05a\") " Dec 06 05:59:03 crc kubenswrapper[4809]: I1206 05:59:03.343204 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8703a5f-60fb-40f5-be0c-d1025852c05a-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "c8703a5f-60fb-40f5-be0c-d1025852c05a" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:59:03 crc kubenswrapper[4809]: I1206 05:59:03.343745 4809 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c8703a5f-60fb-40f5-be0c-d1025852c05a-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 06 05:59:03 crc kubenswrapper[4809]: I1206 05:59:03.343862 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8703a5f-60fb-40f5-be0c-d1025852c05a-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "c8703a5f-60fb-40f5-be0c-d1025852c05a" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:59:03 crc kubenswrapper[4809]: I1206 05:59:03.351249 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8703a5f-60fb-40f5-be0c-d1025852c05a-kube-api-access-55jls" (OuterVolumeSpecName: "kube-api-access-55jls") pod "c8703a5f-60fb-40f5-be0c-d1025852c05a" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a"). InnerVolumeSpecName "kube-api-access-55jls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:59:03 crc kubenswrapper[4809]: I1206 05:59:03.352239 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8703a5f-60fb-40f5-be0c-d1025852c05a-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "c8703a5f-60fb-40f5-be0c-d1025852c05a" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:59:03 crc kubenswrapper[4809]: I1206 05:59:03.352887 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8703a5f-60fb-40f5-be0c-d1025852c05a-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "c8703a5f-60fb-40f5-be0c-d1025852c05a" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:59:03 crc kubenswrapper[4809]: I1206 05:59:03.353224 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8703a5f-60fb-40f5-be0c-d1025852c05a-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "c8703a5f-60fb-40f5-be0c-d1025852c05a" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:59:03 crc kubenswrapper[4809]: I1206 05:59:03.365551 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "c8703a5f-60fb-40f5-be0c-d1025852c05a" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 06 05:59:03 crc kubenswrapper[4809]: I1206 05:59:03.368355 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8703a5f-60fb-40f5-be0c-d1025852c05a-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "c8703a5f-60fb-40f5-be0c-d1025852c05a" (UID: "c8703a5f-60fb-40f5-be0c-d1025852c05a"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:59:03 crc kubenswrapper[4809]: I1206 05:59:03.445416 4809 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c8703a5f-60fb-40f5-be0c-d1025852c05a-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 06 05:59:03 crc kubenswrapper[4809]: I1206 05:59:03.445457 4809 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c8703a5f-60fb-40f5-be0c-d1025852c05a-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 06 05:59:03 crc kubenswrapper[4809]: I1206 05:59:03.445470 4809 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c8703a5f-60fb-40f5-be0c-d1025852c05a-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 05:59:03 crc kubenswrapper[4809]: I1206 05:59:03.445484 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55jls\" (UniqueName: \"kubernetes.io/projected/c8703a5f-60fb-40f5-be0c-d1025852c05a-kube-api-access-55jls\") on node \"crc\" DevicePath \"\"" Dec 06 05:59:03 crc kubenswrapper[4809]: I1206 05:59:03.445496 4809 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c8703a5f-60fb-40f5-be0c-d1025852c05a-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 06 05:59:03 crc kubenswrapper[4809]: I1206 05:59:03.445506 4809 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c8703a5f-60fb-40f5-be0c-d1025852c05a-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 06 05:59:03 crc kubenswrapper[4809]: I1206 05:59:03.701047 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" event={"ID":"c8703a5f-60fb-40f5-be0c-d1025852c05a","Type":"ContainerDied","Data":"e50dd982694f146cca4fadd24a64db7898a56020f167f6645c5fcaa8d16c37b4"} Dec 06 05:59:03 crc kubenswrapper[4809]: I1206 05:59:03.701098 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-kwzrc" Dec 06 05:59:03 crc kubenswrapper[4809]: I1206 05:59:03.701108 4809 scope.go:117] "RemoveContainer" containerID="7b7947c95ae0c2b087e5964c2a4f311424f6e49c17d8601b507e5c18e01992c9" Dec 06 05:59:03 crc kubenswrapper[4809]: I1206 05:59:03.724536 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kwzrc"] Dec 06 05:59:03 crc kubenswrapper[4809]: I1206 05:59:03.727458 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kwzrc"] Dec 06 05:59:05 crc kubenswrapper[4809]: I1206 05:59:05.395497 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8703a5f-60fb-40f5-be0c-d1025852c05a" path="/var/lib/kubelet/pods/c8703a5f-60fb-40f5-be0c-d1025852c05a/volumes" Dec 06 05:59:05 crc kubenswrapper[4809]: I1206 05:59:05.774641 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-6d5b84845-ln5ns"] Dec 06 05:59:05 crc kubenswrapper[4809]: E1206 05:59:05.775153 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8703a5f-60fb-40f5-be0c-d1025852c05a" containerName="registry" Dec 06 05:59:05 crc kubenswrapper[4809]: I1206 05:59:05.775168 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8703a5f-60fb-40f5-be0c-d1025852c05a" containerName="registry" Dec 06 05:59:05 crc kubenswrapper[4809]: I1206 05:59:05.775307 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8703a5f-60fb-40f5-be0c-d1025852c05a" containerName="registry" Dec 06 05:59:05 crc kubenswrapper[4809]: I1206 05:59:05.775945 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-ln5ns" Dec 06 05:59:05 crc kubenswrapper[4809]: I1206 05:59:05.778531 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"kube-root-ca.crt" Dec 06 05:59:05 crc kubenswrapper[4809]: I1206 05:59:05.778694 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"openshift-service-ca.crt" Dec 06 05:59:05 crc kubenswrapper[4809]: I1206 05:59:05.778848 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"telemetry-config" Dec 06 05:59:05 crc kubenswrapper[4809]: I1206 05:59:05.781276 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"cluster-monitoring-operator-dockercfg-wwt9l" Dec 06 05:59:05 crc kubenswrapper[4809]: I1206 05:59:05.783387 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"cluster-monitoring-operator-tls" Dec 06 05:59:05 crc kubenswrapper[4809]: I1206 05:59:05.788260 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-6d5b84845-ln5ns"] Dec 06 05:59:05 crc kubenswrapper[4809]: I1206 05:59:05.872982 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jtmz\" (UniqueName: \"kubernetes.io/projected/887ebcf7-f7b7-49f2-ad22-c69143d49afb-kube-api-access-9jtmz\") pod \"cluster-monitoring-operator-6d5b84845-ln5ns\" (UID: \"887ebcf7-f7b7-49f2-ad22-c69143d49afb\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-ln5ns" Dec 06 05:59:05 crc kubenswrapper[4809]: I1206 05:59:05.873062 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/887ebcf7-f7b7-49f2-ad22-c69143d49afb-telemetry-config\") pod \"cluster-monitoring-operator-6d5b84845-ln5ns\" (UID: \"887ebcf7-f7b7-49f2-ad22-c69143d49afb\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-ln5ns" Dec 06 05:59:05 crc kubenswrapper[4809]: I1206 05:59:05.873092 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/887ebcf7-f7b7-49f2-ad22-c69143d49afb-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-6d5b84845-ln5ns\" (UID: \"887ebcf7-f7b7-49f2-ad22-c69143d49afb\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-ln5ns" Dec 06 05:59:05 crc kubenswrapper[4809]: I1206 05:59:05.973805 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/887ebcf7-f7b7-49f2-ad22-c69143d49afb-telemetry-config\") pod \"cluster-monitoring-operator-6d5b84845-ln5ns\" (UID: \"887ebcf7-f7b7-49f2-ad22-c69143d49afb\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-ln5ns" Dec 06 05:59:05 crc kubenswrapper[4809]: I1206 05:59:05.973847 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/887ebcf7-f7b7-49f2-ad22-c69143d49afb-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-6d5b84845-ln5ns\" (UID: \"887ebcf7-f7b7-49f2-ad22-c69143d49afb\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-ln5ns" Dec 06 05:59:05 crc kubenswrapper[4809]: I1206 05:59:05.973911 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jtmz\" (UniqueName: \"kubernetes.io/projected/887ebcf7-f7b7-49f2-ad22-c69143d49afb-kube-api-access-9jtmz\") pod \"cluster-monitoring-operator-6d5b84845-ln5ns\" (UID: \"887ebcf7-f7b7-49f2-ad22-c69143d49afb\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-ln5ns" Dec 06 05:59:05 crc kubenswrapper[4809]: I1206 05:59:05.975867 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/887ebcf7-f7b7-49f2-ad22-c69143d49afb-telemetry-config\") pod \"cluster-monitoring-operator-6d5b84845-ln5ns\" (UID: \"887ebcf7-f7b7-49f2-ad22-c69143d49afb\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-ln5ns" Dec 06 05:59:05 crc kubenswrapper[4809]: I1206 05:59:05.979492 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/887ebcf7-f7b7-49f2-ad22-c69143d49afb-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-6d5b84845-ln5ns\" (UID: \"887ebcf7-f7b7-49f2-ad22-c69143d49afb\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-ln5ns" Dec 06 05:59:05 crc kubenswrapper[4809]: I1206 05:59:05.996089 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jtmz\" (UniqueName: \"kubernetes.io/projected/887ebcf7-f7b7-49f2-ad22-c69143d49afb-kube-api-access-9jtmz\") pod \"cluster-monitoring-operator-6d5b84845-ln5ns\" (UID: \"887ebcf7-f7b7-49f2-ad22-c69143d49afb\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-ln5ns" Dec 06 05:59:06 crc kubenswrapper[4809]: I1206 05:59:06.094403 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-ln5ns" Dec 06 05:59:06 crc kubenswrapper[4809]: I1206 05:59:06.476306 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-6d5b84845-ln5ns"] Dec 06 05:59:06 crc kubenswrapper[4809]: I1206 05:59:06.487112 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 05:59:06 crc kubenswrapper[4809]: I1206 05:59:06.718388 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-ln5ns" event={"ID":"887ebcf7-f7b7-49f2-ad22-c69143d49afb","Type":"ContainerStarted","Data":"536563de68edac3706e66d49111fe6039cac9eb99f62e9ca6cfb02d981062f37"} Dec 06 05:59:09 crc kubenswrapper[4809]: I1206 05:59:09.257128 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-ljfgz"] Dec 06 05:59:09 crc kubenswrapper[4809]: I1206 05:59:09.258455 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-ljfgz" Dec 06 05:59:09 crc kubenswrapper[4809]: I1206 05:59:09.260183 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-admission-webhook-dockercfg-wkdhn" Dec 06 05:59:09 crc kubenswrapper[4809]: I1206 05:59:09.260886 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-admission-webhook-tls" Dec 06 05:59:09 crc kubenswrapper[4809]: I1206 05:59:09.261838 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-ljfgz"] Dec 06 05:59:09 crc kubenswrapper[4809]: I1206 05:59:09.335632 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/5e0d0108-b36f-47bc-bb5c-9fbc5ed0bf92-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-ljfgz\" (UID: \"5e0d0108-b36f-47bc-bb5c-9fbc5ed0bf92\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-ljfgz" Dec 06 05:59:09 crc kubenswrapper[4809]: I1206 05:59:09.436786 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/5e0d0108-b36f-47bc-bb5c-9fbc5ed0bf92-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-ljfgz\" (UID: \"5e0d0108-b36f-47bc-bb5c-9fbc5ed0bf92\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-ljfgz" Dec 06 05:59:09 crc kubenswrapper[4809]: E1206 05:59:09.436984 4809 secret.go:188] Couldn't get secret openshift-monitoring/prometheus-operator-admission-webhook-tls: secret "prometheus-operator-admission-webhook-tls" not found Dec 06 05:59:09 crc kubenswrapper[4809]: E1206 05:59:09.437299 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5e0d0108-b36f-47bc-bb5c-9fbc5ed0bf92-tls-certificates podName:5e0d0108-b36f-47bc-bb5c-9fbc5ed0bf92 nodeName:}" failed. No retries permitted until 2025-12-06 05:59:09.937277736 +0000 UTC m=+474.826260678 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-certificates" (UniqueName: "kubernetes.io/secret/5e0d0108-b36f-47bc-bb5c-9fbc5ed0bf92-tls-certificates") pod "prometheus-operator-admission-webhook-f54c54754-ljfgz" (UID: "5e0d0108-b36f-47bc-bb5c-9fbc5ed0bf92") : secret "prometheus-operator-admission-webhook-tls" not found Dec 06 05:59:09 crc kubenswrapper[4809]: I1206 05:59:09.739464 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-ln5ns" event={"ID":"887ebcf7-f7b7-49f2-ad22-c69143d49afb","Type":"ContainerStarted","Data":"5014c493618e235f5d0fdfa06f230db5ed379e332235ccd2d391afaeafe243a9"} Dec 06 05:59:09 crc kubenswrapper[4809]: I1206 05:59:09.759614 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-ln5ns" podStartSLOduration=2.547621103 podStartE2EDuration="4.759593512s" podCreationTimestamp="2025-12-06 05:59:05 +0000 UTC" firstStartedPulling="2025-12-06 05:59:06.48682278 +0000 UTC m=+471.375805722" lastFinishedPulling="2025-12-06 05:59:08.698795189 +0000 UTC m=+473.587778131" observedRunningTime="2025-12-06 05:59:09.756887758 +0000 UTC m=+474.645870760" watchObservedRunningTime="2025-12-06 05:59:09.759593512 +0000 UTC m=+474.648576454" Dec 06 05:59:09 crc kubenswrapper[4809]: I1206 05:59:09.943328 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/5e0d0108-b36f-47bc-bb5c-9fbc5ed0bf92-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-ljfgz\" (UID: \"5e0d0108-b36f-47bc-bb5c-9fbc5ed0bf92\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-ljfgz" Dec 06 05:59:09 crc kubenswrapper[4809]: I1206 05:59:09.949482 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/5e0d0108-b36f-47bc-bb5c-9fbc5ed0bf92-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-ljfgz\" (UID: \"5e0d0108-b36f-47bc-bb5c-9fbc5ed0bf92\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-ljfgz" Dec 06 05:59:10 crc kubenswrapper[4809]: I1206 05:59:10.173532 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-ljfgz" Dec 06 05:59:10 crc kubenswrapper[4809]: I1206 05:59:10.585594 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-ljfgz"] Dec 06 05:59:10 crc kubenswrapper[4809]: I1206 05:59:10.744734 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-ljfgz" event={"ID":"5e0d0108-b36f-47bc-bb5c-9fbc5ed0bf92","Type":"ContainerStarted","Data":"b35df68d0b3e2f83a32348cf9da5cf2f064bd4e52af4776f3b825f1220ae9b6e"} Dec 06 05:59:12 crc kubenswrapper[4809]: I1206 05:59:12.757587 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-ljfgz" event={"ID":"5e0d0108-b36f-47bc-bb5c-9fbc5ed0bf92","Type":"ContainerStarted","Data":"df5b73b750eb5706c1ac0a1dffd6ded5f8dec3f1775a781b809e3071747bc55d"} Dec 06 05:59:12 crc kubenswrapper[4809]: I1206 05:59:12.758062 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-ljfgz" Dec 06 05:59:12 crc kubenswrapper[4809]: I1206 05:59:12.763747 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-ljfgz" Dec 06 05:59:12 crc kubenswrapper[4809]: I1206 05:59:12.834052 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-ljfgz" podStartSLOduration=2.627747128 podStartE2EDuration="3.834033237s" podCreationTimestamp="2025-12-06 05:59:09 +0000 UTC" firstStartedPulling="2025-12-06 05:59:10.590601501 +0000 UTC m=+475.479584443" lastFinishedPulling="2025-12-06 05:59:11.79688761 +0000 UTC m=+476.685870552" observedRunningTime="2025-12-06 05:59:12.830758708 +0000 UTC m=+477.719741660" watchObservedRunningTime="2025-12-06 05:59:12.834033237 +0000 UTC m=+477.723016179" Dec 06 05:59:13 crc kubenswrapper[4809]: I1206 05:59:13.324804 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-operator-db54df47d-lbjmx"] Dec 06 05:59:13 crc kubenswrapper[4809]: I1206 05:59:13.326200 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-db54df47d-lbjmx" Dec 06 05:59:13 crc kubenswrapper[4809]: I1206 05:59:13.330161 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-tls" Dec 06 05:59:13 crc kubenswrapper[4809]: I1206 05:59:13.330754 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-kube-rbac-proxy-config" Dec 06 05:59:13 crc kubenswrapper[4809]: I1206 05:59:13.331028 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-dockercfg-jsb2l" Dec 06 05:59:13 crc kubenswrapper[4809]: I1206 05:59:13.331264 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"metrics-client-ca" Dec 06 05:59:13 crc kubenswrapper[4809]: I1206 05:59:13.339910 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-db54df47d-lbjmx"] Dec 06 05:59:13 crc kubenswrapper[4809]: I1206 05:59:13.385599 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/a2c650ac-1257-460a-98c0-9a7db19710e1-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-db54df47d-lbjmx\" (UID: \"a2c650ac-1257-460a-98c0-9a7db19710e1\") " pod="openshift-monitoring/prometheus-operator-db54df47d-lbjmx" Dec 06 05:59:13 crc kubenswrapper[4809]: I1206 05:59:13.385663 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/a2c650ac-1257-460a-98c0-9a7db19710e1-prometheus-operator-tls\") pod \"prometheus-operator-db54df47d-lbjmx\" (UID: \"a2c650ac-1257-460a-98c0-9a7db19710e1\") " pod="openshift-monitoring/prometheus-operator-db54df47d-lbjmx" Dec 06 05:59:13 crc kubenswrapper[4809]: I1206 05:59:13.385689 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/a2c650ac-1257-460a-98c0-9a7db19710e1-metrics-client-ca\") pod \"prometheus-operator-db54df47d-lbjmx\" (UID: \"a2c650ac-1257-460a-98c0-9a7db19710e1\") " pod="openshift-monitoring/prometheus-operator-db54df47d-lbjmx" Dec 06 05:59:13 crc kubenswrapper[4809]: I1206 05:59:13.385899 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96hq7\" (UniqueName: \"kubernetes.io/projected/a2c650ac-1257-460a-98c0-9a7db19710e1-kube-api-access-96hq7\") pod \"prometheus-operator-db54df47d-lbjmx\" (UID: \"a2c650ac-1257-460a-98c0-9a7db19710e1\") " pod="openshift-monitoring/prometheus-operator-db54df47d-lbjmx" Dec 06 05:59:13 crc kubenswrapper[4809]: I1206 05:59:13.486702 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/a2c650ac-1257-460a-98c0-9a7db19710e1-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-db54df47d-lbjmx\" (UID: \"a2c650ac-1257-460a-98c0-9a7db19710e1\") " pod="openshift-monitoring/prometheus-operator-db54df47d-lbjmx" Dec 06 05:59:13 crc kubenswrapper[4809]: I1206 05:59:13.486788 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/a2c650ac-1257-460a-98c0-9a7db19710e1-prometheus-operator-tls\") pod \"prometheus-operator-db54df47d-lbjmx\" (UID: \"a2c650ac-1257-460a-98c0-9a7db19710e1\") " pod="openshift-monitoring/prometheus-operator-db54df47d-lbjmx" Dec 06 05:59:13 crc kubenswrapper[4809]: I1206 05:59:13.486827 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/a2c650ac-1257-460a-98c0-9a7db19710e1-metrics-client-ca\") pod \"prometheus-operator-db54df47d-lbjmx\" (UID: \"a2c650ac-1257-460a-98c0-9a7db19710e1\") " pod="openshift-monitoring/prometheus-operator-db54df47d-lbjmx" Dec 06 05:59:13 crc kubenswrapper[4809]: I1206 05:59:13.486873 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96hq7\" (UniqueName: \"kubernetes.io/projected/a2c650ac-1257-460a-98c0-9a7db19710e1-kube-api-access-96hq7\") pod \"prometheus-operator-db54df47d-lbjmx\" (UID: \"a2c650ac-1257-460a-98c0-9a7db19710e1\") " pod="openshift-monitoring/prometheus-operator-db54df47d-lbjmx" Dec 06 05:59:13 crc kubenswrapper[4809]: I1206 05:59:13.487970 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/a2c650ac-1257-460a-98c0-9a7db19710e1-metrics-client-ca\") pod \"prometheus-operator-db54df47d-lbjmx\" (UID: \"a2c650ac-1257-460a-98c0-9a7db19710e1\") " pod="openshift-monitoring/prometheus-operator-db54df47d-lbjmx" Dec 06 05:59:13 crc kubenswrapper[4809]: I1206 05:59:13.492544 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/a2c650ac-1257-460a-98c0-9a7db19710e1-prometheus-operator-tls\") pod \"prometheus-operator-db54df47d-lbjmx\" (UID: \"a2c650ac-1257-460a-98c0-9a7db19710e1\") " pod="openshift-monitoring/prometheus-operator-db54df47d-lbjmx" Dec 06 05:59:13 crc kubenswrapper[4809]: I1206 05:59:13.492900 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/a2c650ac-1257-460a-98c0-9a7db19710e1-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-db54df47d-lbjmx\" (UID: \"a2c650ac-1257-460a-98c0-9a7db19710e1\") " pod="openshift-monitoring/prometheus-operator-db54df47d-lbjmx" Dec 06 05:59:13 crc kubenswrapper[4809]: I1206 05:59:13.505244 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96hq7\" (UniqueName: \"kubernetes.io/projected/a2c650ac-1257-460a-98c0-9a7db19710e1-kube-api-access-96hq7\") pod \"prometheus-operator-db54df47d-lbjmx\" (UID: \"a2c650ac-1257-460a-98c0-9a7db19710e1\") " pod="openshift-monitoring/prometheus-operator-db54df47d-lbjmx" Dec 06 05:59:13 crc kubenswrapper[4809]: I1206 05:59:13.678893 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-db54df47d-lbjmx" Dec 06 05:59:14 crc kubenswrapper[4809]: I1206 05:59:14.068656 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-db54df47d-lbjmx"] Dec 06 05:59:14 crc kubenswrapper[4809]: W1206 05:59:14.073330 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda2c650ac_1257_460a_98c0_9a7db19710e1.slice/crio-9eb64394d0386db984441b03890300025dfbd3f38ae13624984ce8b88aad21ca WatchSource:0}: Error finding container 9eb64394d0386db984441b03890300025dfbd3f38ae13624984ce8b88aad21ca: Status 404 returned error can't find the container with id 9eb64394d0386db984441b03890300025dfbd3f38ae13624984ce8b88aad21ca Dec 06 05:59:14 crc kubenswrapper[4809]: I1206 05:59:14.779260 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-db54df47d-lbjmx" event={"ID":"a2c650ac-1257-460a-98c0-9a7db19710e1","Type":"ContainerStarted","Data":"9eb64394d0386db984441b03890300025dfbd3f38ae13624984ce8b88aad21ca"} Dec 06 05:59:16 crc kubenswrapper[4809]: I1206 05:59:16.792709 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-db54df47d-lbjmx" event={"ID":"a2c650ac-1257-460a-98c0-9a7db19710e1","Type":"ContainerStarted","Data":"b4eca520b4b86f0f420732b64d124c82eacb2c02fb5f0d0d3ab01d2c388cd7d9"} Dec 06 05:59:16 crc kubenswrapper[4809]: I1206 05:59:16.793596 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-db54df47d-lbjmx" event={"ID":"a2c650ac-1257-460a-98c0-9a7db19710e1","Type":"ContainerStarted","Data":"3b2d9aaf5ca60e68a23c35e473b8c5769a6cfe030433cc6d55ae326a4f9e213f"} Dec 06 05:59:16 crc kubenswrapper[4809]: I1206 05:59:16.810417 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-operator-db54df47d-lbjmx" podStartSLOduration=1.96512055 podStartE2EDuration="3.810391934s" podCreationTimestamp="2025-12-06 05:59:13 +0000 UTC" firstStartedPulling="2025-12-06 05:59:14.075561205 +0000 UTC m=+478.964544147" lastFinishedPulling="2025-12-06 05:59:15.920832589 +0000 UTC m=+480.809815531" observedRunningTime="2025-12-06 05:59:16.806786497 +0000 UTC m=+481.695769459" watchObservedRunningTime="2025-12-06 05:59:16.810391934 +0000 UTC m=+481.699374876" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.680856 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/openshift-state-metrics-566fddb674-xmx2t"] Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.682341 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/openshift-state-metrics-566fddb674-xmx2t" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.687713 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"openshift-state-metrics-dockercfg-drvd4" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.687730 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"openshift-state-metrics-kube-rbac-proxy-config" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.687780 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"openshift-state-metrics-tls" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.718839 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/kube-state-metrics-777cb5bd5d-78wr9"] Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.719912 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-78wr9" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.721517 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-state-metrics-kube-rbac-proxy-config" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.721840 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-state-metrics-dockercfg-jb77m" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.721975 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"kube-state-metrics-custom-resource-state-configmap" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.724091 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-state-metrics-tls" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.738497 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/node-exporter-cw8sk"] Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.739468 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-cw8sk" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.742200 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"node-exporter-kube-rbac-proxy-config" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.742397 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"node-exporter-tls" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.746152 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/kube-state-metrics-777cb5bd5d-78wr9"] Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.746693 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"node-exporter-dockercfg-5ngt4" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.753441 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/openshift-state-metrics-566fddb674-xmx2t"] Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.754829 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/528ba802-293a-488e-8856-14b997f7956f-node-exporter-tls\") pod \"node-exporter-cw8sk\" (UID: \"528ba802-293a-488e-8856-14b997f7956f\") " pod="openshift-monitoring/node-exporter-cw8sk" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.754880 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/528ba802-293a-488e-8856-14b997f7956f-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-cw8sk\" (UID: \"528ba802-293a-488e-8856-14b997f7956f\") " pod="openshift-monitoring/node-exporter-cw8sk" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.754903 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/5d806c2b-b90f-4541-a9a9-2cf8e86ac947-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-777cb5bd5d-78wr9\" (UID: \"5d806c2b-b90f-4541-a9a9-2cf8e86ac947\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-78wr9" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.754938 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/528ba802-293a-488e-8856-14b997f7956f-metrics-client-ca\") pod \"node-exporter-cw8sk\" (UID: \"528ba802-293a-488e-8856-14b997f7956f\") " pod="openshift-monitoring/node-exporter-cw8sk" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.754957 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/528ba802-293a-488e-8856-14b997f7956f-sys\") pod \"node-exporter-cw8sk\" (UID: \"528ba802-293a-488e-8856-14b997f7956f\") " pod="openshift-monitoring/node-exporter-cw8sk" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.754976 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/e1147859-11de-422f-9fb6-84ad24e7f49c-metrics-client-ca\") pod \"openshift-state-metrics-566fddb674-xmx2t\" (UID: \"e1147859-11de-422f-9fb6-84ad24e7f49c\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-xmx2t" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.754990 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/528ba802-293a-488e-8856-14b997f7956f-root\") pod \"node-exporter-cw8sk\" (UID: \"528ba802-293a-488e-8856-14b997f7956f\") " pod="openshift-monitoring/node-exporter-cw8sk" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.755007 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/528ba802-293a-488e-8856-14b997f7956f-node-exporter-wtmp\") pod \"node-exporter-cw8sk\" (UID: \"528ba802-293a-488e-8856-14b997f7956f\") " pod="openshift-monitoring/node-exporter-cw8sk" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.755021 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zpxt\" (UniqueName: \"kubernetes.io/projected/e1147859-11de-422f-9fb6-84ad24e7f49c-kube-api-access-4zpxt\") pod \"openshift-state-metrics-566fddb674-xmx2t\" (UID: \"e1147859-11de-422f-9fb6-84ad24e7f49c\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-xmx2t" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.755041 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xt5q4\" (UniqueName: \"kubernetes.io/projected/528ba802-293a-488e-8856-14b997f7956f-kube-api-access-xt5q4\") pod \"node-exporter-cw8sk\" (UID: \"528ba802-293a-488e-8856-14b997f7956f\") " pod="openshift-monitoring/node-exporter-cw8sk" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.755060 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4wnx\" (UniqueName: \"kubernetes.io/projected/5d806c2b-b90f-4541-a9a9-2cf8e86ac947-kube-api-access-j4wnx\") pod \"kube-state-metrics-777cb5bd5d-78wr9\" (UID: \"5d806c2b-b90f-4541-a9a9-2cf8e86ac947\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-78wr9" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.755075 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/5d806c2b-b90f-4541-a9a9-2cf8e86ac947-metrics-client-ca\") pod \"kube-state-metrics-777cb5bd5d-78wr9\" (UID: \"5d806c2b-b90f-4541-a9a9-2cf8e86ac947\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-78wr9" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.755096 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/5d806c2b-b90f-4541-a9a9-2cf8e86ac947-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-777cb5bd5d-78wr9\" (UID: \"5d806c2b-b90f-4541-a9a9-2cf8e86ac947\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-78wr9" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.755114 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/e1147859-11de-422f-9fb6-84ad24e7f49c-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-566fddb674-xmx2t\" (UID: \"e1147859-11de-422f-9fb6-84ad24e7f49c\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-xmx2t" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.755138 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/e1147859-11de-422f-9fb6-84ad24e7f49c-openshift-state-metrics-tls\") pod \"openshift-state-metrics-566fddb674-xmx2t\" (UID: \"e1147859-11de-422f-9fb6-84ad24e7f49c\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-xmx2t" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.755156 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/528ba802-293a-488e-8856-14b997f7956f-node-exporter-textfile\") pod \"node-exporter-cw8sk\" (UID: \"528ba802-293a-488e-8856-14b997f7956f\") " pod="openshift-monitoring/node-exporter-cw8sk" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.755175 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/5d806c2b-b90f-4541-a9a9-2cf8e86ac947-kube-state-metrics-tls\") pod \"kube-state-metrics-777cb5bd5d-78wr9\" (UID: \"5d806c2b-b90f-4541-a9a9-2cf8e86ac947\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-78wr9" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.755189 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/5d806c2b-b90f-4541-a9a9-2cf8e86ac947-volume-directive-shadow\") pod \"kube-state-metrics-777cb5bd5d-78wr9\" (UID: \"5d806c2b-b90f-4541-a9a9-2cf8e86ac947\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-78wr9" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.855658 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/528ba802-293a-488e-8856-14b997f7956f-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-cw8sk\" (UID: \"528ba802-293a-488e-8856-14b997f7956f\") " pod="openshift-monitoring/node-exporter-cw8sk" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.855714 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/5d806c2b-b90f-4541-a9a9-2cf8e86ac947-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-777cb5bd5d-78wr9\" (UID: \"5d806c2b-b90f-4541-a9a9-2cf8e86ac947\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-78wr9" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.855743 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/528ba802-293a-488e-8856-14b997f7956f-metrics-client-ca\") pod \"node-exporter-cw8sk\" (UID: \"528ba802-293a-488e-8856-14b997f7956f\") " pod="openshift-monitoring/node-exporter-cw8sk" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.855760 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/528ba802-293a-488e-8856-14b997f7956f-sys\") pod \"node-exporter-cw8sk\" (UID: \"528ba802-293a-488e-8856-14b997f7956f\") " pod="openshift-monitoring/node-exporter-cw8sk" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.856041 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/e1147859-11de-422f-9fb6-84ad24e7f49c-metrics-client-ca\") pod \"openshift-state-metrics-566fddb674-xmx2t\" (UID: \"e1147859-11de-422f-9fb6-84ad24e7f49c\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-xmx2t" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.856155 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/528ba802-293a-488e-8856-14b997f7956f-root\") pod \"node-exporter-cw8sk\" (UID: \"528ba802-293a-488e-8856-14b997f7956f\") " pod="openshift-monitoring/node-exporter-cw8sk" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.856251 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/528ba802-293a-488e-8856-14b997f7956f-node-exporter-wtmp\") pod \"node-exporter-cw8sk\" (UID: \"528ba802-293a-488e-8856-14b997f7956f\") " pod="openshift-monitoring/node-exporter-cw8sk" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.856340 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zpxt\" (UniqueName: \"kubernetes.io/projected/e1147859-11de-422f-9fb6-84ad24e7f49c-kube-api-access-4zpxt\") pod \"openshift-state-metrics-566fddb674-xmx2t\" (UID: \"e1147859-11de-422f-9fb6-84ad24e7f49c\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-xmx2t" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.856455 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xt5q4\" (UniqueName: \"kubernetes.io/projected/528ba802-293a-488e-8856-14b997f7956f-kube-api-access-xt5q4\") pod \"node-exporter-cw8sk\" (UID: \"528ba802-293a-488e-8856-14b997f7956f\") " pod="openshift-monitoring/node-exporter-cw8sk" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.856548 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4wnx\" (UniqueName: \"kubernetes.io/projected/5d806c2b-b90f-4541-a9a9-2cf8e86ac947-kube-api-access-j4wnx\") pod \"kube-state-metrics-777cb5bd5d-78wr9\" (UID: \"5d806c2b-b90f-4541-a9a9-2cf8e86ac947\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-78wr9" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.856632 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/5d806c2b-b90f-4541-a9a9-2cf8e86ac947-metrics-client-ca\") pod \"kube-state-metrics-777cb5bd5d-78wr9\" (UID: \"5d806c2b-b90f-4541-a9a9-2cf8e86ac947\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-78wr9" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.856720 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/5d806c2b-b90f-4541-a9a9-2cf8e86ac947-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-777cb5bd5d-78wr9\" (UID: \"5d806c2b-b90f-4541-a9a9-2cf8e86ac947\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-78wr9" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.856820 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/e1147859-11de-422f-9fb6-84ad24e7f49c-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-566fddb674-xmx2t\" (UID: \"e1147859-11de-422f-9fb6-84ad24e7f49c\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-xmx2t" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.856273 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"root\" (UniqueName: \"kubernetes.io/host-path/528ba802-293a-488e-8856-14b997f7956f-root\") pod \"node-exporter-cw8sk\" (UID: \"528ba802-293a-488e-8856-14b997f7956f\") " pod="openshift-monitoring/node-exporter-cw8sk" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.856490 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/528ba802-293a-488e-8856-14b997f7956f-node-exporter-wtmp\") pod \"node-exporter-cw8sk\" (UID: \"528ba802-293a-488e-8856-14b997f7956f\") " pod="openshift-monitoring/node-exporter-cw8sk" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.856263 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/528ba802-293a-488e-8856-14b997f7956f-sys\") pod \"node-exporter-cw8sk\" (UID: \"528ba802-293a-488e-8856-14b997f7956f\") " pod="openshift-monitoring/node-exporter-cw8sk" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.856880 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/e1147859-11de-422f-9fb6-84ad24e7f49c-metrics-client-ca\") pod \"openshift-state-metrics-566fddb674-xmx2t\" (UID: \"e1147859-11de-422f-9fb6-84ad24e7f49c\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-xmx2t" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.856955 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/e1147859-11de-422f-9fb6-84ad24e7f49c-openshift-state-metrics-tls\") pod \"openshift-state-metrics-566fddb674-xmx2t\" (UID: \"e1147859-11de-422f-9fb6-84ad24e7f49c\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-xmx2t" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.857049 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/528ba802-293a-488e-8856-14b997f7956f-node-exporter-textfile\") pod \"node-exporter-cw8sk\" (UID: \"528ba802-293a-488e-8856-14b997f7956f\") " pod="openshift-monitoring/node-exporter-cw8sk" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.857078 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/5d806c2b-b90f-4541-a9a9-2cf8e86ac947-volume-directive-shadow\") pod \"kube-state-metrics-777cb5bd5d-78wr9\" (UID: \"5d806c2b-b90f-4541-a9a9-2cf8e86ac947\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-78wr9" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.857092 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/5d806c2b-b90f-4541-a9a9-2cf8e86ac947-kube-state-metrics-tls\") pod \"kube-state-metrics-777cb5bd5d-78wr9\" (UID: \"5d806c2b-b90f-4541-a9a9-2cf8e86ac947\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-78wr9" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.857130 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/528ba802-293a-488e-8856-14b997f7956f-node-exporter-tls\") pod \"node-exporter-cw8sk\" (UID: \"528ba802-293a-488e-8856-14b997f7956f\") " pod="openshift-monitoring/node-exporter-cw8sk" Dec 06 05:59:18 crc kubenswrapper[4809]: E1206 05:59:18.857268 4809 secret.go:188] Couldn't get secret openshift-monitoring/node-exporter-tls: secret "node-exporter-tls" not found Dec 06 05:59:18 crc kubenswrapper[4809]: E1206 05:59:18.857313 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/528ba802-293a-488e-8856-14b997f7956f-node-exporter-tls podName:528ba802-293a-488e-8856-14b997f7956f nodeName:}" failed. No retries permitted until 2025-12-06 05:59:19.357299043 +0000 UTC m=+484.246281985 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "node-exporter-tls" (UniqueName: "kubernetes.io/secret/528ba802-293a-488e-8856-14b997f7956f-node-exporter-tls") pod "node-exporter-cw8sk" (UID: "528ba802-293a-488e-8856-14b997f7956f") : secret "node-exporter-tls" not found Dec 06 05:59:18 crc kubenswrapper[4809]: E1206 05:59:18.857434 4809 secret.go:188] Couldn't get secret openshift-monitoring/kube-state-metrics-tls: secret "kube-state-metrics-tls" not found Dec 06 05:59:18 crc kubenswrapper[4809]: E1206 05:59:18.857459 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5d806c2b-b90f-4541-a9a9-2cf8e86ac947-kube-state-metrics-tls podName:5d806c2b-b90f-4541-a9a9-2cf8e86ac947 nodeName:}" failed. No retries permitted until 2025-12-06 05:59:19.357452948 +0000 UTC m=+484.246435890 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-state-metrics-tls" (UniqueName: "kubernetes.io/secret/5d806c2b-b90f-4541-a9a9-2cf8e86ac947-kube-state-metrics-tls") pod "kube-state-metrics-777cb5bd5d-78wr9" (UID: "5d806c2b-b90f-4541-a9a9-2cf8e86ac947") : secret "kube-state-metrics-tls" not found Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.857577 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/5d806c2b-b90f-4541-a9a9-2cf8e86ac947-metrics-client-ca\") pod \"kube-state-metrics-777cb5bd5d-78wr9\" (UID: \"5d806c2b-b90f-4541-a9a9-2cf8e86ac947\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-78wr9" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.857690 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/528ba802-293a-488e-8856-14b997f7956f-node-exporter-textfile\") pod \"node-exporter-cw8sk\" (UID: \"528ba802-293a-488e-8856-14b997f7956f\") " pod="openshift-monitoring/node-exporter-cw8sk" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.857770 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/528ba802-293a-488e-8856-14b997f7956f-metrics-client-ca\") pod \"node-exporter-cw8sk\" (UID: \"528ba802-293a-488e-8856-14b997f7956f\") " pod="openshift-monitoring/node-exporter-cw8sk" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.858092 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/5d806c2b-b90f-4541-a9a9-2cf8e86ac947-volume-directive-shadow\") pod \"kube-state-metrics-777cb5bd5d-78wr9\" (UID: \"5d806c2b-b90f-4541-a9a9-2cf8e86ac947\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-78wr9" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.858537 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/5d806c2b-b90f-4541-a9a9-2cf8e86ac947-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-777cb5bd5d-78wr9\" (UID: \"5d806c2b-b90f-4541-a9a9-2cf8e86ac947\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-78wr9" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.862258 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/e1147859-11de-422f-9fb6-84ad24e7f49c-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-566fddb674-xmx2t\" (UID: \"e1147859-11de-422f-9fb6-84ad24e7f49c\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-xmx2t" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.862568 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/528ba802-293a-488e-8856-14b997f7956f-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-cw8sk\" (UID: \"528ba802-293a-488e-8856-14b997f7956f\") " pod="openshift-monitoring/node-exporter-cw8sk" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.863564 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/5d806c2b-b90f-4541-a9a9-2cf8e86ac947-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-777cb5bd5d-78wr9\" (UID: \"5d806c2b-b90f-4541-a9a9-2cf8e86ac947\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-78wr9" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.864544 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/e1147859-11de-422f-9fb6-84ad24e7f49c-openshift-state-metrics-tls\") pod \"openshift-state-metrics-566fddb674-xmx2t\" (UID: \"e1147859-11de-422f-9fb6-84ad24e7f49c\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-xmx2t" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.874080 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zpxt\" (UniqueName: \"kubernetes.io/projected/e1147859-11de-422f-9fb6-84ad24e7f49c-kube-api-access-4zpxt\") pod \"openshift-state-metrics-566fddb674-xmx2t\" (UID: \"e1147859-11de-422f-9fb6-84ad24e7f49c\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-xmx2t" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.875493 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xt5q4\" (UniqueName: \"kubernetes.io/projected/528ba802-293a-488e-8856-14b997f7956f-kube-api-access-xt5q4\") pod \"node-exporter-cw8sk\" (UID: \"528ba802-293a-488e-8856-14b997f7956f\") " pod="openshift-monitoring/node-exporter-cw8sk" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.876188 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4wnx\" (UniqueName: \"kubernetes.io/projected/5d806c2b-b90f-4541-a9a9-2cf8e86ac947-kube-api-access-j4wnx\") pod \"kube-state-metrics-777cb5bd5d-78wr9\" (UID: \"5d806c2b-b90f-4541-a9a9-2cf8e86ac947\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-78wr9" Dec 06 05:59:18 crc kubenswrapper[4809]: I1206 05:59:18.998455 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/openshift-state-metrics-566fddb674-xmx2t" Dec 06 05:59:19 crc kubenswrapper[4809]: I1206 05:59:19.362053 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/5d806c2b-b90f-4541-a9a9-2cf8e86ac947-kube-state-metrics-tls\") pod \"kube-state-metrics-777cb5bd5d-78wr9\" (UID: \"5d806c2b-b90f-4541-a9a9-2cf8e86ac947\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-78wr9" Dec 06 05:59:19 crc kubenswrapper[4809]: I1206 05:59:19.362350 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/528ba802-293a-488e-8856-14b997f7956f-node-exporter-tls\") pod \"node-exporter-cw8sk\" (UID: \"528ba802-293a-488e-8856-14b997f7956f\") " pod="openshift-monitoring/node-exporter-cw8sk" Dec 06 05:59:19 crc kubenswrapper[4809]: I1206 05:59:19.367313 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/5d806c2b-b90f-4541-a9a9-2cf8e86ac947-kube-state-metrics-tls\") pod \"kube-state-metrics-777cb5bd5d-78wr9\" (UID: \"5d806c2b-b90f-4541-a9a9-2cf8e86ac947\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-78wr9" Dec 06 05:59:19 crc kubenswrapper[4809]: I1206 05:59:19.373680 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/528ba802-293a-488e-8856-14b997f7956f-node-exporter-tls\") pod \"node-exporter-cw8sk\" (UID: \"528ba802-293a-488e-8856-14b997f7956f\") " pod="openshift-monitoring/node-exporter-cw8sk" Dec 06 05:59:19 crc kubenswrapper[4809]: I1206 05:59:19.391367 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-cw8sk" Dec 06 05:59:19 crc kubenswrapper[4809]: I1206 05:59:19.447254 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/openshift-state-metrics-566fddb674-xmx2t"] Dec 06 05:59:19 crc kubenswrapper[4809]: I1206 05:59:19.633443 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-78wr9" Dec 06 05:59:19 crc kubenswrapper[4809]: I1206 05:59:19.743537 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Dec 06 05:59:19 crc kubenswrapper[4809]: I1206 05:59:19.745526 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Dec 06 05:59:19 crc kubenswrapper[4809]: I1206 05:59:19.751266 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-dockercfg-2mdlt" Dec 06 05:59:19 crc kubenswrapper[4809]: I1206 05:59:19.757774 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-tls-assets-0" Dec 06 05:59:19 crc kubenswrapper[4809]: I1206 05:59:19.757820 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-kube-rbac-proxy-web" Dec 06 05:59:19 crc kubenswrapper[4809]: I1206 05:59:19.757894 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-kube-rbac-proxy-metric" Dec 06 05:59:19 crc kubenswrapper[4809]: I1206 05:59:19.757955 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-kube-rbac-proxy" Dec 06 05:59:19 crc kubenswrapper[4809]: I1206 05:59:19.758021 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-generated" Dec 06 05:59:19 crc kubenswrapper[4809]: I1206 05:59:19.758165 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-tls" Dec 06 05:59:19 crc kubenswrapper[4809]: I1206 05:59:19.758316 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-web-config" Dec 06 05:59:19 crc kubenswrapper[4809]: I1206 05:59:19.763281 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"alertmanager-trusted-ca-bundle" Dec 06 05:59:19 crc kubenswrapper[4809]: I1206 05:59:19.777674 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:19.817545 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-xmx2t" event={"ID":"e1147859-11de-422f-9fb6-84ad24e7f49c","Type":"ContainerStarted","Data":"8a4c47a5f934c9bd535964ddb4fc65fbdaa3d1f2e59b6692572c2a8bf1ef16d0"} Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:19.818276 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-cw8sk" event={"ID":"528ba802-293a-488e-8856-14b997f7956f","Type":"ContainerStarted","Data":"789a565b4e01727f1b07b284fa415d161d6325f7cb44a0e210d2d18440faad53"} Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:19.874865 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b-config-out\") pod \"alertmanager-main-0\" (UID: \"cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:19.874916 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:19.874953 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b-config-volume\") pod \"alertmanager-main-0\" (UID: \"cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:19.874975 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b-tls-assets\") pod \"alertmanager-main-0\" (UID: \"cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:19.874999 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fl4z\" (UniqueName: \"kubernetes.io/projected/cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b-kube-api-access-9fl4z\") pod \"alertmanager-main-0\" (UID: \"cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:19.875026 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:19.875050 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:19.875072 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b-web-config\") pod \"alertmanager-main-0\" (UID: \"cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:19.875103 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:19.875118 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:19.875165 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:19.875180 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:19.975925 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:19.975999 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b-config-volume\") pod \"alertmanager-main-0\" (UID: \"cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:19.976024 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b-tls-assets\") pod \"alertmanager-main-0\" (UID: \"cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:19.976053 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fl4z\" (UniqueName: \"kubernetes.io/projected/cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b-kube-api-access-9fl4z\") pod \"alertmanager-main-0\" (UID: \"cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:19.976089 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:19.976125 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:19.976154 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b-web-config\") pod \"alertmanager-main-0\" (UID: \"cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:19.976210 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:19.976230 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:19.976259 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:19.976279 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:19.976304 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b-config-out\") pod \"alertmanager-main-0\" (UID: \"cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:19.977651 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:19.978065 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:19.978411 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:19.981147 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b-config-out\") pod \"alertmanager-main-0\" (UID: \"cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:19.981304 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:19.982040 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b-config-volume\") pod \"alertmanager-main-0\" (UID: \"cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:19.989637 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b-tls-assets\") pod \"alertmanager-main-0\" (UID: \"cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:19.989700 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:19.989803 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:19.989849 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b-web-config\") pod \"alertmanager-main-0\" (UID: \"cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:19.989866 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:19.999810 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fl4z\" (UniqueName: \"kubernetes.io/projected/cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b-kube-api-access-9fl4z\") pod \"alertmanager-main-0\" (UID: \"cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:20.069493 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:20.084461 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/kube-state-metrics-777cb5bd5d-78wr9"] Dec 06 05:59:22 crc kubenswrapper[4809]: W1206 05:59:20.091986 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5d806c2b_b90f_4541_a9a9_2cf8e86ac947.slice/crio-f03de5cdf8ef8dedeb94449ec38721470c03aea80459b1f60877ff61d41a3ae9 WatchSource:0}: Error finding container f03de5cdf8ef8dedeb94449ec38721470c03aea80459b1f60877ff61d41a3ae9: Status 404 returned error can't find the container with id f03de5cdf8ef8dedeb94449ec38721470c03aea80459b1f60877ff61d41a3ae9 Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:20.657024 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/thanos-querier-d74dd8556-59sjd"] Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:20.658862 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/thanos-querier-d74dd8556-59sjd" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:20.660738 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-grpc-tls-b91cvo0hiehkl" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:20.661283 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy-web" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:20.661516 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-dockercfg-4twrj" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:20.661670 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-tls" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:20.661719 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy-metrics" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:20.661862 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:20.662013 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy-rules" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:20.676092 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/thanos-querier-d74dd8556-59sjd"] Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:20.787715 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/ec930236-fceb-44dd-90c3-13a25d7cd9f2-secret-grpc-tls\") pod \"thanos-querier-d74dd8556-59sjd\" (UID: \"ec930236-fceb-44dd-90c3-13a25d7cd9f2\") " pod="openshift-monitoring/thanos-querier-d74dd8556-59sjd" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:20.787760 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/ec930236-fceb-44dd-90c3-13a25d7cd9f2-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-d74dd8556-59sjd\" (UID: \"ec930236-fceb-44dd-90c3-13a25d7cd9f2\") " pod="openshift-monitoring/thanos-querier-d74dd8556-59sjd" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:20.787783 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/ec930236-fceb-44dd-90c3-13a25d7cd9f2-metrics-client-ca\") pod \"thanos-querier-d74dd8556-59sjd\" (UID: \"ec930236-fceb-44dd-90c3-13a25d7cd9f2\") " pod="openshift-monitoring/thanos-querier-d74dd8556-59sjd" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:20.787802 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/ec930236-fceb-44dd-90c3-13a25d7cd9f2-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-d74dd8556-59sjd\" (UID: \"ec930236-fceb-44dd-90c3-13a25d7cd9f2\") " pod="openshift-monitoring/thanos-querier-d74dd8556-59sjd" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:20.787871 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/ec930236-fceb-44dd-90c3-13a25d7cd9f2-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-d74dd8556-59sjd\" (UID: \"ec930236-fceb-44dd-90c3-13a25d7cd9f2\") " pod="openshift-monitoring/thanos-querier-d74dd8556-59sjd" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:20.787907 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fszs\" (UniqueName: \"kubernetes.io/projected/ec930236-fceb-44dd-90c3-13a25d7cd9f2-kube-api-access-9fszs\") pod \"thanos-querier-d74dd8556-59sjd\" (UID: \"ec930236-fceb-44dd-90c3-13a25d7cd9f2\") " pod="openshift-monitoring/thanos-querier-d74dd8556-59sjd" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:20.787945 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/ec930236-fceb-44dd-90c3-13a25d7cd9f2-secret-thanos-querier-tls\") pod \"thanos-querier-d74dd8556-59sjd\" (UID: \"ec930236-fceb-44dd-90c3-13a25d7cd9f2\") " pod="openshift-monitoring/thanos-querier-d74dd8556-59sjd" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:20.788040 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/ec930236-fceb-44dd-90c3-13a25d7cd9f2-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-d74dd8556-59sjd\" (UID: \"ec930236-fceb-44dd-90c3-13a25d7cd9f2\") " pod="openshift-monitoring/thanos-querier-d74dd8556-59sjd" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:20.828582 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-78wr9" event={"ID":"5d806c2b-b90f-4541-a9a9-2cf8e86ac947","Type":"ContainerStarted","Data":"f03de5cdf8ef8dedeb94449ec38721470c03aea80459b1f60877ff61d41a3ae9"} Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:20.830926 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-xmx2t" event={"ID":"e1147859-11de-422f-9fb6-84ad24e7f49c","Type":"ContainerStarted","Data":"26425accdc10e113c32a7fc9e601fb60090f23379c4f8d00b191b5c0662ff456"} Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:20.888916 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/ec930236-fceb-44dd-90c3-13a25d7cd9f2-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-d74dd8556-59sjd\" (UID: \"ec930236-fceb-44dd-90c3-13a25d7cd9f2\") " pod="openshift-monitoring/thanos-querier-d74dd8556-59sjd" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:20.888986 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/ec930236-fceb-44dd-90c3-13a25d7cd9f2-secret-grpc-tls\") pod \"thanos-querier-d74dd8556-59sjd\" (UID: \"ec930236-fceb-44dd-90c3-13a25d7cd9f2\") " pod="openshift-monitoring/thanos-querier-d74dd8556-59sjd" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:20.889022 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/ec930236-fceb-44dd-90c3-13a25d7cd9f2-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-d74dd8556-59sjd\" (UID: \"ec930236-fceb-44dd-90c3-13a25d7cd9f2\") " pod="openshift-monitoring/thanos-querier-d74dd8556-59sjd" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:20.889041 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/ec930236-fceb-44dd-90c3-13a25d7cd9f2-metrics-client-ca\") pod \"thanos-querier-d74dd8556-59sjd\" (UID: \"ec930236-fceb-44dd-90c3-13a25d7cd9f2\") " pod="openshift-monitoring/thanos-querier-d74dd8556-59sjd" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:20.889063 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/ec930236-fceb-44dd-90c3-13a25d7cd9f2-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-d74dd8556-59sjd\" (UID: \"ec930236-fceb-44dd-90c3-13a25d7cd9f2\") " pod="openshift-monitoring/thanos-querier-d74dd8556-59sjd" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:20.889083 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/ec930236-fceb-44dd-90c3-13a25d7cd9f2-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-d74dd8556-59sjd\" (UID: \"ec930236-fceb-44dd-90c3-13a25d7cd9f2\") " pod="openshift-monitoring/thanos-querier-d74dd8556-59sjd" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:20.889102 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fszs\" (UniqueName: \"kubernetes.io/projected/ec930236-fceb-44dd-90c3-13a25d7cd9f2-kube-api-access-9fszs\") pod \"thanos-querier-d74dd8556-59sjd\" (UID: \"ec930236-fceb-44dd-90c3-13a25d7cd9f2\") " pod="openshift-monitoring/thanos-querier-d74dd8556-59sjd" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:20.889121 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/ec930236-fceb-44dd-90c3-13a25d7cd9f2-secret-thanos-querier-tls\") pod \"thanos-querier-d74dd8556-59sjd\" (UID: \"ec930236-fceb-44dd-90c3-13a25d7cd9f2\") " pod="openshift-monitoring/thanos-querier-d74dd8556-59sjd" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:20.890173 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/ec930236-fceb-44dd-90c3-13a25d7cd9f2-metrics-client-ca\") pod \"thanos-querier-d74dd8556-59sjd\" (UID: \"ec930236-fceb-44dd-90c3-13a25d7cd9f2\") " pod="openshift-monitoring/thanos-querier-d74dd8556-59sjd" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:20.898136 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/ec930236-fceb-44dd-90c3-13a25d7cd9f2-secret-grpc-tls\") pod \"thanos-querier-d74dd8556-59sjd\" (UID: \"ec930236-fceb-44dd-90c3-13a25d7cd9f2\") " pod="openshift-monitoring/thanos-querier-d74dd8556-59sjd" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:20.903431 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/ec930236-fceb-44dd-90c3-13a25d7cd9f2-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-d74dd8556-59sjd\" (UID: \"ec930236-fceb-44dd-90c3-13a25d7cd9f2\") " pod="openshift-monitoring/thanos-querier-d74dd8556-59sjd" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:20.903729 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/ec930236-fceb-44dd-90c3-13a25d7cd9f2-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-d74dd8556-59sjd\" (UID: \"ec930236-fceb-44dd-90c3-13a25d7cd9f2\") " pod="openshift-monitoring/thanos-querier-d74dd8556-59sjd" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:20.903770 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/ec930236-fceb-44dd-90c3-13a25d7cd9f2-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-d74dd8556-59sjd\" (UID: \"ec930236-fceb-44dd-90c3-13a25d7cd9f2\") " pod="openshift-monitoring/thanos-querier-d74dd8556-59sjd" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:20.904025 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/ec930236-fceb-44dd-90c3-13a25d7cd9f2-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-d74dd8556-59sjd\" (UID: \"ec930236-fceb-44dd-90c3-13a25d7cd9f2\") " pod="openshift-monitoring/thanos-querier-d74dd8556-59sjd" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:20.911150 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fszs\" (UniqueName: \"kubernetes.io/projected/ec930236-fceb-44dd-90c3-13a25d7cd9f2-kube-api-access-9fszs\") pod \"thanos-querier-d74dd8556-59sjd\" (UID: \"ec930236-fceb-44dd-90c3-13a25d7cd9f2\") " pod="openshift-monitoring/thanos-querier-d74dd8556-59sjd" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:20.912069 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/ec930236-fceb-44dd-90c3-13a25d7cd9f2-secret-thanos-querier-tls\") pod \"thanos-querier-d74dd8556-59sjd\" (UID: \"ec930236-fceb-44dd-90c3-13a25d7cd9f2\") " pod="openshift-monitoring/thanos-querier-d74dd8556-59sjd" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:20.976579 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/thanos-querier-d74dd8556-59sjd" Dec 06 05:59:22 crc kubenswrapper[4809]: I1206 05:59:22.844776 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-xmx2t" event={"ID":"e1147859-11de-422f-9fb6-84ad24e7f49c","Type":"ContainerStarted","Data":"5e83c615db602354bac1ac6c4c80ed8ea44f78ca08e32a32115a8f160164ce4c"} Dec 06 05:59:23 crc kubenswrapper[4809]: I1206 05:59:23.101433 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/thanos-querier-d74dd8556-59sjd"] Dec 06 05:59:23 crc kubenswrapper[4809]: I1206 05:59:23.104811 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Dec 06 05:59:23 crc kubenswrapper[4809]: W1206 05:59:23.107431 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podec930236_fceb_44dd_90c3_13a25d7cd9f2.slice/crio-d59a0f1996090d7539777754dc098c06f545bd8a40be8a3a65e201d21852d55d WatchSource:0}: Error finding container d59a0f1996090d7539777754dc098c06f545bd8a40be8a3a65e201d21852d55d: Status 404 returned error can't find the container with id d59a0f1996090d7539777754dc098c06f545bd8a40be8a3a65e201d21852d55d Dec 06 05:59:23 crc kubenswrapper[4809]: W1206 05:59:23.108218 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcff2bf4d_ccb5_46ed_a2fc_38ddbb4a9c4b.slice/crio-2eb0f04761ba9cac66c3a18dba99055c0e2479158e7364adc09a646bb544b06b WatchSource:0}: Error finding container 2eb0f04761ba9cac66c3a18dba99055c0e2479158e7364adc09a646bb544b06b: Status 404 returned error can't find the container with id 2eb0f04761ba9cac66c3a18dba99055c0e2479158e7364adc09a646bb544b06b Dec 06 05:59:23 crc kubenswrapper[4809]: I1206 05:59:23.853372 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b","Type":"ContainerStarted","Data":"2eb0f04761ba9cac66c3a18dba99055c0e2479158e7364adc09a646bb544b06b"} Dec 06 05:59:23 crc kubenswrapper[4809]: I1206 05:59:23.855629 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-d74dd8556-59sjd" event={"ID":"ec930236-fceb-44dd-90c3-13a25d7cd9f2","Type":"ContainerStarted","Data":"d59a0f1996090d7539777754dc098c06f545bd8a40be8a3a65e201d21852d55d"} Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.092219 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/metrics-server-556bbc5579-8wblw"] Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.093339 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/metrics-server-556bbc5579-8wblw" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.095780 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-client-certs" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.096566 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"kubelet-serving-ca-bundle" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.097243 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-server-dockercfg-g4cq5" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.097604 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-server-8i541qel85nq1" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.097265 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-server-tls" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.097314 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"metrics-server-audit-profiles" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.116608 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/metrics-server-556bbc5579-8wblw"] Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.142611 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4f96141-d70e-4715-9e30-516c9e783cc9-client-ca-bundle\") pod \"metrics-server-556bbc5579-8wblw\" (UID: \"b4f96141-d70e-4715-9e30-516c9e783cc9\") " pod="openshift-monitoring/metrics-server-556bbc5579-8wblw" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.142874 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/b4f96141-d70e-4715-9e30-516c9e783cc9-audit-log\") pod \"metrics-server-556bbc5579-8wblw\" (UID: \"b4f96141-d70e-4715-9e30-516c9e783cc9\") " pod="openshift-monitoring/metrics-server-556bbc5579-8wblw" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.143033 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b4f96141-d70e-4715-9e30-516c9e783cc9-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-556bbc5579-8wblw\" (UID: \"b4f96141-d70e-4715-9e30-516c9e783cc9\") " pod="openshift-monitoring/metrics-server-556bbc5579-8wblw" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.143164 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqqvm\" (UniqueName: \"kubernetes.io/projected/b4f96141-d70e-4715-9e30-516c9e783cc9-kube-api-access-nqqvm\") pod \"metrics-server-556bbc5579-8wblw\" (UID: \"b4f96141-d70e-4715-9e30-516c9e783cc9\") " pod="openshift-monitoring/metrics-server-556bbc5579-8wblw" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.143262 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/b4f96141-d70e-4715-9e30-516c9e783cc9-metrics-server-audit-profiles\") pod \"metrics-server-556bbc5579-8wblw\" (UID: \"b4f96141-d70e-4715-9e30-516c9e783cc9\") " pod="openshift-monitoring/metrics-server-556bbc5579-8wblw" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.143344 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/b4f96141-d70e-4715-9e30-516c9e783cc9-secret-metrics-client-certs\") pod \"metrics-server-556bbc5579-8wblw\" (UID: \"b4f96141-d70e-4715-9e30-516c9e783cc9\") " pod="openshift-monitoring/metrics-server-556bbc5579-8wblw" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.143446 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/b4f96141-d70e-4715-9e30-516c9e783cc9-secret-metrics-server-tls\") pod \"metrics-server-556bbc5579-8wblw\" (UID: \"b4f96141-d70e-4715-9e30-516c9e783cc9\") " pod="openshift-monitoring/metrics-server-556bbc5579-8wblw" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.244434 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/b4f96141-d70e-4715-9e30-516c9e783cc9-secret-metrics-server-tls\") pod \"metrics-server-556bbc5579-8wblw\" (UID: \"b4f96141-d70e-4715-9e30-516c9e783cc9\") " pod="openshift-monitoring/metrics-server-556bbc5579-8wblw" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.244784 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4f96141-d70e-4715-9e30-516c9e783cc9-client-ca-bundle\") pod \"metrics-server-556bbc5579-8wblw\" (UID: \"b4f96141-d70e-4715-9e30-516c9e783cc9\") " pod="openshift-monitoring/metrics-server-556bbc5579-8wblw" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.245016 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/b4f96141-d70e-4715-9e30-516c9e783cc9-audit-log\") pod \"metrics-server-556bbc5579-8wblw\" (UID: \"b4f96141-d70e-4715-9e30-516c9e783cc9\") " pod="openshift-monitoring/metrics-server-556bbc5579-8wblw" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.245305 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b4f96141-d70e-4715-9e30-516c9e783cc9-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-556bbc5579-8wblw\" (UID: \"b4f96141-d70e-4715-9e30-516c9e783cc9\") " pod="openshift-monitoring/metrics-server-556bbc5579-8wblw" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.245528 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqqvm\" (UniqueName: \"kubernetes.io/projected/b4f96141-d70e-4715-9e30-516c9e783cc9-kube-api-access-nqqvm\") pod \"metrics-server-556bbc5579-8wblw\" (UID: \"b4f96141-d70e-4715-9e30-516c9e783cc9\") " pod="openshift-monitoring/metrics-server-556bbc5579-8wblw" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.245743 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/b4f96141-d70e-4715-9e30-516c9e783cc9-metrics-server-audit-profiles\") pod \"metrics-server-556bbc5579-8wblw\" (UID: \"b4f96141-d70e-4715-9e30-516c9e783cc9\") " pod="openshift-monitoring/metrics-server-556bbc5579-8wblw" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.246001 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/b4f96141-d70e-4715-9e30-516c9e783cc9-secret-metrics-client-certs\") pod \"metrics-server-556bbc5579-8wblw\" (UID: \"b4f96141-d70e-4715-9e30-516c9e783cc9\") " pod="openshift-monitoring/metrics-server-556bbc5579-8wblw" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.245657 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/b4f96141-d70e-4715-9e30-516c9e783cc9-audit-log\") pod \"metrics-server-556bbc5579-8wblw\" (UID: \"b4f96141-d70e-4715-9e30-516c9e783cc9\") " pod="openshift-monitoring/metrics-server-556bbc5579-8wblw" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.247061 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/b4f96141-d70e-4715-9e30-516c9e783cc9-metrics-server-audit-profiles\") pod \"metrics-server-556bbc5579-8wblw\" (UID: \"b4f96141-d70e-4715-9e30-516c9e783cc9\") " pod="openshift-monitoring/metrics-server-556bbc5579-8wblw" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.247433 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b4f96141-d70e-4715-9e30-516c9e783cc9-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-556bbc5579-8wblw\" (UID: \"b4f96141-d70e-4715-9e30-516c9e783cc9\") " pod="openshift-monitoring/metrics-server-556bbc5579-8wblw" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.250160 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/b4f96141-d70e-4715-9e30-516c9e783cc9-secret-metrics-server-tls\") pod \"metrics-server-556bbc5579-8wblw\" (UID: \"b4f96141-d70e-4715-9e30-516c9e783cc9\") " pod="openshift-monitoring/metrics-server-556bbc5579-8wblw" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.252079 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4f96141-d70e-4715-9e30-516c9e783cc9-client-ca-bundle\") pod \"metrics-server-556bbc5579-8wblw\" (UID: \"b4f96141-d70e-4715-9e30-516c9e783cc9\") " pod="openshift-monitoring/metrics-server-556bbc5579-8wblw" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.252906 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/b4f96141-d70e-4715-9e30-516c9e783cc9-secret-metrics-client-certs\") pod \"metrics-server-556bbc5579-8wblw\" (UID: \"b4f96141-d70e-4715-9e30-516c9e783cc9\") " pod="openshift-monitoring/metrics-server-556bbc5579-8wblw" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.276629 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqqvm\" (UniqueName: \"kubernetes.io/projected/b4f96141-d70e-4715-9e30-516c9e783cc9-kube-api-access-nqqvm\") pod \"metrics-server-556bbc5579-8wblw\" (UID: \"b4f96141-d70e-4715-9e30-516c9e783cc9\") " pod="openshift-monitoring/metrics-server-556bbc5579-8wblw" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.329436 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-7dc5585756-dv7bd"] Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.330403 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7dc5585756-dv7bd" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.347831 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txb97\" (UniqueName: \"kubernetes.io/projected/0a07f515-be4e-4978-bd8b-ffac881b7a3f-kube-api-access-txb97\") pod \"console-7dc5585756-dv7bd\" (UID: \"0a07f515-be4e-4978-bd8b-ffac881b7a3f\") " pod="openshift-console/console-7dc5585756-dv7bd" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.347971 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0a07f515-be4e-4978-bd8b-ffac881b7a3f-service-ca\") pod \"console-7dc5585756-dv7bd\" (UID: \"0a07f515-be4e-4978-bd8b-ffac881b7a3f\") " pod="openshift-console/console-7dc5585756-dv7bd" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.348069 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0a07f515-be4e-4978-bd8b-ffac881b7a3f-trusted-ca-bundle\") pod \"console-7dc5585756-dv7bd\" (UID: \"0a07f515-be4e-4978-bd8b-ffac881b7a3f\") " pod="openshift-console/console-7dc5585756-dv7bd" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.348188 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0a07f515-be4e-4978-bd8b-ffac881b7a3f-console-config\") pod \"console-7dc5585756-dv7bd\" (UID: \"0a07f515-be4e-4978-bd8b-ffac881b7a3f\") " pod="openshift-console/console-7dc5585756-dv7bd" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.348258 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0a07f515-be4e-4978-bd8b-ffac881b7a3f-console-oauth-config\") pod \"console-7dc5585756-dv7bd\" (UID: \"0a07f515-be4e-4978-bd8b-ffac881b7a3f\") " pod="openshift-console/console-7dc5585756-dv7bd" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.348324 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0a07f515-be4e-4978-bd8b-ffac881b7a3f-console-serving-cert\") pod \"console-7dc5585756-dv7bd\" (UID: \"0a07f515-be4e-4978-bd8b-ffac881b7a3f\") " pod="openshift-console/console-7dc5585756-dv7bd" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.348429 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0a07f515-be4e-4978-bd8b-ffac881b7a3f-oauth-serving-cert\") pod \"console-7dc5585756-dv7bd\" (UID: \"0a07f515-be4e-4978-bd8b-ffac881b7a3f\") " pod="openshift-console/console-7dc5585756-dv7bd" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.348480 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7dc5585756-dv7bd"] Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.421754 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/metrics-server-556bbc5579-8wblw" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.450185 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0a07f515-be4e-4978-bd8b-ffac881b7a3f-trusted-ca-bundle\") pod \"console-7dc5585756-dv7bd\" (UID: \"0a07f515-be4e-4978-bd8b-ffac881b7a3f\") " pod="openshift-console/console-7dc5585756-dv7bd" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.450257 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0a07f515-be4e-4978-bd8b-ffac881b7a3f-console-config\") pod \"console-7dc5585756-dv7bd\" (UID: \"0a07f515-be4e-4978-bd8b-ffac881b7a3f\") " pod="openshift-console/console-7dc5585756-dv7bd" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.450287 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0a07f515-be4e-4978-bd8b-ffac881b7a3f-console-oauth-config\") pod \"console-7dc5585756-dv7bd\" (UID: \"0a07f515-be4e-4978-bd8b-ffac881b7a3f\") " pod="openshift-console/console-7dc5585756-dv7bd" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.450314 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0a07f515-be4e-4978-bd8b-ffac881b7a3f-console-serving-cert\") pod \"console-7dc5585756-dv7bd\" (UID: \"0a07f515-be4e-4978-bd8b-ffac881b7a3f\") " pod="openshift-console/console-7dc5585756-dv7bd" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.450365 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0a07f515-be4e-4978-bd8b-ffac881b7a3f-oauth-serving-cert\") pod \"console-7dc5585756-dv7bd\" (UID: \"0a07f515-be4e-4978-bd8b-ffac881b7a3f\") " pod="openshift-console/console-7dc5585756-dv7bd" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.450383 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txb97\" (UniqueName: \"kubernetes.io/projected/0a07f515-be4e-4978-bd8b-ffac881b7a3f-kube-api-access-txb97\") pod \"console-7dc5585756-dv7bd\" (UID: \"0a07f515-be4e-4978-bd8b-ffac881b7a3f\") " pod="openshift-console/console-7dc5585756-dv7bd" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.450418 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0a07f515-be4e-4978-bd8b-ffac881b7a3f-service-ca\") pod \"console-7dc5585756-dv7bd\" (UID: \"0a07f515-be4e-4978-bd8b-ffac881b7a3f\") " pod="openshift-console/console-7dc5585756-dv7bd" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.451471 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0a07f515-be4e-4978-bd8b-ffac881b7a3f-service-ca\") pod \"console-7dc5585756-dv7bd\" (UID: \"0a07f515-be4e-4978-bd8b-ffac881b7a3f\") " pod="openshift-console/console-7dc5585756-dv7bd" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.452092 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0a07f515-be4e-4978-bd8b-ffac881b7a3f-console-config\") pod \"console-7dc5585756-dv7bd\" (UID: \"0a07f515-be4e-4978-bd8b-ffac881b7a3f\") " pod="openshift-console/console-7dc5585756-dv7bd" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.454290 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0a07f515-be4e-4978-bd8b-ffac881b7a3f-oauth-serving-cert\") pod \"console-7dc5585756-dv7bd\" (UID: \"0a07f515-be4e-4978-bd8b-ffac881b7a3f\") " pod="openshift-console/console-7dc5585756-dv7bd" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.455617 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0a07f515-be4e-4978-bd8b-ffac881b7a3f-trusted-ca-bundle\") pod \"console-7dc5585756-dv7bd\" (UID: \"0a07f515-be4e-4978-bd8b-ffac881b7a3f\") " pod="openshift-console/console-7dc5585756-dv7bd" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.458410 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0a07f515-be4e-4978-bd8b-ffac881b7a3f-console-oauth-config\") pod \"console-7dc5585756-dv7bd\" (UID: \"0a07f515-be4e-4978-bd8b-ffac881b7a3f\") " pod="openshift-console/console-7dc5585756-dv7bd" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.462553 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0a07f515-be4e-4978-bd8b-ffac881b7a3f-console-serving-cert\") pod \"console-7dc5585756-dv7bd\" (UID: \"0a07f515-be4e-4978-bd8b-ffac881b7a3f\") " pod="openshift-console/console-7dc5585756-dv7bd" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.478701 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txb97\" (UniqueName: \"kubernetes.io/projected/0a07f515-be4e-4978-bd8b-ffac881b7a3f-kube-api-access-txb97\") pod \"console-7dc5585756-dv7bd\" (UID: \"0a07f515-be4e-4978-bd8b-ffac881b7a3f\") " pod="openshift-console/console-7dc5585756-dv7bd" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.488040 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/monitoring-plugin-5998c84788-zjqnq"] Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.488870 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/monitoring-plugin-5998c84788-zjqnq" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.491352 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"default-dockercfg-6tstp" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.492677 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"monitoring-plugin-cert" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.493894 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/monitoring-plugin-5998c84788-zjqnq"] Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.552923 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/bb66c253-f1a0-4412-93f7-82b8fb03fb33-monitoring-plugin-cert\") pod \"monitoring-plugin-5998c84788-zjqnq\" (UID: \"bb66c253-f1a0-4412-93f7-82b8fb03fb33\") " pod="openshift-monitoring/monitoring-plugin-5998c84788-zjqnq" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.654008 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7dc5585756-dv7bd" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.654192 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/bb66c253-f1a0-4412-93f7-82b8fb03fb33-monitoring-plugin-cert\") pod \"monitoring-plugin-5998c84788-zjqnq\" (UID: \"bb66c253-f1a0-4412-93f7-82b8fb03fb33\") " pod="openshift-monitoring/monitoring-plugin-5998c84788-zjqnq" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.658610 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/bb66c253-f1a0-4412-93f7-82b8fb03fb33-monitoring-plugin-cert\") pod \"monitoring-plugin-5998c84788-zjqnq\" (UID: \"bb66c253-f1a0-4412-93f7-82b8fb03fb33\") " pod="openshift-monitoring/monitoring-plugin-5998c84788-zjqnq" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.719641 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/metrics-server-556bbc5579-8wblw"] Dec 06 05:59:24 crc kubenswrapper[4809]: W1206 05:59:24.736371 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4f96141_d70e_4715_9e30_516c9e783cc9.slice/crio-10de1a6b1cd16889197a3cc256f859f32bb0d71309f0b6c0d610ac672a06fd40 WatchSource:0}: Error finding container 10de1a6b1cd16889197a3cc256f859f32bb0d71309f0b6c0d610ac672a06fd40: Status 404 returned error can't find the container with id 10de1a6b1cd16889197a3cc256f859f32bb0d71309f0b6c0d610ac672a06fd40 Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.825440 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/monitoring-plugin-5998c84788-zjqnq" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.866605 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-556bbc5579-8wblw" event={"ID":"b4f96141-d70e-4715-9e30-516c9e783cc9","Type":"ContainerStarted","Data":"10de1a6b1cd16889197a3cc256f859f32bb0d71309f0b6c0d610ac672a06fd40"} Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.988651 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.993161 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.998856 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-rbac-proxy" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.999113 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-thanos-sidecar-tls" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.999329 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s" Dec 06 05:59:24 crc kubenswrapper[4809]: I1206 05:59:24.999401 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-thanos-prometheus-http-client-file" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.000264 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-dockercfg-dsptv" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.000311 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-kube-rbac-proxy-web" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.000489 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"serving-certs-ca-bundle" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.000800 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-grpc-tls-dmjms8q80klni" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.001970 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-tls" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.002410 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-web-config" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.002596 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-tls-assets-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.028992 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"prometheus-k8s-rulefiles-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.032637 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"prometheus-trusted-ca-bundle" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.041689 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.065598 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7dc5585756-dv7bd"] Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.067856 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-679cn\" (UniqueName: \"kubernetes.io/projected/869feda2-ec9d-44de-9533-27a4d7b990c1-kube-api-access-679cn\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.067909 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/869feda2-ec9d-44de-9533-27a4d7b990c1-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.067979 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/869feda2-ec9d-44de-9533-27a4d7b990c1-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.068000 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/869feda2-ec9d-44de-9533-27a4d7b990c1-web-config\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.068029 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/869feda2-ec9d-44de-9533-27a4d7b990c1-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.068044 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/869feda2-ec9d-44de-9533-27a4d7b990c1-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.068102 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/869feda2-ec9d-44de-9533-27a4d7b990c1-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.068180 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/869feda2-ec9d-44de-9533-27a4d7b990c1-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.068218 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/869feda2-ec9d-44de-9533-27a4d7b990c1-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.068243 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/869feda2-ec9d-44de-9533-27a4d7b990c1-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.068261 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/869feda2-ec9d-44de-9533-27a4d7b990c1-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.068288 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/869feda2-ec9d-44de-9533-27a4d7b990c1-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.068346 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/869feda2-ec9d-44de-9533-27a4d7b990c1-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.068365 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/869feda2-ec9d-44de-9533-27a4d7b990c1-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.068379 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/869feda2-ec9d-44de-9533-27a4d7b990c1-config-out\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.068393 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/869feda2-ec9d-44de-9533-27a4d7b990c1-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.068423 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/869feda2-ec9d-44de-9533-27a4d7b990c1-config\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.068455 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/869feda2-ec9d-44de-9533-27a4d7b990c1-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.169689 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/869feda2-ec9d-44de-9533-27a4d7b990c1-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.169747 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/869feda2-ec9d-44de-9533-27a4d7b990c1-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.169774 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/869feda2-ec9d-44de-9533-27a4d7b990c1-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.169805 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/869feda2-ec9d-44de-9533-27a4d7b990c1-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.169837 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/869feda2-ec9d-44de-9533-27a4d7b990c1-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.169869 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/869feda2-ec9d-44de-9533-27a4d7b990c1-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.169915 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/869feda2-ec9d-44de-9533-27a4d7b990c1-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.169959 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/869feda2-ec9d-44de-9533-27a4d7b990c1-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.169979 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/869feda2-ec9d-44de-9533-27a4d7b990c1-config-out\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.170000 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/869feda2-ec9d-44de-9533-27a4d7b990c1-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.170028 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/869feda2-ec9d-44de-9533-27a4d7b990c1-config\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.170056 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/869feda2-ec9d-44de-9533-27a4d7b990c1-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.170083 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-679cn\" (UniqueName: \"kubernetes.io/projected/869feda2-ec9d-44de-9533-27a4d7b990c1-kube-api-access-679cn\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.170106 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/869feda2-ec9d-44de-9533-27a4d7b990c1-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.170138 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/869feda2-ec9d-44de-9533-27a4d7b990c1-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.170158 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/869feda2-ec9d-44de-9533-27a4d7b990c1-web-config\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.170183 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/869feda2-ec9d-44de-9533-27a4d7b990c1-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.170201 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/869feda2-ec9d-44de-9533-27a4d7b990c1-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.171224 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/869feda2-ec9d-44de-9533-27a4d7b990c1-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.171236 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/869feda2-ec9d-44de-9533-27a4d7b990c1-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.177621 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/869feda2-ec9d-44de-9533-27a4d7b990c1-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.178028 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/869feda2-ec9d-44de-9533-27a4d7b990c1-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.178192 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/869feda2-ec9d-44de-9533-27a4d7b990c1-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.178496 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/869feda2-ec9d-44de-9533-27a4d7b990c1-config\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.178737 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/869feda2-ec9d-44de-9533-27a4d7b990c1-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.178959 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/869feda2-ec9d-44de-9533-27a4d7b990c1-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.181486 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/869feda2-ec9d-44de-9533-27a4d7b990c1-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.181773 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/869feda2-ec9d-44de-9533-27a4d7b990c1-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.181734 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/869feda2-ec9d-44de-9533-27a4d7b990c1-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.181519 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/869feda2-ec9d-44de-9533-27a4d7b990c1-config-out\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.181991 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/869feda2-ec9d-44de-9533-27a4d7b990c1-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.182416 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/869feda2-ec9d-44de-9533-27a4d7b990c1-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.182550 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/869feda2-ec9d-44de-9533-27a4d7b990c1-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.182572 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/869feda2-ec9d-44de-9533-27a4d7b990c1-web-config\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.184404 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/869feda2-ec9d-44de-9533-27a4d7b990c1-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.189074 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-679cn\" (UniqueName: \"kubernetes.io/projected/869feda2-ec9d-44de-9533-27a4d7b990c1-kube-api-access-679cn\") pod \"prometheus-k8s-0\" (UID: \"869feda2-ec9d-44de-9533-27a4d7b990c1\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.296454 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/monitoring-plugin-5998c84788-zjqnq"] Dec 06 05:59:25 crc kubenswrapper[4809]: W1206 05:59:25.298733 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb66c253_f1a0_4412_93f7_82b8fb03fb33.slice/crio-85b79d73077fe12efa8f9a5b4b5a1cc5420c26a7eed9b2a538000412993de349 WatchSource:0}: Error finding container 85b79d73077fe12efa8f9a5b4b5a1cc5420c26a7eed9b2a538000412993de349: Status 404 returned error can't find the container with id 85b79d73077fe12efa8f9a5b4b5a1cc5420c26a7eed9b2a538000412993de349 Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.320185 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.783710 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.873512 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"869feda2-ec9d-44de-9533-27a4d7b990c1","Type":"ContainerStarted","Data":"8cfbe1e2c315a72762222be3580af793ea4b28921f591c7d7c2f443d36337994"} Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.874764 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/monitoring-plugin-5998c84788-zjqnq" event={"ID":"bb66c253-f1a0-4412-93f7-82b8fb03fb33","Type":"ContainerStarted","Data":"85b79d73077fe12efa8f9a5b4b5a1cc5420c26a7eed9b2a538000412993de349"} Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.876151 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-78wr9" event={"ID":"5d806c2b-b90f-4541-a9a9-2cf8e86ac947","Type":"ContainerStarted","Data":"e660253c11cb7547d9db9ebfb8f8d7c103e827498a562dcc739e1790ee4e8f0c"} Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.877236 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7dc5585756-dv7bd" event={"ID":"0a07f515-be4e-4978-bd8b-ffac881b7a3f","Type":"ContainerStarted","Data":"795c66f39956b918963aaa2311783d6be38583742482763abbd4629b96013ab5"} Dec 06 05:59:25 crc kubenswrapper[4809]: I1206 05:59:25.879182 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-cw8sk" event={"ID":"528ba802-293a-488e-8856-14b997f7956f","Type":"ContainerStarted","Data":"28fabd140e10fe5a3cbe42cee38be51ecf666e90cf88ade061c3370a8ca1c4c5"} Dec 06 05:59:26 crc kubenswrapper[4809]: I1206 05:59:26.896475 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-78wr9" event={"ID":"5d806c2b-b90f-4541-a9a9-2cf8e86ac947","Type":"ContainerStarted","Data":"b20cf2f171d24c23bd0af5fcc3fb8a6a6662a70b06443e88f1dcd76364507f08"} Dec 06 05:59:26 crc kubenswrapper[4809]: I1206 05:59:26.896939 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-78wr9" event={"ID":"5d806c2b-b90f-4541-a9a9-2cf8e86ac947","Type":"ContainerStarted","Data":"e88fdd12b3ed5a23ae3fa509929c1d8ccfdc0d1e37cea2cd87bfc5bcd895e982"} Dec 06 05:59:26 crc kubenswrapper[4809]: I1206 05:59:26.901090 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7dc5585756-dv7bd" event={"ID":"0a07f515-be4e-4978-bd8b-ffac881b7a3f","Type":"ContainerStarted","Data":"b47eaa725139cb91ace9f7260c25c06f4a0be75487d9511699bf9058c5316d54"} Dec 06 05:59:26 crc kubenswrapper[4809]: I1206 05:59:26.904823 4809 generic.go:334] "Generic (PLEG): container finished" podID="528ba802-293a-488e-8856-14b997f7956f" containerID="28fabd140e10fe5a3cbe42cee38be51ecf666e90cf88ade061c3370a8ca1c4c5" exitCode=0 Dec 06 05:59:26 crc kubenswrapper[4809]: I1206 05:59:26.904882 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-cw8sk" event={"ID":"528ba802-293a-488e-8856-14b997f7956f","Type":"ContainerDied","Data":"28fabd140e10fe5a3cbe42cee38be51ecf666e90cf88ade061c3370a8ca1c4c5"} Dec 06 05:59:26 crc kubenswrapper[4809]: I1206 05:59:26.938325 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-7dc5585756-dv7bd" podStartSLOduration=2.938291426 podStartE2EDuration="2.938291426s" podCreationTimestamp="2025-12-06 05:59:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:59:26.937243418 +0000 UTC m=+491.826226360" watchObservedRunningTime="2025-12-06 05:59:26.938291426 +0000 UTC m=+491.827274368" Dec 06 05:59:26 crc kubenswrapper[4809]: I1206 05:59:26.940106 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-78wr9" podStartSLOduration=6.175502984 podStartE2EDuration="8.940101725s" podCreationTimestamp="2025-12-06 05:59:18 +0000 UTC" firstStartedPulling="2025-12-06 05:59:20.094229197 +0000 UTC m=+484.983212139" lastFinishedPulling="2025-12-06 05:59:22.858827938 +0000 UTC m=+487.747810880" observedRunningTime="2025-12-06 05:59:26.918898768 +0000 UTC m=+491.807881710" watchObservedRunningTime="2025-12-06 05:59:26.940101725 +0000 UTC m=+491.829084667" Dec 06 05:59:27 crc kubenswrapper[4809]: I1206 05:59:27.913706 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-cw8sk" event={"ID":"528ba802-293a-488e-8856-14b997f7956f","Type":"ContainerStarted","Data":"3179f3551ae60bd33e1eb94ebdc4bdb1f40b0e747e5bdb17a0231b54416fe34c"} Dec 06 05:59:27 crc kubenswrapper[4809]: I1206 05:59:27.914083 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-cw8sk" event={"ID":"528ba802-293a-488e-8856-14b997f7956f","Type":"ContainerStarted","Data":"46fafa554abca086cf383c9b43f029a0366ab41233cd7c9428b9c4d6bc5862d1"} Dec 06 05:59:30 crc kubenswrapper[4809]: I1206 05:59:30.931791 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-d74dd8556-59sjd" event={"ID":"ec930236-fceb-44dd-90c3-13a25d7cd9f2","Type":"ContainerStarted","Data":"e824675c0f4b0d57078dbb9f522a8cdb67c709aec814450ea7d3f04bc8a486b6"} Dec 06 05:59:30 crc kubenswrapper[4809]: I1206 05:59:30.932680 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-d74dd8556-59sjd" event={"ID":"ec930236-fceb-44dd-90c3-13a25d7cd9f2","Type":"ContainerStarted","Data":"fd71a55a72d12d27570ce5aacad65dc0ce381eed2ecb5af6bf963fe76a5a9f10"} Dec 06 05:59:30 crc kubenswrapper[4809]: I1206 05:59:30.932702 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-d74dd8556-59sjd" event={"ID":"ec930236-fceb-44dd-90c3-13a25d7cd9f2","Type":"ContainerStarted","Data":"8d7c2abd5e6c0fa6092cfa7ab341044172b6e2751d81ca44b84672cc7da3e69c"} Dec 06 05:59:30 crc kubenswrapper[4809]: I1206 05:59:30.934422 4809 generic.go:334] "Generic (PLEG): container finished" podID="869feda2-ec9d-44de-9533-27a4d7b990c1" containerID="1d79c10ab11219e9c073657705a0e46216a085bf44d95351a6757115ef4fa5c0" exitCode=0 Dec 06 05:59:30 crc kubenswrapper[4809]: I1206 05:59:30.934502 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"869feda2-ec9d-44de-9533-27a4d7b990c1","Type":"ContainerDied","Data":"1d79c10ab11219e9c073657705a0e46216a085bf44d95351a6757115ef4fa5c0"} Dec 06 05:59:30 crc kubenswrapper[4809]: I1206 05:59:30.936247 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/monitoring-plugin-5998c84788-zjqnq" event={"ID":"bb66c253-f1a0-4412-93f7-82b8fb03fb33","Type":"ContainerStarted","Data":"69e7afaeb6c5f41dedd054fc4eb7cdab6e7176b65a4f5d7146524c6e7f5f7704"} Dec 06 05:59:30 crc kubenswrapper[4809]: I1206 05:59:30.936669 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/monitoring-plugin-5998c84788-zjqnq" Dec 06 05:59:30 crc kubenswrapper[4809]: I1206 05:59:30.938765 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-xmx2t" event={"ID":"e1147859-11de-422f-9fb6-84ad24e7f49c","Type":"ContainerStarted","Data":"aa8f0471dbd304cdf6697fa34e200e07c7b9ab5baeb9ca273442267feb145ca1"} Dec 06 05:59:30 crc kubenswrapper[4809]: I1206 05:59:30.940635 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-556bbc5579-8wblw" event={"ID":"b4f96141-d70e-4715-9e30-516c9e783cc9","Type":"ContainerStarted","Data":"cac60d4c52fd5a110675ebf867d8dabeb845a5056ba0753e52a1a1f592f0790d"} Dec 06 05:59:30 crc kubenswrapper[4809]: I1206 05:59:30.943145 4809 generic.go:334] "Generic (PLEG): container finished" podID="cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b" containerID="808af2c17dfd3e7c8786faae8f15d9d2e7c995f701cc9f0c2178e2c95ac78904" exitCode=0 Dec 06 05:59:30 crc kubenswrapper[4809]: I1206 05:59:30.943184 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b","Type":"ContainerDied","Data":"808af2c17dfd3e7c8786faae8f15d9d2e7c995f701cc9f0c2178e2c95ac78904"} Dec 06 05:59:30 crc kubenswrapper[4809]: I1206 05:59:30.943817 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/monitoring-plugin-5998c84788-zjqnq" Dec 06 05:59:30 crc kubenswrapper[4809]: I1206 05:59:30.964099 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/node-exporter-cw8sk" podStartSLOduration=9.523106014 podStartE2EDuration="12.964079471s" podCreationTimestamp="2025-12-06 05:59:18 +0000 UTC" firstStartedPulling="2025-12-06 05:59:19.412054153 +0000 UTC m=+484.301037095" lastFinishedPulling="2025-12-06 05:59:22.85302761 +0000 UTC m=+487.742010552" observedRunningTime="2025-12-06 05:59:27.940915853 +0000 UTC m=+492.829898795" watchObservedRunningTime="2025-12-06 05:59:30.964079471 +0000 UTC m=+495.853062413" Dec 06 05:59:30 crc kubenswrapper[4809]: I1206 05:59:30.977565 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/monitoring-plugin-5998c84788-zjqnq" podStartSLOduration=2.401002591 podStartE2EDuration="6.977543288s" podCreationTimestamp="2025-12-06 05:59:24 +0000 UTC" firstStartedPulling="2025-12-06 05:59:25.301111283 +0000 UTC m=+490.190094235" lastFinishedPulling="2025-12-06 05:59:29.87765199 +0000 UTC m=+494.766634932" observedRunningTime="2025-12-06 05:59:30.975123892 +0000 UTC m=+495.864106854" watchObservedRunningTime="2025-12-06 05:59:30.977543288 +0000 UTC m=+495.866526230" Dec 06 05:59:31 crc kubenswrapper[4809]: I1206 05:59:31.027371 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/openshift-state-metrics-566fddb674-xmx2t" podStartSLOduration=5.324299363 podStartE2EDuration="13.027352355s" podCreationTimestamp="2025-12-06 05:59:18 +0000 UTC" firstStartedPulling="2025-12-06 05:59:22.149988768 +0000 UTC m=+487.038971710" lastFinishedPulling="2025-12-06 05:59:29.85304176 +0000 UTC m=+494.742024702" observedRunningTime="2025-12-06 05:59:31.023331706 +0000 UTC m=+495.912314648" watchObservedRunningTime="2025-12-06 05:59:31.027352355 +0000 UTC m=+495.916335307" Dec 06 05:59:31 crc kubenswrapper[4809]: I1206 05:59:31.041576 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/metrics-server-556bbc5579-8wblw" podStartSLOduration=1.8947288119999999 podStartE2EDuration="7.041555822s" podCreationTimestamp="2025-12-06 05:59:24 +0000 UTC" firstStartedPulling="2025-12-06 05:59:24.738916511 +0000 UTC m=+489.627899453" lastFinishedPulling="2025-12-06 05:59:29.885743521 +0000 UTC m=+494.774726463" observedRunningTime="2025-12-06 05:59:31.038089098 +0000 UTC m=+495.927072050" watchObservedRunningTime="2025-12-06 05:59:31.041555822 +0000 UTC m=+495.930538774" Dec 06 05:59:34 crc kubenswrapper[4809]: I1206 05:59:34.655817 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-7dc5585756-dv7bd" Dec 06 05:59:34 crc kubenswrapper[4809]: I1206 05:59:34.657601 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-7dc5585756-dv7bd" Dec 06 05:59:34 crc kubenswrapper[4809]: I1206 05:59:34.662920 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-7dc5585756-dv7bd" Dec 06 05:59:34 crc kubenswrapper[4809]: I1206 05:59:34.972361 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-7dc5585756-dv7bd" Dec 06 05:59:35 crc kubenswrapper[4809]: I1206 05:59:35.047013 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-4pwcw"] Dec 06 05:59:35 crc kubenswrapper[4809]: I1206 05:59:35.978682 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-d74dd8556-59sjd" event={"ID":"ec930236-fceb-44dd-90c3-13a25d7cd9f2","Type":"ContainerStarted","Data":"f056187d378c05601a74928ba2d2f48a8e963720ef254dad689e493f44c94f1f"} Dec 06 05:59:35 crc kubenswrapper[4809]: I1206 05:59:35.979032 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-d74dd8556-59sjd" event={"ID":"ec930236-fceb-44dd-90c3-13a25d7cd9f2","Type":"ContainerStarted","Data":"4808118b46ad85c46841c5f5629d3ecd63d9de62cc3046c3c171c96d62116eca"} Dec 06 05:59:37 crc kubenswrapper[4809]: I1206 05:59:37.991389 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b","Type":"ContainerStarted","Data":"8ea504e587b22bfd1ef5ff29bdf21943112139f36b73ce0718084642ea0dc537"} Dec 06 05:59:37 crc kubenswrapper[4809]: I1206 05:59:37.992146 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b","Type":"ContainerStarted","Data":"606eef1b86f2e959fd75d07cab6b2eb791d18171ff9cb61c9e6c8a75e5be8c82"} Dec 06 05:59:37 crc kubenswrapper[4809]: I1206 05:59:37.992175 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b","Type":"ContainerStarted","Data":"4252fed893c347da600e071a9e4100c7b7c419fa16e8fb574209be63067a7dcc"} Dec 06 05:59:37 crc kubenswrapper[4809]: I1206 05:59:37.992189 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b","Type":"ContainerStarted","Data":"71d3c3ce68ec6a823a5babc02d4dd3b48988495e129f7ceaea00bb2123dca317"} Dec 06 05:59:37 crc kubenswrapper[4809]: I1206 05:59:37.992201 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b","Type":"ContainerStarted","Data":"2e1e9eafccd2b3c01bf8bdb7c8cd13f947d590adb7c8936039ca9d5385e5a8e5"} Dec 06 05:59:37 crc kubenswrapper[4809]: I1206 05:59:37.992213 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"cff2bf4d-ccb5-46ed-a2fc-38ddbb4a9c4b","Type":"ContainerStarted","Data":"f65c9f63f0c05581c918b1728e7c6acd577fbe1991892a91544825b1a259a27a"} Dec 06 05:59:37 crc kubenswrapper[4809]: I1206 05:59:37.998118 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-d74dd8556-59sjd" event={"ID":"ec930236-fceb-44dd-90c3-13a25d7cd9f2","Type":"ContainerStarted","Data":"d06a81da14d8cb1c8193c0a7a88b1f53c755b492069428fc6da30e9217961544"} Dec 06 05:59:37 crc kubenswrapper[4809]: I1206 05:59:37.999406 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/thanos-querier-d74dd8556-59sjd" Dec 06 05:59:38 crc kubenswrapper[4809]: I1206 05:59:38.007508 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"869feda2-ec9d-44de-9533-27a4d7b990c1","Type":"ContainerStarted","Data":"c3ffdd935fa61106de3e708776db4ec78f552dfae0c45cd2c98c16104a3642c1"} Dec 06 05:59:38 crc kubenswrapper[4809]: I1206 05:59:38.007545 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"869feda2-ec9d-44de-9533-27a4d7b990c1","Type":"ContainerStarted","Data":"6307328b4258ef2a0ae80f8283453f58c14829a0ec9be18e43a501b2b9fd7040"} Dec 06 05:59:38 crc kubenswrapper[4809]: I1206 05:59:38.007558 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"869feda2-ec9d-44de-9533-27a4d7b990c1","Type":"ContainerStarted","Data":"284259b8e5d231ad33c982670e31c61566177af1f29399d38cda9ab9470d6855"} Dec 06 05:59:38 crc kubenswrapper[4809]: I1206 05:59:38.007570 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"869feda2-ec9d-44de-9533-27a4d7b990c1","Type":"ContainerStarted","Data":"2bac19ae9a28e4d592ca2ae57c919cb2c94b75eaca168d289f136a4b7be9a1b6"} Dec 06 05:59:38 crc kubenswrapper[4809]: I1206 05:59:38.007581 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"869feda2-ec9d-44de-9533-27a4d7b990c1","Type":"ContainerStarted","Data":"a13ae564d1b094cbb67a76a515806e146cce795ebeca86cf898672ed18f873b8"} Dec 06 05:59:38 crc kubenswrapper[4809]: I1206 05:59:38.007593 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"869feda2-ec9d-44de-9533-27a4d7b990c1","Type":"ContainerStarted","Data":"3fafe0a511e3074d660f5ccb4036652361effcb0f4ec496f8c7f708bdabda5f6"} Dec 06 05:59:38 crc kubenswrapper[4809]: I1206 05:59:38.008561 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/thanos-querier-d74dd8556-59sjd" Dec 06 05:59:38 crc kubenswrapper[4809]: I1206 05:59:38.027899 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/alertmanager-main-0" podStartSLOduration=5.070285406 podStartE2EDuration="19.027878978s" podCreationTimestamp="2025-12-06 05:59:19 +0000 UTC" firstStartedPulling="2025-12-06 05:59:23.112261335 +0000 UTC m=+488.001244277" lastFinishedPulling="2025-12-06 05:59:37.069854907 +0000 UTC m=+501.958837849" observedRunningTime="2025-12-06 05:59:38.022181803 +0000 UTC m=+502.911164755" watchObservedRunningTime="2025-12-06 05:59:38.027878978 +0000 UTC m=+502.916861940" Dec 06 05:59:38 crc kubenswrapper[4809]: I1206 05:59:38.045619 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/thanos-querier-d74dd8556-59sjd" podStartSLOduration=9.418265648 podStartE2EDuration="18.045600331s" podCreationTimestamp="2025-12-06 05:59:20 +0000 UTC" firstStartedPulling="2025-12-06 05:59:23.109551811 +0000 UTC m=+487.998534753" lastFinishedPulling="2025-12-06 05:59:31.736886484 +0000 UTC m=+496.625869436" observedRunningTime="2025-12-06 05:59:38.042638251 +0000 UTC m=+502.931621193" watchObservedRunningTime="2025-12-06 05:59:38.045600331 +0000 UTC m=+502.934583273" Dec 06 05:59:38 crc kubenswrapper[4809]: I1206 05:59:38.082147 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-k8s-0" podStartSLOduration=2.824223898 podStartE2EDuration="14.082124237s" podCreationTimestamp="2025-12-06 05:59:24 +0000 UTC" firstStartedPulling="2025-12-06 05:59:25.797069262 +0000 UTC m=+490.686052214" lastFinishedPulling="2025-12-06 05:59:37.054969611 +0000 UTC m=+501.943952553" observedRunningTime="2025-12-06 05:59:38.076497254 +0000 UTC m=+502.965480216" watchObservedRunningTime="2025-12-06 05:59:38.082124237 +0000 UTC m=+502.971107179" Dec 06 05:59:40 crc kubenswrapper[4809]: I1206 05:59:40.320709 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-k8s-0" Dec 06 05:59:44 crc kubenswrapper[4809]: I1206 05:59:44.421995 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/metrics-server-556bbc5579-8wblw" Dec 06 05:59:44 crc kubenswrapper[4809]: I1206 05:59:44.422415 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/metrics-server-556bbc5579-8wblw" Dec 06 06:00:00 crc kubenswrapper[4809]: I1206 06:00:00.094650 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-4pwcw" podUID="8e03364a-1840-4c91-8b2b-c7e2071fc73b" containerName="console" containerID="cri-o://4007a4fcd8930f1f9d5f287878aac2fe01b09a892931d190ed6ac6d779049f92" gracePeriod=15 Dec 06 06:00:00 crc kubenswrapper[4809]: I1206 06:00:00.203372 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416680-97j2f"] Dec 06 06:00:00 crc kubenswrapper[4809]: I1206 06:00:00.205839 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416680-97j2f" Dec 06 06:00:00 crc kubenswrapper[4809]: I1206 06:00:00.209851 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 06:00:00 crc kubenswrapper[4809]: I1206 06:00:00.211896 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 06:00:00 crc kubenswrapper[4809]: I1206 06:00:00.213088 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416680-97j2f"] Dec 06 06:00:00 crc kubenswrapper[4809]: I1206 06:00:00.262448 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c52c53b2-f8ba-4272-863b-f8ba48c1b3cc-secret-volume\") pod \"collect-profiles-29416680-97j2f\" (UID: \"c52c53b2-f8ba-4272-863b-f8ba48c1b3cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416680-97j2f" Dec 06 06:00:00 crc kubenswrapper[4809]: I1206 06:00:00.262859 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c52c53b2-f8ba-4272-863b-f8ba48c1b3cc-config-volume\") pod \"collect-profiles-29416680-97j2f\" (UID: \"c52c53b2-f8ba-4272-863b-f8ba48c1b3cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416680-97j2f" Dec 06 06:00:00 crc kubenswrapper[4809]: I1206 06:00:00.262906 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9624\" (UniqueName: \"kubernetes.io/projected/c52c53b2-f8ba-4272-863b-f8ba48c1b3cc-kube-api-access-d9624\") pod \"collect-profiles-29416680-97j2f\" (UID: \"c52c53b2-f8ba-4272-863b-f8ba48c1b3cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416680-97j2f" Dec 06 06:00:00 crc kubenswrapper[4809]: I1206 06:00:00.364190 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c52c53b2-f8ba-4272-863b-f8ba48c1b3cc-secret-volume\") pod \"collect-profiles-29416680-97j2f\" (UID: \"c52c53b2-f8ba-4272-863b-f8ba48c1b3cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416680-97j2f" Dec 06 06:00:00 crc kubenswrapper[4809]: I1206 06:00:00.364254 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c52c53b2-f8ba-4272-863b-f8ba48c1b3cc-config-volume\") pod \"collect-profiles-29416680-97j2f\" (UID: \"c52c53b2-f8ba-4272-863b-f8ba48c1b3cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416680-97j2f" Dec 06 06:00:00 crc kubenswrapper[4809]: I1206 06:00:00.365513 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c52c53b2-f8ba-4272-863b-f8ba48c1b3cc-config-volume\") pod \"collect-profiles-29416680-97j2f\" (UID: \"c52c53b2-f8ba-4272-863b-f8ba48c1b3cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416680-97j2f" Dec 06 06:00:00 crc kubenswrapper[4809]: I1206 06:00:00.365596 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9624\" (UniqueName: \"kubernetes.io/projected/c52c53b2-f8ba-4272-863b-f8ba48c1b3cc-kube-api-access-d9624\") pod \"collect-profiles-29416680-97j2f\" (UID: \"c52c53b2-f8ba-4272-863b-f8ba48c1b3cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416680-97j2f" Dec 06 06:00:00 crc kubenswrapper[4809]: I1206 06:00:00.370200 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c52c53b2-f8ba-4272-863b-f8ba48c1b3cc-secret-volume\") pod \"collect-profiles-29416680-97j2f\" (UID: \"c52c53b2-f8ba-4272-863b-f8ba48c1b3cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416680-97j2f" Dec 06 06:00:00 crc kubenswrapper[4809]: I1206 06:00:00.385877 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9624\" (UniqueName: \"kubernetes.io/projected/c52c53b2-f8ba-4272-863b-f8ba48c1b3cc-kube-api-access-d9624\") pod \"collect-profiles-29416680-97j2f\" (UID: \"c52c53b2-f8ba-4272-863b-f8ba48c1b3cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416680-97j2f" Dec 06 06:00:00 crc kubenswrapper[4809]: I1206 06:00:00.460178 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-4pwcw_8e03364a-1840-4c91-8b2b-c7e2071fc73b/console/0.log" Dec 06 06:00:00 crc kubenswrapper[4809]: I1206 06:00:00.460262 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-4pwcw" Dec 06 06:00:00 crc kubenswrapper[4809]: I1206 06:00:00.549714 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416680-97j2f" Dec 06 06:00:00 crc kubenswrapper[4809]: I1206 06:00:00.572264 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-glkvh\" (UniqueName: \"kubernetes.io/projected/8e03364a-1840-4c91-8b2b-c7e2071fc73b-kube-api-access-glkvh\") pod \"8e03364a-1840-4c91-8b2b-c7e2071fc73b\" (UID: \"8e03364a-1840-4c91-8b2b-c7e2071fc73b\") " Dec 06 06:00:00 crc kubenswrapper[4809]: I1206 06:00:00.572358 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8e03364a-1840-4c91-8b2b-c7e2071fc73b-console-serving-cert\") pod \"8e03364a-1840-4c91-8b2b-c7e2071fc73b\" (UID: \"8e03364a-1840-4c91-8b2b-c7e2071fc73b\") " Dec 06 06:00:00 crc kubenswrapper[4809]: I1206 06:00:00.572414 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8e03364a-1840-4c91-8b2b-c7e2071fc73b-oauth-serving-cert\") pod \"8e03364a-1840-4c91-8b2b-c7e2071fc73b\" (UID: \"8e03364a-1840-4c91-8b2b-c7e2071fc73b\") " Dec 06 06:00:00 crc kubenswrapper[4809]: I1206 06:00:00.572457 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8e03364a-1840-4c91-8b2b-c7e2071fc73b-console-config\") pod \"8e03364a-1840-4c91-8b2b-c7e2071fc73b\" (UID: \"8e03364a-1840-4c91-8b2b-c7e2071fc73b\") " Dec 06 06:00:00 crc kubenswrapper[4809]: I1206 06:00:00.572506 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e03364a-1840-4c91-8b2b-c7e2071fc73b-trusted-ca-bundle\") pod \"8e03364a-1840-4c91-8b2b-c7e2071fc73b\" (UID: \"8e03364a-1840-4c91-8b2b-c7e2071fc73b\") " Dec 06 06:00:00 crc kubenswrapper[4809]: I1206 06:00:00.572533 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8e03364a-1840-4c91-8b2b-c7e2071fc73b-service-ca\") pod \"8e03364a-1840-4c91-8b2b-c7e2071fc73b\" (UID: \"8e03364a-1840-4c91-8b2b-c7e2071fc73b\") " Dec 06 06:00:00 crc kubenswrapper[4809]: I1206 06:00:00.572565 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8e03364a-1840-4c91-8b2b-c7e2071fc73b-console-oauth-config\") pod \"8e03364a-1840-4c91-8b2b-c7e2071fc73b\" (UID: \"8e03364a-1840-4c91-8b2b-c7e2071fc73b\") " Dec 06 06:00:00 crc kubenswrapper[4809]: I1206 06:00:00.573495 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e03364a-1840-4c91-8b2b-c7e2071fc73b-console-config" (OuterVolumeSpecName: "console-config") pod "8e03364a-1840-4c91-8b2b-c7e2071fc73b" (UID: "8e03364a-1840-4c91-8b2b-c7e2071fc73b"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:00:00 crc kubenswrapper[4809]: I1206 06:00:00.573532 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e03364a-1840-4c91-8b2b-c7e2071fc73b-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "8e03364a-1840-4c91-8b2b-c7e2071fc73b" (UID: "8e03364a-1840-4c91-8b2b-c7e2071fc73b"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:00:00 crc kubenswrapper[4809]: I1206 06:00:00.573737 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e03364a-1840-4c91-8b2b-c7e2071fc73b-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "8e03364a-1840-4c91-8b2b-c7e2071fc73b" (UID: "8e03364a-1840-4c91-8b2b-c7e2071fc73b"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:00:00 crc kubenswrapper[4809]: I1206 06:00:00.573902 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e03364a-1840-4c91-8b2b-c7e2071fc73b-service-ca" (OuterVolumeSpecName: "service-ca") pod "8e03364a-1840-4c91-8b2b-c7e2071fc73b" (UID: "8e03364a-1840-4c91-8b2b-c7e2071fc73b"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:00:00 crc kubenswrapper[4809]: I1206 06:00:00.576945 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e03364a-1840-4c91-8b2b-c7e2071fc73b-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "8e03364a-1840-4c91-8b2b-c7e2071fc73b" (UID: "8e03364a-1840-4c91-8b2b-c7e2071fc73b"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:00:00 crc kubenswrapper[4809]: I1206 06:00:00.577015 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e03364a-1840-4c91-8b2b-c7e2071fc73b-kube-api-access-glkvh" (OuterVolumeSpecName: "kube-api-access-glkvh") pod "8e03364a-1840-4c91-8b2b-c7e2071fc73b" (UID: "8e03364a-1840-4c91-8b2b-c7e2071fc73b"). InnerVolumeSpecName "kube-api-access-glkvh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:00:00 crc kubenswrapper[4809]: I1206 06:00:00.577531 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e03364a-1840-4c91-8b2b-c7e2071fc73b-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "8e03364a-1840-4c91-8b2b-c7e2071fc73b" (UID: "8e03364a-1840-4c91-8b2b-c7e2071fc73b"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:00:00 crc kubenswrapper[4809]: I1206 06:00:00.673645 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-glkvh\" (UniqueName: \"kubernetes.io/projected/8e03364a-1840-4c91-8b2b-c7e2071fc73b-kube-api-access-glkvh\") on node \"crc\" DevicePath \"\"" Dec 06 06:00:00 crc kubenswrapper[4809]: I1206 06:00:00.673898 4809 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8e03364a-1840-4c91-8b2b-c7e2071fc73b-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 06:00:00 crc kubenswrapper[4809]: I1206 06:00:00.673908 4809 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8e03364a-1840-4c91-8b2b-c7e2071fc73b-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 06:00:00 crc kubenswrapper[4809]: I1206 06:00:00.673916 4809 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8e03364a-1840-4c91-8b2b-c7e2071fc73b-console-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:00:00 crc kubenswrapper[4809]: I1206 06:00:00.673924 4809 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e03364a-1840-4c91-8b2b-c7e2071fc73b-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:00:00 crc kubenswrapper[4809]: I1206 06:00:00.673947 4809 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8e03364a-1840-4c91-8b2b-c7e2071fc73b-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 06:00:00 crc kubenswrapper[4809]: I1206 06:00:00.673955 4809 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8e03364a-1840-4c91-8b2b-c7e2071fc73b-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:00:00 crc kubenswrapper[4809]: I1206 06:00:00.925300 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416680-97j2f"] Dec 06 06:00:01 crc kubenswrapper[4809]: I1206 06:00:01.154427 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-4pwcw_8e03364a-1840-4c91-8b2b-c7e2071fc73b/console/0.log" Dec 06 06:00:01 crc kubenswrapper[4809]: I1206 06:00:01.154487 4809 generic.go:334] "Generic (PLEG): container finished" podID="8e03364a-1840-4c91-8b2b-c7e2071fc73b" containerID="4007a4fcd8930f1f9d5f287878aac2fe01b09a892931d190ed6ac6d779049f92" exitCode=2 Dec 06 06:00:01 crc kubenswrapper[4809]: I1206 06:00:01.154530 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-4pwcw" event={"ID":"8e03364a-1840-4c91-8b2b-c7e2071fc73b","Type":"ContainerDied","Data":"4007a4fcd8930f1f9d5f287878aac2fe01b09a892931d190ed6ac6d779049f92"} Dec 06 06:00:01 crc kubenswrapper[4809]: I1206 06:00:01.154566 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-4pwcw" Dec 06 06:00:01 crc kubenswrapper[4809]: I1206 06:00:01.154584 4809 scope.go:117] "RemoveContainer" containerID="4007a4fcd8930f1f9d5f287878aac2fe01b09a892931d190ed6ac6d779049f92" Dec 06 06:00:01 crc kubenswrapper[4809]: I1206 06:00:01.154571 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-4pwcw" event={"ID":"8e03364a-1840-4c91-8b2b-c7e2071fc73b","Type":"ContainerDied","Data":"16846d1b1f5b3efa865f1863bb744eaa55819731e81fe580e571a1e290114936"} Dec 06 06:00:01 crc kubenswrapper[4809]: I1206 06:00:01.157926 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416680-97j2f" event={"ID":"c52c53b2-f8ba-4272-863b-f8ba48c1b3cc","Type":"ContainerStarted","Data":"34e04c3ffeae84899ea75dfbf59cbd3770a5c9f7104321b0bf1f6a4bad7da62f"} Dec 06 06:00:01 crc kubenswrapper[4809]: I1206 06:00:01.157976 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416680-97j2f" event={"ID":"c52c53b2-f8ba-4272-863b-f8ba48c1b3cc","Type":"ContainerStarted","Data":"eef9b0a52c252eeaac902d9c4e4fe086819c052dab223ed001e781a1fab9fbfa"} Dec 06 06:00:01 crc kubenswrapper[4809]: I1206 06:00:01.176081 4809 scope.go:117] "RemoveContainer" containerID="4007a4fcd8930f1f9d5f287878aac2fe01b09a892931d190ed6ac6d779049f92" Dec 06 06:00:01 crc kubenswrapper[4809]: E1206 06:00:01.176533 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4007a4fcd8930f1f9d5f287878aac2fe01b09a892931d190ed6ac6d779049f92\": container with ID starting with 4007a4fcd8930f1f9d5f287878aac2fe01b09a892931d190ed6ac6d779049f92 not found: ID does not exist" containerID="4007a4fcd8930f1f9d5f287878aac2fe01b09a892931d190ed6ac6d779049f92" Dec 06 06:00:01 crc kubenswrapper[4809]: I1206 06:00:01.176566 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4007a4fcd8930f1f9d5f287878aac2fe01b09a892931d190ed6ac6d779049f92"} err="failed to get container status \"4007a4fcd8930f1f9d5f287878aac2fe01b09a892931d190ed6ac6d779049f92\": rpc error: code = NotFound desc = could not find container \"4007a4fcd8930f1f9d5f287878aac2fe01b09a892931d190ed6ac6d779049f92\": container with ID starting with 4007a4fcd8930f1f9d5f287878aac2fe01b09a892931d190ed6ac6d779049f92 not found: ID does not exist" Dec 06 06:00:01 crc kubenswrapper[4809]: I1206 06:00:01.185037 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29416680-97j2f" podStartSLOduration=1.185015999 podStartE2EDuration="1.185015999s" podCreationTimestamp="2025-12-06 06:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:00:01.174082991 +0000 UTC m=+526.063065933" watchObservedRunningTime="2025-12-06 06:00:01.185015999 +0000 UTC m=+526.073998941" Dec 06 06:00:01 crc kubenswrapper[4809]: I1206 06:00:01.190261 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-4pwcw"] Dec 06 06:00:01 crc kubenswrapper[4809]: I1206 06:00:01.197609 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-4pwcw"] Dec 06 06:00:01 crc kubenswrapper[4809]: I1206 06:00:01.397862 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e03364a-1840-4c91-8b2b-c7e2071fc73b" path="/var/lib/kubelet/pods/8e03364a-1840-4c91-8b2b-c7e2071fc73b/volumes" Dec 06 06:00:02 crc kubenswrapper[4809]: I1206 06:00:02.167244 4809 generic.go:334] "Generic (PLEG): container finished" podID="c52c53b2-f8ba-4272-863b-f8ba48c1b3cc" containerID="34e04c3ffeae84899ea75dfbf59cbd3770a5c9f7104321b0bf1f6a4bad7da62f" exitCode=0 Dec 06 06:00:02 crc kubenswrapper[4809]: I1206 06:00:02.168129 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416680-97j2f" event={"ID":"c52c53b2-f8ba-4272-863b-f8ba48c1b3cc","Type":"ContainerDied","Data":"34e04c3ffeae84899ea75dfbf59cbd3770a5c9f7104321b0bf1f6a4bad7da62f"} Dec 06 06:00:03 crc kubenswrapper[4809]: I1206 06:00:03.429289 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416680-97j2f" Dec 06 06:00:03 crc kubenswrapper[4809]: I1206 06:00:03.514611 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c52c53b2-f8ba-4272-863b-f8ba48c1b3cc-secret-volume\") pod \"c52c53b2-f8ba-4272-863b-f8ba48c1b3cc\" (UID: \"c52c53b2-f8ba-4272-863b-f8ba48c1b3cc\") " Dec 06 06:00:03 crc kubenswrapper[4809]: I1206 06:00:03.514705 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9624\" (UniqueName: \"kubernetes.io/projected/c52c53b2-f8ba-4272-863b-f8ba48c1b3cc-kube-api-access-d9624\") pod \"c52c53b2-f8ba-4272-863b-f8ba48c1b3cc\" (UID: \"c52c53b2-f8ba-4272-863b-f8ba48c1b3cc\") " Dec 06 06:00:03 crc kubenswrapper[4809]: I1206 06:00:03.514843 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c52c53b2-f8ba-4272-863b-f8ba48c1b3cc-config-volume\") pod \"c52c53b2-f8ba-4272-863b-f8ba48c1b3cc\" (UID: \"c52c53b2-f8ba-4272-863b-f8ba48c1b3cc\") " Dec 06 06:00:03 crc kubenswrapper[4809]: I1206 06:00:03.515737 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c52c53b2-f8ba-4272-863b-f8ba48c1b3cc-config-volume" (OuterVolumeSpecName: "config-volume") pod "c52c53b2-f8ba-4272-863b-f8ba48c1b3cc" (UID: "c52c53b2-f8ba-4272-863b-f8ba48c1b3cc"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:00:03 crc kubenswrapper[4809]: I1206 06:00:03.516357 4809 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c52c53b2-f8ba-4272-863b-f8ba48c1b3cc-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 06:00:03 crc kubenswrapper[4809]: I1206 06:00:03.519887 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c52c53b2-f8ba-4272-863b-f8ba48c1b3cc-kube-api-access-d9624" (OuterVolumeSpecName: "kube-api-access-d9624") pod "c52c53b2-f8ba-4272-863b-f8ba48c1b3cc" (UID: "c52c53b2-f8ba-4272-863b-f8ba48c1b3cc"). InnerVolumeSpecName "kube-api-access-d9624". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:00:03 crc kubenswrapper[4809]: I1206 06:00:03.520382 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c52c53b2-f8ba-4272-863b-f8ba48c1b3cc-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c52c53b2-f8ba-4272-863b-f8ba48c1b3cc" (UID: "c52c53b2-f8ba-4272-863b-f8ba48c1b3cc"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:00:03 crc kubenswrapper[4809]: I1206 06:00:03.617079 4809 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c52c53b2-f8ba-4272-863b-f8ba48c1b3cc-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 06:00:03 crc kubenswrapper[4809]: I1206 06:00:03.617113 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9624\" (UniqueName: \"kubernetes.io/projected/c52c53b2-f8ba-4272-863b-f8ba48c1b3cc-kube-api-access-d9624\") on node \"crc\" DevicePath \"\"" Dec 06 06:00:04 crc kubenswrapper[4809]: I1206 06:00:04.178637 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416680-97j2f" event={"ID":"c52c53b2-f8ba-4272-863b-f8ba48c1b3cc","Type":"ContainerDied","Data":"eef9b0a52c252eeaac902d9c4e4fe086819c052dab223ed001e781a1fab9fbfa"} Dec 06 06:00:04 crc kubenswrapper[4809]: I1206 06:00:04.178679 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eef9b0a52c252eeaac902d9c4e4fe086819c052dab223ed001e781a1fab9fbfa" Dec 06 06:00:04 crc kubenswrapper[4809]: I1206 06:00:04.178728 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416680-97j2f" Dec 06 06:00:04 crc kubenswrapper[4809]: I1206 06:00:04.430095 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/metrics-server-556bbc5579-8wblw" Dec 06 06:00:04 crc kubenswrapper[4809]: I1206 06:00:04.438055 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/metrics-server-556bbc5579-8wblw" Dec 06 06:00:15 crc kubenswrapper[4809]: I1206 06:00:15.928045 4809 scope.go:117] "RemoveContainer" containerID="b4f1ca19e4f9be63b5be9830ba9302245bb78083537d87f172a89ffc807173d9" Dec 06 06:00:25 crc kubenswrapper[4809]: I1206 06:00:25.320540 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/prometheus-k8s-0" Dec 06 06:00:25 crc kubenswrapper[4809]: I1206 06:00:25.352996 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/prometheus-k8s-0" Dec 06 06:00:25 crc kubenswrapper[4809]: I1206 06:00:25.382299 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-k8s-0" Dec 06 06:01:04 crc kubenswrapper[4809]: I1206 06:01:04.497036 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:01:04 crc kubenswrapper[4809]: I1206 06:01:04.497694 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:01:10 crc kubenswrapper[4809]: I1206 06:01:10.125445 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-664586d6fb-zgzv8"] Dec 06 06:01:10 crc kubenswrapper[4809]: E1206 06:01:10.126223 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e03364a-1840-4c91-8b2b-c7e2071fc73b" containerName="console" Dec 06 06:01:10 crc kubenswrapper[4809]: I1206 06:01:10.126239 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e03364a-1840-4c91-8b2b-c7e2071fc73b" containerName="console" Dec 06 06:01:10 crc kubenswrapper[4809]: E1206 06:01:10.126264 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c52c53b2-f8ba-4272-863b-f8ba48c1b3cc" containerName="collect-profiles" Dec 06 06:01:10 crc kubenswrapper[4809]: I1206 06:01:10.126271 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c52c53b2-f8ba-4272-863b-f8ba48c1b3cc" containerName="collect-profiles" Dec 06 06:01:10 crc kubenswrapper[4809]: I1206 06:01:10.126367 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="c52c53b2-f8ba-4272-863b-f8ba48c1b3cc" containerName="collect-profiles" Dec 06 06:01:10 crc kubenswrapper[4809]: I1206 06:01:10.126389 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e03364a-1840-4c91-8b2b-c7e2071fc73b" containerName="console" Dec 06 06:01:10 crc kubenswrapper[4809]: I1206 06:01:10.126822 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-664586d6fb-zgzv8" Dec 06 06:01:10 crc kubenswrapper[4809]: I1206 06:01:10.144063 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-664586d6fb-zgzv8"] Dec 06 06:01:10 crc kubenswrapper[4809]: I1206 06:01:10.236209 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwd2s\" (UniqueName: \"kubernetes.io/projected/7b54a496-5ecf-48c6-a9ba-d30828286c79-kube-api-access-lwd2s\") pod \"console-664586d6fb-zgzv8\" (UID: \"7b54a496-5ecf-48c6-a9ba-d30828286c79\") " pod="openshift-console/console-664586d6fb-zgzv8" Dec 06 06:01:10 crc kubenswrapper[4809]: I1206 06:01:10.236286 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7b54a496-5ecf-48c6-a9ba-d30828286c79-console-oauth-config\") pod \"console-664586d6fb-zgzv8\" (UID: \"7b54a496-5ecf-48c6-a9ba-d30828286c79\") " pod="openshift-console/console-664586d6fb-zgzv8" Dec 06 06:01:10 crc kubenswrapper[4809]: I1206 06:01:10.236315 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7b54a496-5ecf-48c6-a9ba-d30828286c79-oauth-serving-cert\") pod \"console-664586d6fb-zgzv8\" (UID: \"7b54a496-5ecf-48c6-a9ba-d30828286c79\") " pod="openshift-console/console-664586d6fb-zgzv8" Dec 06 06:01:10 crc kubenswrapper[4809]: I1206 06:01:10.236338 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7b54a496-5ecf-48c6-a9ba-d30828286c79-console-config\") pod \"console-664586d6fb-zgzv8\" (UID: \"7b54a496-5ecf-48c6-a9ba-d30828286c79\") " pod="openshift-console/console-664586d6fb-zgzv8" Dec 06 06:01:10 crc kubenswrapper[4809]: I1206 06:01:10.236405 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7b54a496-5ecf-48c6-a9ba-d30828286c79-console-serving-cert\") pod \"console-664586d6fb-zgzv8\" (UID: \"7b54a496-5ecf-48c6-a9ba-d30828286c79\") " pod="openshift-console/console-664586d6fb-zgzv8" Dec 06 06:01:10 crc kubenswrapper[4809]: I1206 06:01:10.236440 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7b54a496-5ecf-48c6-a9ba-d30828286c79-service-ca\") pod \"console-664586d6fb-zgzv8\" (UID: \"7b54a496-5ecf-48c6-a9ba-d30828286c79\") " pod="openshift-console/console-664586d6fb-zgzv8" Dec 06 06:01:10 crc kubenswrapper[4809]: I1206 06:01:10.236498 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7b54a496-5ecf-48c6-a9ba-d30828286c79-trusted-ca-bundle\") pod \"console-664586d6fb-zgzv8\" (UID: \"7b54a496-5ecf-48c6-a9ba-d30828286c79\") " pod="openshift-console/console-664586d6fb-zgzv8" Dec 06 06:01:10 crc kubenswrapper[4809]: I1206 06:01:10.337508 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7b54a496-5ecf-48c6-a9ba-d30828286c79-trusted-ca-bundle\") pod \"console-664586d6fb-zgzv8\" (UID: \"7b54a496-5ecf-48c6-a9ba-d30828286c79\") " pod="openshift-console/console-664586d6fb-zgzv8" Dec 06 06:01:10 crc kubenswrapper[4809]: I1206 06:01:10.337599 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwd2s\" (UniqueName: \"kubernetes.io/projected/7b54a496-5ecf-48c6-a9ba-d30828286c79-kube-api-access-lwd2s\") pod \"console-664586d6fb-zgzv8\" (UID: \"7b54a496-5ecf-48c6-a9ba-d30828286c79\") " pod="openshift-console/console-664586d6fb-zgzv8" Dec 06 06:01:10 crc kubenswrapper[4809]: I1206 06:01:10.337626 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7b54a496-5ecf-48c6-a9ba-d30828286c79-oauth-serving-cert\") pod \"console-664586d6fb-zgzv8\" (UID: \"7b54a496-5ecf-48c6-a9ba-d30828286c79\") " pod="openshift-console/console-664586d6fb-zgzv8" Dec 06 06:01:10 crc kubenswrapper[4809]: I1206 06:01:10.337641 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7b54a496-5ecf-48c6-a9ba-d30828286c79-console-oauth-config\") pod \"console-664586d6fb-zgzv8\" (UID: \"7b54a496-5ecf-48c6-a9ba-d30828286c79\") " pod="openshift-console/console-664586d6fb-zgzv8" Dec 06 06:01:10 crc kubenswrapper[4809]: I1206 06:01:10.337659 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7b54a496-5ecf-48c6-a9ba-d30828286c79-console-config\") pod \"console-664586d6fb-zgzv8\" (UID: \"7b54a496-5ecf-48c6-a9ba-d30828286c79\") " pod="openshift-console/console-664586d6fb-zgzv8" Dec 06 06:01:10 crc kubenswrapper[4809]: I1206 06:01:10.337681 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7b54a496-5ecf-48c6-a9ba-d30828286c79-console-serving-cert\") pod \"console-664586d6fb-zgzv8\" (UID: \"7b54a496-5ecf-48c6-a9ba-d30828286c79\") " pod="openshift-console/console-664586d6fb-zgzv8" Dec 06 06:01:10 crc kubenswrapper[4809]: I1206 06:01:10.337695 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7b54a496-5ecf-48c6-a9ba-d30828286c79-service-ca\") pod \"console-664586d6fb-zgzv8\" (UID: \"7b54a496-5ecf-48c6-a9ba-d30828286c79\") " pod="openshift-console/console-664586d6fb-zgzv8" Dec 06 06:01:10 crc kubenswrapper[4809]: I1206 06:01:10.338558 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7b54a496-5ecf-48c6-a9ba-d30828286c79-service-ca\") pod \"console-664586d6fb-zgzv8\" (UID: \"7b54a496-5ecf-48c6-a9ba-d30828286c79\") " pod="openshift-console/console-664586d6fb-zgzv8" Dec 06 06:01:10 crc kubenswrapper[4809]: I1206 06:01:10.339287 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7b54a496-5ecf-48c6-a9ba-d30828286c79-trusted-ca-bundle\") pod \"console-664586d6fb-zgzv8\" (UID: \"7b54a496-5ecf-48c6-a9ba-d30828286c79\") " pod="openshift-console/console-664586d6fb-zgzv8" Dec 06 06:01:10 crc kubenswrapper[4809]: I1206 06:01:10.340145 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7b54a496-5ecf-48c6-a9ba-d30828286c79-oauth-serving-cert\") pod \"console-664586d6fb-zgzv8\" (UID: \"7b54a496-5ecf-48c6-a9ba-d30828286c79\") " pod="openshift-console/console-664586d6fb-zgzv8" Dec 06 06:01:10 crc kubenswrapper[4809]: I1206 06:01:10.341602 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7b54a496-5ecf-48c6-a9ba-d30828286c79-console-config\") pod \"console-664586d6fb-zgzv8\" (UID: \"7b54a496-5ecf-48c6-a9ba-d30828286c79\") " pod="openshift-console/console-664586d6fb-zgzv8" Dec 06 06:01:10 crc kubenswrapper[4809]: I1206 06:01:10.351874 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7b54a496-5ecf-48c6-a9ba-d30828286c79-console-serving-cert\") pod \"console-664586d6fb-zgzv8\" (UID: \"7b54a496-5ecf-48c6-a9ba-d30828286c79\") " pod="openshift-console/console-664586d6fb-zgzv8" Dec 06 06:01:10 crc kubenswrapper[4809]: I1206 06:01:10.355201 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7b54a496-5ecf-48c6-a9ba-d30828286c79-console-oauth-config\") pod \"console-664586d6fb-zgzv8\" (UID: \"7b54a496-5ecf-48c6-a9ba-d30828286c79\") " pod="openshift-console/console-664586d6fb-zgzv8" Dec 06 06:01:10 crc kubenswrapper[4809]: I1206 06:01:10.357782 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwd2s\" (UniqueName: \"kubernetes.io/projected/7b54a496-5ecf-48c6-a9ba-d30828286c79-kube-api-access-lwd2s\") pod \"console-664586d6fb-zgzv8\" (UID: \"7b54a496-5ecf-48c6-a9ba-d30828286c79\") " pod="openshift-console/console-664586d6fb-zgzv8" Dec 06 06:01:10 crc kubenswrapper[4809]: I1206 06:01:10.446319 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-664586d6fb-zgzv8" Dec 06 06:01:10 crc kubenswrapper[4809]: I1206 06:01:10.861152 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-664586d6fb-zgzv8"] Dec 06 06:01:11 crc kubenswrapper[4809]: I1206 06:01:11.616038 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-664586d6fb-zgzv8" event={"ID":"7b54a496-5ecf-48c6-a9ba-d30828286c79","Type":"ContainerStarted","Data":"b89bc9800243faf10718de24d8ac3f3d8d6489019f431edef21c84ce711eb4d4"} Dec 06 06:01:11 crc kubenswrapper[4809]: I1206 06:01:11.616081 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-664586d6fb-zgzv8" event={"ID":"7b54a496-5ecf-48c6-a9ba-d30828286c79","Type":"ContainerStarted","Data":"4ef66f19a1e70d8f9939127aba4edacd0985245ec312ed68c403b0bfd23a9866"} Dec 06 06:01:11 crc kubenswrapper[4809]: I1206 06:01:11.643114 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-664586d6fb-zgzv8" podStartSLOduration=1.643096398 podStartE2EDuration="1.643096398s" podCreationTimestamp="2025-12-06 06:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:01:11.638839272 +0000 UTC m=+596.527822224" watchObservedRunningTime="2025-12-06 06:01:11.643096398 +0000 UTC m=+596.532079340" Dec 06 06:01:20 crc kubenswrapper[4809]: I1206 06:01:20.446991 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-664586d6fb-zgzv8" Dec 06 06:01:20 crc kubenswrapper[4809]: I1206 06:01:20.447429 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-664586d6fb-zgzv8" Dec 06 06:01:20 crc kubenswrapper[4809]: I1206 06:01:20.451254 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-664586d6fb-zgzv8" Dec 06 06:01:20 crc kubenswrapper[4809]: I1206 06:01:20.675023 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-664586d6fb-zgzv8" Dec 06 06:01:20 crc kubenswrapper[4809]: I1206 06:01:20.725000 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-7dc5585756-dv7bd"] Dec 06 06:01:34 crc kubenswrapper[4809]: I1206 06:01:34.495995 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:01:34 crc kubenswrapper[4809]: I1206 06:01:34.498035 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:01:45 crc kubenswrapper[4809]: I1206 06:01:45.765294 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-7dc5585756-dv7bd" podUID="0a07f515-be4e-4978-bd8b-ffac881b7a3f" containerName="console" containerID="cri-o://b47eaa725139cb91ace9f7260c25c06f4a0be75487d9511699bf9058c5316d54" gracePeriod=15 Dec 06 06:01:46 crc kubenswrapper[4809]: I1206 06:01:46.108583 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-7dc5585756-dv7bd_0a07f515-be4e-4978-bd8b-ffac881b7a3f/console/0.log" Dec 06 06:01:46 crc kubenswrapper[4809]: I1206 06:01:46.108889 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7dc5585756-dv7bd" Dec 06 06:01:46 crc kubenswrapper[4809]: I1206 06:01:46.282269 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0a07f515-be4e-4978-bd8b-ffac881b7a3f-console-oauth-config\") pod \"0a07f515-be4e-4978-bd8b-ffac881b7a3f\" (UID: \"0a07f515-be4e-4978-bd8b-ffac881b7a3f\") " Dec 06 06:01:46 crc kubenswrapper[4809]: I1206 06:01:46.282333 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0a07f515-be4e-4978-bd8b-ffac881b7a3f-oauth-serving-cert\") pod \"0a07f515-be4e-4978-bd8b-ffac881b7a3f\" (UID: \"0a07f515-be4e-4978-bd8b-ffac881b7a3f\") " Dec 06 06:01:46 crc kubenswrapper[4809]: I1206 06:01:46.282393 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0a07f515-be4e-4978-bd8b-ffac881b7a3f-service-ca\") pod \"0a07f515-be4e-4978-bd8b-ffac881b7a3f\" (UID: \"0a07f515-be4e-4978-bd8b-ffac881b7a3f\") " Dec 06 06:01:46 crc kubenswrapper[4809]: I1206 06:01:46.282415 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0a07f515-be4e-4978-bd8b-ffac881b7a3f-console-serving-cert\") pod \"0a07f515-be4e-4978-bd8b-ffac881b7a3f\" (UID: \"0a07f515-be4e-4978-bd8b-ffac881b7a3f\") " Dec 06 06:01:46 crc kubenswrapper[4809]: I1206 06:01:46.282455 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-txb97\" (UniqueName: \"kubernetes.io/projected/0a07f515-be4e-4978-bd8b-ffac881b7a3f-kube-api-access-txb97\") pod \"0a07f515-be4e-4978-bd8b-ffac881b7a3f\" (UID: \"0a07f515-be4e-4978-bd8b-ffac881b7a3f\") " Dec 06 06:01:46 crc kubenswrapper[4809]: I1206 06:01:46.282472 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0a07f515-be4e-4978-bd8b-ffac881b7a3f-trusted-ca-bundle\") pod \"0a07f515-be4e-4978-bd8b-ffac881b7a3f\" (UID: \"0a07f515-be4e-4978-bd8b-ffac881b7a3f\") " Dec 06 06:01:46 crc kubenswrapper[4809]: I1206 06:01:46.282555 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0a07f515-be4e-4978-bd8b-ffac881b7a3f-console-config\") pod \"0a07f515-be4e-4978-bd8b-ffac881b7a3f\" (UID: \"0a07f515-be4e-4978-bd8b-ffac881b7a3f\") " Dec 06 06:01:46 crc kubenswrapper[4809]: I1206 06:01:46.283192 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a07f515-be4e-4978-bd8b-ffac881b7a3f-console-config" (OuterVolumeSpecName: "console-config") pod "0a07f515-be4e-4978-bd8b-ffac881b7a3f" (UID: "0a07f515-be4e-4978-bd8b-ffac881b7a3f"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:01:46 crc kubenswrapper[4809]: I1206 06:01:46.283462 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a07f515-be4e-4978-bd8b-ffac881b7a3f-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "0a07f515-be4e-4978-bd8b-ffac881b7a3f" (UID: "0a07f515-be4e-4978-bd8b-ffac881b7a3f"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:01:46 crc kubenswrapper[4809]: I1206 06:01:46.283570 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a07f515-be4e-4978-bd8b-ffac881b7a3f-service-ca" (OuterVolumeSpecName: "service-ca") pod "0a07f515-be4e-4978-bd8b-ffac881b7a3f" (UID: "0a07f515-be4e-4978-bd8b-ffac881b7a3f"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:01:46 crc kubenswrapper[4809]: I1206 06:01:46.283829 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a07f515-be4e-4978-bd8b-ffac881b7a3f-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "0a07f515-be4e-4978-bd8b-ffac881b7a3f" (UID: "0a07f515-be4e-4978-bd8b-ffac881b7a3f"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:01:46 crc kubenswrapper[4809]: I1206 06:01:46.290119 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a07f515-be4e-4978-bd8b-ffac881b7a3f-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "0a07f515-be4e-4978-bd8b-ffac881b7a3f" (UID: "0a07f515-be4e-4978-bd8b-ffac881b7a3f"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:01:46 crc kubenswrapper[4809]: I1206 06:01:46.290197 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a07f515-be4e-4978-bd8b-ffac881b7a3f-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "0a07f515-be4e-4978-bd8b-ffac881b7a3f" (UID: "0a07f515-be4e-4978-bd8b-ffac881b7a3f"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:01:46 crc kubenswrapper[4809]: I1206 06:01:46.290221 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a07f515-be4e-4978-bd8b-ffac881b7a3f-kube-api-access-txb97" (OuterVolumeSpecName: "kube-api-access-txb97") pod "0a07f515-be4e-4978-bd8b-ffac881b7a3f" (UID: "0a07f515-be4e-4978-bd8b-ffac881b7a3f"). InnerVolumeSpecName "kube-api-access-txb97". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:01:46 crc kubenswrapper[4809]: I1206 06:01:46.384390 4809 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0a07f515-be4e-4978-bd8b-ffac881b7a3f-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:46 crc kubenswrapper[4809]: I1206 06:01:46.384427 4809 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0a07f515-be4e-4978-bd8b-ffac881b7a3f-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:46 crc kubenswrapper[4809]: I1206 06:01:46.384436 4809 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0a07f515-be4e-4978-bd8b-ffac881b7a3f-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:46 crc kubenswrapper[4809]: I1206 06:01:46.384445 4809 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0a07f515-be4e-4978-bd8b-ffac881b7a3f-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:46 crc kubenswrapper[4809]: I1206 06:01:46.384455 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-txb97\" (UniqueName: \"kubernetes.io/projected/0a07f515-be4e-4978-bd8b-ffac881b7a3f-kube-api-access-txb97\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:46 crc kubenswrapper[4809]: I1206 06:01:46.384465 4809 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0a07f515-be4e-4978-bd8b-ffac881b7a3f-console-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:46 crc kubenswrapper[4809]: I1206 06:01:46.384472 4809 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0a07f515-be4e-4978-bd8b-ffac881b7a3f-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:46 crc kubenswrapper[4809]: I1206 06:01:46.832813 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-7dc5585756-dv7bd_0a07f515-be4e-4978-bd8b-ffac881b7a3f/console/0.log" Dec 06 06:01:46 crc kubenswrapper[4809]: I1206 06:01:46.833219 4809 generic.go:334] "Generic (PLEG): container finished" podID="0a07f515-be4e-4978-bd8b-ffac881b7a3f" containerID="b47eaa725139cb91ace9f7260c25c06f4a0be75487d9511699bf9058c5316d54" exitCode=2 Dec 06 06:01:46 crc kubenswrapper[4809]: I1206 06:01:46.833258 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7dc5585756-dv7bd" event={"ID":"0a07f515-be4e-4978-bd8b-ffac881b7a3f","Type":"ContainerDied","Data":"b47eaa725139cb91ace9f7260c25c06f4a0be75487d9511699bf9058c5316d54"} Dec 06 06:01:46 crc kubenswrapper[4809]: I1206 06:01:46.833287 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7dc5585756-dv7bd" event={"ID":"0a07f515-be4e-4978-bd8b-ffac881b7a3f","Type":"ContainerDied","Data":"795c66f39956b918963aaa2311783d6be38583742482763abbd4629b96013ab5"} Dec 06 06:01:46 crc kubenswrapper[4809]: I1206 06:01:46.833306 4809 scope.go:117] "RemoveContainer" containerID="b47eaa725139cb91ace9f7260c25c06f4a0be75487d9511699bf9058c5316d54" Dec 06 06:01:46 crc kubenswrapper[4809]: I1206 06:01:46.833367 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7dc5585756-dv7bd" Dec 06 06:01:46 crc kubenswrapper[4809]: I1206 06:01:46.867100 4809 scope.go:117] "RemoveContainer" containerID="b47eaa725139cb91ace9f7260c25c06f4a0be75487d9511699bf9058c5316d54" Dec 06 06:01:46 crc kubenswrapper[4809]: E1206 06:01:46.867992 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b47eaa725139cb91ace9f7260c25c06f4a0be75487d9511699bf9058c5316d54\": container with ID starting with b47eaa725139cb91ace9f7260c25c06f4a0be75487d9511699bf9058c5316d54 not found: ID does not exist" containerID="b47eaa725139cb91ace9f7260c25c06f4a0be75487d9511699bf9058c5316d54" Dec 06 06:01:46 crc kubenswrapper[4809]: I1206 06:01:46.868055 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b47eaa725139cb91ace9f7260c25c06f4a0be75487d9511699bf9058c5316d54"} err="failed to get container status \"b47eaa725139cb91ace9f7260c25c06f4a0be75487d9511699bf9058c5316d54\": rpc error: code = NotFound desc = could not find container \"b47eaa725139cb91ace9f7260c25c06f4a0be75487d9511699bf9058c5316d54\": container with ID starting with b47eaa725139cb91ace9f7260c25c06f4a0be75487d9511699bf9058c5316d54 not found: ID does not exist" Dec 06 06:01:46 crc kubenswrapper[4809]: I1206 06:01:46.872658 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-7dc5585756-dv7bd"] Dec 06 06:01:46 crc kubenswrapper[4809]: I1206 06:01:46.878782 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-7dc5585756-dv7bd"] Dec 06 06:01:47 crc kubenswrapper[4809]: I1206 06:01:47.398451 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a07f515-be4e-4978-bd8b-ffac881b7a3f" path="/var/lib/kubelet/pods/0a07f515-be4e-4978-bd8b-ffac881b7a3f/volumes" Dec 06 06:02:04 crc kubenswrapper[4809]: I1206 06:02:04.496872 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:02:04 crc kubenswrapper[4809]: I1206 06:02:04.497620 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:02:04 crc kubenswrapper[4809]: I1206 06:02:04.497691 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-npms2" Dec 06 06:02:04 crc kubenswrapper[4809]: I1206 06:02:04.498577 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e6b027954a4e3b35818e27fb0ec5b9b81b52eb65e6a41e8dc2459f7e54dc0c30"} pod="openshift-machine-config-operator/machine-config-daemon-npms2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 06:02:04 crc kubenswrapper[4809]: I1206 06:02:04.498666 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" containerID="cri-o://e6b027954a4e3b35818e27fb0ec5b9b81b52eb65e6a41e8dc2459f7e54dc0c30" gracePeriod=600 Dec 06 06:02:04 crc kubenswrapper[4809]: I1206 06:02:04.944263 4809 generic.go:334] "Generic (PLEG): container finished" podID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerID="e6b027954a4e3b35818e27fb0ec5b9b81b52eb65e6a41e8dc2459f7e54dc0c30" exitCode=0 Dec 06 06:02:04 crc kubenswrapper[4809]: I1206 06:02:04.944358 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" event={"ID":"cbf4cf62-024e-4703-a8b8-9aecda9cd26a","Type":"ContainerDied","Data":"e6b027954a4e3b35818e27fb0ec5b9b81b52eb65e6a41e8dc2459f7e54dc0c30"} Dec 06 06:02:04 crc kubenswrapper[4809]: I1206 06:02:04.944625 4809 scope.go:117] "RemoveContainer" containerID="2442b241343358fa541b07551125d7c63b82fcdcd98be18afedabbdc8da517a8" Dec 06 06:02:05 crc kubenswrapper[4809]: I1206 06:02:05.953237 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" event={"ID":"cbf4cf62-024e-4703-a8b8-9aecda9cd26a","Type":"ContainerStarted","Data":"49ef8bff82d74abdc9d792d5696488d1189b4c9562bc3eb566b2639019bb757d"} Dec 06 06:04:01 crc kubenswrapper[4809]: I1206 06:04:01.641089 4809 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 06 06:04:34 crc kubenswrapper[4809]: I1206 06:04:34.496899 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:04:34 crc kubenswrapper[4809]: I1206 06:04:34.497494 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:05:00 crc kubenswrapper[4809]: I1206 06:05:00.334530 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210t4wcj"] Dec 06 06:05:00 crc kubenswrapper[4809]: E1206 06:05:00.335383 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a07f515-be4e-4978-bd8b-ffac881b7a3f" containerName="console" Dec 06 06:05:00 crc kubenswrapper[4809]: I1206 06:05:00.335414 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a07f515-be4e-4978-bd8b-ffac881b7a3f" containerName="console" Dec 06 06:05:00 crc kubenswrapper[4809]: I1206 06:05:00.335563 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a07f515-be4e-4978-bd8b-ffac881b7a3f" containerName="console" Dec 06 06:05:00 crc kubenswrapper[4809]: I1206 06:05:00.336585 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210t4wcj" Dec 06 06:05:00 crc kubenswrapper[4809]: I1206 06:05:00.338864 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 06 06:05:00 crc kubenswrapper[4809]: I1206 06:05:00.346074 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210t4wcj"] Dec 06 06:05:00 crc kubenswrapper[4809]: I1206 06:05:00.414277 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztkpb\" (UniqueName: \"kubernetes.io/projected/64828a07-1605-4981-a9ca-9442ca370be5-kube-api-access-ztkpb\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210t4wcj\" (UID: \"64828a07-1605-4981-a9ca-9442ca370be5\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210t4wcj" Dec 06 06:05:00 crc kubenswrapper[4809]: I1206 06:05:00.414385 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/64828a07-1605-4981-a9ca-9442ca370be5-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210t4wcj\" (UID: \"64828a07-1605-4981-a9ca-9442ca370be5\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210t4wcj" Dec 06 06:05:00 crc kubenswrapper[4809]: I1206 06:05:00.414445 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/64828a07-1605-4981-a9ca-9442ca370be5-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210t4wcj\" (UID: \"64828a07-1605-4981-a9ca-9442ca370be5\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210t4wcj" Dec 06 06:05:00 crc kubenswrapper[4809]: I1206 06:05:00.515557 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/64828a07-1605-4981-a9ca-9442ca370be5-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210t4wcj\" (UID: \"64828a07-1605-4981-a9ca-9442ca370be5\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210t4wcj" Dec 06 06:05:00 crc kubenswrapper[4809]: I1206 06:05:00.515654 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztkpb\" (UniqueName: \"kubernetes.io/projected/64828a07-1605-4981-a9ca-9442ca370be5-kube-api-access-ztkpb\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210t4wcj\" (UID: \"64828a07-1605-4981-a9ca-9442ca370be5\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210t4wcj" Dec 06 06:05:00 crc kubenswrapper[4809]: I1206 06:05:00.515721 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/64828a07-1605-4981-a9ca-9442ca370be5-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210t4wcj\" (UID: \"64828a07-1605-4981-a9ca-9442ca370be5\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210t4wcj" Dec 06 06:05:00 crc kubenswrapper[4809]: I1206 06:05:00.516195 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/64828a07-1605-4981-a9ca-9442ca370be5-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210t4wcj\" (UID: \"64828a07-1605-4981-a9ca-9442ca370be5\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210t4wcj" Dec 06 06:05:00 crc kubenswrapper[4809]: I1206 06:05:00.516207 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/64828a07-1605-4981-a9ca-9442ca370be5-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210t4wcj\" (UID: \"64828a07-1605-4981-a9ca-9442ca370be5\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210t4wcj" Dec 06 06:05:00 crc kubenswrapper[4809]: I1206 06:05:00.536367 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztkpb\" (UniqueName: \"kubernetes.io/projected/64828a07-1605-4981-a9ca-9442ca370be5-kube-api-access-ztkpb\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210t4wcj\" (UID: \"64828a07-1605-4981-a9ca-9442ca370be5\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210t4wcj" Dec 06 06:05:00 crc kubenswrapper[4809]: I1206 06:05:00.657229 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210t4wcj" Dec 06 06:05:00 crc kubenswrapper[4809]: I1206 06:05:00.913082 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210t4wcj"] Dec 06 06:05:01 crc kubenswrapper[4809]: I1206 06:05:01.009254 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210t4wcj" event={"ID":"64828a07-1605-4981-a9ca-9442ca370be5","Type":"ContainerStarted","Data":"0c7c6ed8e57884b8a920d1b1c708da0e68cfc9d336d979d5a0ac929dafcb4531"} Dec 06 06:05:02 crc kubenswrapper[4809]: I1206 06:05:02.015256 4809 generic.go:334] "Generic (PLEG): container finished" podID="64828a07-1605-4981-a9ca-9442ca370be5" containerID="0656cef5435be29ad092c38959b90d4acbe9b7205bb840f6cfab14c2514c8e73" exitCode=0 Dec 06 06:05:02 crc kubenswrapper[4809]: I1206 06:05:02.015300 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210t4wcj" event={"ID":"64828a07-1605-4981-a9ca-9442ca370be5","Type":"ContainerDied","Data":"0656cef5435be29ad092c38959b90d4acbe9b7205bb840f6cfab14c2514c8e73"} Dec 06 06:05:02 crc kubenswrapper[4809]: I1206 06:05:02.016582 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 06:05:02 crc kubenswrapper[4809]: I1206 06:05:02.698773 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-j88qv"] Dec 06 06:05:02 crc kubenswrapper[4809]: I1206 06:05:02.703846 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j88qv" Dec 06 06:05:02 crc kubenswrapper[4809]: I1206 06:05:02.714355 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j88qv"] Dec 06 06:05:02 crc kubenswrapper[4809]: I1206 06:05:02.755573 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qqjk\" (UniqueName: \"kubernetes.io/projected/3c5e2355-685f-45fe-98d4-ee0e0406075d-kube-api-access-9qqjk\") pod \"redhat-operators-j88qv\" (UID: \"3c5e2355-685f-45fe-98d4-ee0e0406075d\") " pod="openshift-marketplace/redhat-operators-j88qv" Dec 06 06:05:02 crc kubenswrapper[4809]: I1206 06:05:02.755697 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c5e2355-685f-45fe-98d4-ee0e0406075d-catalog-content\") pod \"redhat-operators-j88qv\" (UID: \"3c5e2355-685f-45fe-98d4-ee0e0406075d\") " pod="openshift-marketplace/redhat-operators-j88qv" Dec 06 06:05:02 crc kubenswrapper[4809]: I1206 06:05:02.755794 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c5e2355-685f-45fe-98d4-ee0e0406075d-utilities\") pod \"redhat-operators-j88qv\" (UID: \"3c5e2355-685f-45fe-98d4-ee0e0406075d\") " pod="openshift-marketplace/redhat-operators-j88qv" Dec 06 06:05:02 crc kubenswrapper[4809]: I1206 06:05:02.857222 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c5e2355-685f-45fe-98d4-ee0e0406075d-utilities\") pod \"redhat-operators-j88qv\" (UID: \"3c5e2355-685f-45fe-98d4-ee0e0406075d\") " pod="openshift-marketplace/redhat-operators-j88qv" Dec 06 06:05:02 crc kubenswrapper[4809]: I1206 06:05:02.857308 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qqjk\" (UniqueName: \"kubernetes.io/projected/3c5e2355-685f-45fe-98d4-ee0e0406075d-kube-api-access-9qqjk\") pod \"redhat-operators-j88qv\" (UID: \"3c5e2355-685f-45fe-98d4-ee0e0406075d\") " pod="openshift-marketplace/redhat-operators-j88qv" Dec 06 06:05:02 crc kubenswrapper[4809]: I1206 06:05:02.857351 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c5e2355-685f-45fe-98d4-ee0e0406075d-catalog-content\") pod \"redhat-operators-j88qv\" (UID: \"3c5e2355-685f-45fe-98d4-ee0e0406075d\") " pod="openshift-marketplace/redhat-operators-j88qv" Dec 06 06:05:02 crc kubenswrapper[4809]: I1206 06:05:02.857861 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c5e2355-685f-45fe-98d4-ee0e0406075d-catalog-content\") pod \"redhat-operators-j88qv\" (UID: \"3c5e2355-685f-45fe-98d4-ee0e0406075d\") " pod="openshift-marketplace/redhat-operators-j88qv" Dec 06 06:05:02 crc kubenswrapper[4809]: I1206 06:05:02.858151 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c5e2355-685f-45fe-98d4-ee0e0406075d-utilities\") pod \"redhat-operators-j88qv\" (UID: \"3c5e2355-685f-45fe-98d4-ee0e0406075d\") " pod="openshift-marketplace/redhat-operators-j88qv" Dec 06 06:05:02 crc kubenswrapper[4809]: I1206 06:05:02.897338 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qqjk\" (UniqueName: \"kubernetes.io/projected/3c5e2355-685f-45fe-98d4-ee0e0406075d-kube-api-access-9qqjk\") pod \"redhat-operators-j88qv\" (UID: \"3c5e2355-685f-45fe-98d4-ee0e0406075d\") " pod="openshift-marketplace/redhat-operators-j88qv" Dec 06 06:05:03 crc kubenswrapper[4809]: I1206 06:05:03.021437 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j88qv" Dec 06 06:05:03 crc kubenswrapper[4809]: I1206 06:05:03.232453 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j88qv"] Dec 06 06:05:04 crc kubenswrapper[4809]: I1206 06:05:04.028390 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j88qv" event={"ID":"3c5e2355-685f-45fe-98d4-ee0e0406075d","Type":"ContainerStarted","Data":"a1b7e312ad90ffbadb566f9fdcdd9eebe5f2803a5e7d2932b68c5bb3f6fd7781"} Dec 06 06:05:04 crc kubenswrapper[4809]: I1206 06:05:04.497005 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:05:04 crc kubenswrapper[4809]: I1206 06:05:04.497371 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:05:07 crc kubenswrapper[4809]: I1206 06:05:07.051241 4809 generic.go:334] "Generic (PLEG): container finished" podID="3c5e2355-685f-45fe-98d4-ee0e0406075d" containerID="8ef2d96a8b34261df3655cd85565e0a2642719e0394a5536122dba6ca33a07a0" exitCode=0 Dec 06 06:05:07 crc kubenswrapper[4809]: I1206 06:05:07.051318 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j88qv" event={"ID":"3c5e2355-685f-45fe-98d4-ee0e0406075d","Type":"ContainerDied","Data":"8ef2d96a8b34261df3655cd85565e0a2642719e0394a5536122dba6ca33a07a0"} Dec 06 06:05:09 crc kubenswrapper[4809]: I1206 06:05:09.085704 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210t4wcj" event={"ID":"64828a07-1605-4981-a9ca-9442ca370be5","Type":"ContainerStarted","Data":"9452b5710f0ea3bc42e0e949319c3c39936fbaccdde52734ebc49d04bf3812fa"} Dec 06 06:05:10 crc kubenswrapper[4809]: I1206 06:05:10.095294 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j88qv" event={"ID":"3c5e2355-685f-45fe-98d4-ee0e0406075d","Type":"ContainerStarted","Data":"e6a979852ffe0757e6bd2d20d548ff969847341def135a6c0dd423eb2ef71951"} Dec 06 06:05:10 crc kubenswrapper[4809]: I1206 06:05:10.097181 4809 generic.go:334] "Generic (PLEG): container finished" podID="64828a07-1605-4981-a9ca-9442ca370be5" containerID="9452b5710f0ea3bc42e0e949319c3c39936fbaccdde52734ebc49d04bf3812fa" exitCode=0 Dec 06 06:05:10 crc kubenswrapper[4809]: I1206 06:05:10.097226 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210t4wcj" event={"ID":"64828a07-1605-4981-a9ca-9442ca370be5","Type":"ContainerDied","Data":"9452b5710f0ea3bc42e0e949319c3c39936fbaccdde52734ebc49d04bf3812fa"} Dec 06 06:05:11 crc kubenswrapper[4809]: I1206 06:05:11.107849 4809 generic.go:334] "Generic (PLEG): container finished" podID="64828a07-1605-4981-a9ca-9442ca370be5" containerID="21cabb88bdd30ddc2eea10397bc28e8a8053a5db0c935dfc4a9d86e49f046bca" exitCode=0 Dec 06 06:05:11 crc kubenswrapper[4809]: I1206 06:05:11.108273 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210t4wcj" event={"ID":"64828a07-1605-4981-a9ca-9442ca370be5","Type":"ContainerDied","Data":"21cabb88bdd30ddc2eea10397bc28e8a8053a5db0c935dfc4a9d86e49f046bca"} Dec 06 06:05:11 crc kubenswrapper[4809]: I1206 06:05:11.110946 4809 generic.go:334] "Generic (PLEG): container finished" podID="3c5e2355-685f-45fe-98d4-ee0e0406075d" containerID="e6a979852ffe0757e6bd2d20d548ff969847341def135a6c0dd423eb2ef71951" exitCode=0 Dec 06 06:05:11 crc kubenswrapper[4809]: I1206 06:05:11.110995 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j88qv" event={"ID":"3c5e2355-685f-45fe-98d4-ee0e0406075d","Type":"ContainerDied","Data":"e6a979852ffe0757e6bd2d20d548ff969847341def135a6c0dd423eb2ef71951"} Dec 06 06:05:11 crc kubenswrapper[4809]: I1206 06:05:11.414296 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-5xvvb"] Dec 06 06:05:11 crc kubenswrapper[4809]: I1206 06:05:11.415002 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerName="ovn-controller" containerID="cri-o://5be86696b7a6518c5df930accb2c6890d58cfc37104fd34a359724b9e30231b1" gracePeriod=30 Dec 06 06:05:11 crc kubenswrapper[4809]: I1206 06:05:11.415180 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://83d58d489f532b78aca59f5b92584aad1dc292e6700c112c531cbeb3209fd90c" gracePeriod=30 Dec 06 06:05:11 crc kubenswrapper[4809]: I1206 06:05:11.415248 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerName="northd" containerID="cri-o://ab9b3bd4c8a6d6af7939b06fe01b8a0a50d81bb8c669bc6a27f486953c234e11" gracePeriod=30 Dec 06 06:05:11 crc kubenswrapper[4809]: I1206 06:05:11.415318 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerName="ovn-acl-logging" containerID="cri-o://0a3e4453935871bb84d68bc4a8b6c98f299b7c8094ba82291bc90e258e1fec55" gracePeriod=30 Dec 06 06:05:11 crc kubenswrapper[4809]: I1206 06:05:11.415222 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerName="kube-rbac-proxy-node" containerID="cri-o://c981561d51870deffeb446e8ca769b7ac0a1a16edbe4e982e0f2c84a45043db9" gracePeriod=30 Dec 06 06:05:11 crc kubenswrapper[4809]: I1206 06:05:11.415196 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerName="nbdb" containerID="cri-o://3dcd421b38e5f9ba0c542a09f0047bde6c043942a096d11575afb7423ab2552c" gracePeriod=30 Dec 06 06:05:11 crc kubenswrapper[4809]: I1206 06:05:11.415227 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerName="sbdb" containerID="cri-o://dee8b965be3b2486f1f532f2bdfbbcbd84f9b3b711a11c25e86dae1284314d4c" gracePeriod=30 Dec 06 06:05:11 crc kubenswrapper[4809]: I1206 06:05:11.465374 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerName="ovnkube-controller" containerID="cri-o://1cf68aa264056dd94a3e6687d35def2e9eb5580a4ee72a3acd46a061f4d06437" gracePeriod=30 Dec 06 06:05:12 crc kubenswrapper[4809]: I1206 06:05:12.119076 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j88qv" event={"ID":"3c5e2355-685f-45fe-98d4-ee0e0406075d","Type":"ContainerStarted","Data":"d86f9164d10d620c990222a1d697eef8646f314e472220d3c0405add4a8eb3fe"} Dec 06 06:05:12 crc kubenswrapper[4809]: I1206 06:05:12.122840 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5xvvb_ffb3a4f1-1e22-4220-ac80-3b2d69d2db99/ovnkube-controller/3.log" Dec 06 06:05:12 crc kubenswrapper[4809]: I1206 06:05:12.128028 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5xvvb_ffb3a4f1-1e22-4220-ac80-3b2d69d2db99/ovn-acl-logging/0.log" Dec 06 06:05:12 crc kubenswrapper[4809]: I1206 06:05:12.128603 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5xvvb_ffb3a4f1-1e22-4220-ac80-3b2d69d2db99/ovn-controller/0.log" Dec 06 06:05:12 crc kubenswrapper[4809]: I1206 06:05:12.129212 4809 generic.go:334] "Generic (PLEG): container finished" podID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerID="1cf68aa264056dd94a3e6687d35def2e9eb5580a4ee72a3acd46a061f4d06437" exitCode=0 Dec 06 06:05:12 crc kubenswrapper[4809]: I1206 06:05:12.129243 4809 generic.go:334] "Generic (PLEG): container finished" podID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerID="dee8b965be3b2486f1f532f2bdfbbcbd84f9b3b711a11c25e86dae1284314d4c" exitCode=0 Dec 06 06:05:12 crc kubenswrapper[4809]: I1206 06:05:12.129253 4809 generic.go:334] "Generic (PLEG): container finished" podID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerID="3dcd421b38e5f9ba0c542a09f0047bde6c043942a096d11575afb7423ab2552c" exitCode=0 Dec 06 06:05:12 crc kubenswrapper[4809]: I1206 06:05:12.129263 4809 generic.go:334] "Generic (PLEG): container finished" podID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerID="ab9b3bd4c8a6d6af7939b06fe01b8a0a50d81bb8c669bc6a27f486953c234e11" exitCode=0 Dec 06 06:05:12 crc kubenswrapper[4809]: I1206 06:05:12.129273 4809 generic.go:334] "Generic (PLEG): container finished" podID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerID="0a3e4453935871bb84d68bc4a8b6c98f299b7c8094ba82291bc90e258e1fec55" exitCode=143 Dec 06 06:05:12 crc kubenswrapper[4809]: I1206 06:05:12.129285 4809 generic.go:334] "Generic (PLEG): container finished" podID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerID="5be86696b7a6518c5df930accb2c6890d58cfc37104fd34a359724b9e30231b1" exitCode=143 Dec 06 06:05:12 crc kubenswrapper[4809]: I1206 06:05:12.129291 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" event={"ID":"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99","Type":"ContainerDied","Data":"1cf68aa264056dd94a3e6687d35def2e9eb5580a4ee72a3acd46a061f4d06437"} Dec 06 06:05:12 crc kubenswrapper[4809]: I1206 06:05:12.129337 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" event={"ID":"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99","Type":"ContainerDied","Data":"dee8b965be3b2486f1f532f2bdfbbcbd84f9b3b711a11c25e86dae1284314d4c"} Dec 06 06:05:12 crc kubenswrapper[4809]: I1206 06:05:12.129349 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" event={"ID":"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99","Type":"ContainerDied","Data":"3dcd421b38e5f9ba0c542a09f0047bde6c043942a096d11575afb7423ab2552c"} Dec 06 06:05:12 crc kubenswrapper[4809]: I1206 06:05:12.129359 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" event={"ID":"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99","Type":"ContainerDied","Data":"ab9b3bd4c8a6d6af7939b06fe01b8a0a50d81bb8c669bc6a27f486953c234e11"} Dec 06 06:05:12 crc kubenswrapper[4809]: I1206 06:05:12.129370 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" event={"ID":"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99","Type":"ContainerDied","Data":"0a3e4453935871bb84d68bc4a8b6c98f299b7c8094ba82291bc90e258e1fec55"} Dec 06 06:05:12 crc kubenswrapper[4809]: I1206 06:05:12.129381 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" event={"ID":"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99","Type":"ContainerDied","Data":"5be86696b7a6518c5df930accb2c6890d58cfc37104fd34a359724b9e30231b1"} Dec 06 06:05:12 crc kubenswrapper[4809]: I1206 06:05:12.129398 4809 scope.go:117] "RemoveContainer" containerID="4a4a7efb6435b711f6fc3d07f0f82258ae97be843a5917fe8017c2afd690c200" Dec 06 06:05:12 crc kubenswrapper[4809]: I1206 06:05:12.131400 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-cb4vc_ed42d686-41af-470e-910d-a1fabbec66b0/kube-multus/2.log" Dec 06 06:05:12 crc kubenswrapper[4809]: I1206 06:05:12.131856 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-cb4vc_ed42d686-41af-470e-910d-a1fabbec66b0/kube-multus/1.log" Dec 06 06:05:12 crc kubenswrapper[4809]: I1206 06:05:12.131894 4809 generic.go:334] "Generic (PLEG): container finished" podID="ed42d686-41af-470e-910d-a1fabbec66b0" containerID="b950762e7c636b40a9c1701711f6bee3048580e99b743938cbcb59c938281b24" exitCode=2 Dec 06 06:05:12 crc kubenswrapper[4809]: I1206 06:05:12.131966 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-cb4vc" event={"ID":"ed42d686-41af-470e-910d-a1fabbec66b0","Type":"ContainerDied","Data":"b950762e7c636b40a9c1701711f6bee3048580e99b743938cbcb59c938281b24"} Dec 06 06:05:12 crc kubenswrapper[4809]: I1206 06:05:12.132878 4809 scope.go:117] "RemoveContainer" containerID="b950762e7c636b40a9c1701711f6bee3048580e99b743938cbcb59c938281b24" Dec 06 06:05:12 crc kubenswrapper[4809]: I1206 06:05:12.144777 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-j88qv" podStartSLOduration=6.289282064 podStartE2EDuration="10.144756258s" podCreationTimestamp="2025-12-06 06:05:02 +0000 UTC" firstStartedPulling="2025-12-06 06:05:07.698875567 +0000 UTC m=+832.587858539" lastFinishedPulling="2025-12-06 06:05:11.554349791 +0000 UTC m=+836.443332733" observedRunningTime="2025-12-06 06:05:12.143050402 +0000 UTC m=+837.032033354" watchObservedRunningTime="2025-12-06 06:05:12.144756258 +0000 UTC m=+837.033739200" Dec 06 06:05:12 crc kubenswrapper[4809]: I1206 06:05:12.203652 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210t4wcj" Dec 06 06:05:12 crc kubenswrapper[4809]: I1206 06:05:12.314667 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/64828a07-1605-4981-a9ca-9442ca370be5-util\") pod \"64828a07-1605-4981-a9ca-9442ca370be5\" (UID: \"64828a07-1605-4981-a9ca-9442ca370be5\") " Dec 06 06:05:12 crc kubenswrapper[4809]: I1206 06:05:12.314818 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ztkpb\" (UniqueName: \"kubernetes.io/projected/64828a07-1605-4981-a9ca-9442ca370be5-kube-api-access-ztkpb\") pod \"64828a07-1605-4981-a9ca-9442ca370be5\" (UID: \"64828a07-1605-4981-a9ca-9442ca370be5\") " Dec 06 06:05:12 crc kubenswrapper[4809]: I1206 06:05:12.314890 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/64828a07-1605-4981-a9ca-9442ca370be5-bundle\") pod \"64828a07-1605-4981-a9ca-9442ca370be5\" (UID: \"64828a07-1605-4981-a9ca-9442ca370be5\") " Dec 06 06:05:12 crc kubenswrapper[4809]: I1206 06:05:12.319005 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64828a07-1605-4981-a9ca-9442ca370be5-bundle" (OuterVolumeSpecName: "bundle") pod "64828a07-1605-4981-a9ca-9442ca370be5" (UID: "64828a07-1605-4981-a9ca-9442ca370be5"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:05:12 crc kubenswrapper[4809]: I1206 06:05:12.323702 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64828a07-1605-4981-a9ca-9442ca370be5-kube-api-access-ztkpb" (OuterVolumeSpecName: "kube-api-access-ztkpb") pod "64828a07-1605-4981-a9ca-9442ca370be5" (UID: "64828a07-1605-4981-a9ca-9442ca370be5"). InnerVolumeSpecName "kube-api-access-ztkpb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:05:12 crc kubenswrapper[4809]: I1206 06:05:12.324677 4809 scope.go:117] "RemoveContainer" containerID="2edf7fa65761bc560178a40857780a5668e183a4b01974935cdbc3540915114f" Dec 06 06:05:12 crc kubenswrapper[4809]: I1206 06:05:12.327897 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64828a07-1605-4981-a9ca-9442ca370be5-util" (OuterVolumeSpecName: "util") pod "64828a07-1605-4981-a9ca-9442ca370be5" (UID: "64828a07-1605-4981-a9ca-9442ca370be5"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:05:12 crc kubenswrapper[4809]: I1206 06:05:12.416412 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ztkpb\" (UniqueName: \"kubernetes.io/projected/64828a07-1605-4981-a9ca-9442ca370be5-kube-api-access-ztkpb\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:12 crc kubenswrapper[4809]: I1206 06:05:12.416437 4809 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/64828a07-1605-4981-a9ca-9442ca370be5-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:12 crc kubenswrapper[4809]: I1206 06:05:12.416447 4809 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/64828a07-1605-4981-a9ca-9442ca370be5-util\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.021944 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-j88qv" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.022319 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-j88qv" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.140821 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-cb4vc_ed42d686-41af-470e-910d-a1fabbec66b0/kube-multus/2.log" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.140922 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-cb4vc" event={"ID":"ed42d686-41af-470e-910d-a1fabbec66b0","Type":"ContainerStarted","Data":"75e41cbfa1f23c1829931f0b0230f3e783d25cb76c9a7da1d743da9dceffdf3d"} Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.143710 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210t4wcj" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.143710 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210t4wcj" event={"ID":"64828a07-1605-4981-a9ca-9442ca370be5","Type":"ContainerDied","Data":"0c7c6ed8e57884b8a920d1b1c708da0e68cfc9d336d979d5a0ac929dafcb4531"} Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.143849 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0c7c6ed8e57884b8a920d1b1c708da0e68cfc9d336d979d5a0ac929dafcb4531" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.147473 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5xvvb_ffb3a4f1-1e22-4220-ac80-3b2d69d2db99/ovn-acl-logging/0.log" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.147907 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5xvvb_ffb3a4f1-1e22-4220-ac80-3b2d69d2db99/ovn-controller/0.log" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.148234 4809 generic.go:334] "Generic (PLEG): container finished" podID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerID="83d58d489f532b78aca59f5b92584aad1dc292e6700c112c531cbeb3209fd90c" exitCode=0 Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.148258 4809 generic.go:334] "Generic (PLEG): container finished" podID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerID="c981561d51870deffeb446e8ca769b7ac0a1a16edbe4e982e0f2c84a45043db9" exitCode=0 Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.148833 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" event={"ID":"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99","Type":"ContainerDied","Data":"83d58d489f532b78aca59f5b92584aad1dc292e6700c112c531cbeb3209fd90c"} Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.148861 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" event={"ID":"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99","Type":"ContainerDied","Data":"c981561d51870deffeb446e8ca769b7ac0a1a16edbe4e982e0f2c84a45043db9"} Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.340621 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5xvvb_ffb3a4f1-1e22-4220-ac80-3b2d69d2db99/ovn-acl-logging/0.log" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.341311 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5xvvb_ffb3a4f1-1e22-4220-ac80-3b2d69d2db99/ovn-controller/0.log" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.341816 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.400759 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-vlv76"] Dec 06 06:05:13 crc kubenswrapper[4809]: E1206 06:05:13.401551 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64828a07-1605-4981-a9ca-9442ca370be5" containerName="util" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.401649 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="64828a07-1605-4981-a9ca-9442ca370be5" containerName="util" Dec 06 06:05:13 crc kubenswrapper[4809]: E1206 06:05:13.401744 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerName="nbdb" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.401858 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerName="nbdb" Dec 06 06:05:13 crc kubenswrapper[4809]: E1206 06:05:13.401962 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerName="ovnkube-controller" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.402041 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerName="ovnkube-controller" Dec 06 06:05:13 crc kubenswrapper[4809]: E1206 06:05:13.402123 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerName="kube-rbac-proxy-ovn-metrics" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.402194 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerName="kube-rbac-proxy-ovn-metrics" Dec 06 06:05:13 crc kubenswrapper[4809]: E1206 06:05:13.402272 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerName="ovn-acl-logging" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.402350 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerName="ovn-acl-logging" Dec 06 06:05:13 crc kubenswrapper[4809]: E1206 06:05:13.402442 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerName="ovnkube-controller" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.402532 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerName="ovnkube-controller" Dec 06 06:05:13 crc kubenswrapper[4809]: E1206 06:05:13.402612 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64828a07-1605-4981-a9ca-9442ca370be5" containerName="pull" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.402686 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="64828a07-1605-4981-a9ca-9442ca370be5" containerName="pull" Dec 06 06:05:13 crc kubenswrapper[4809]: E1206 06:05:13.402764 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerName="kube-rbac-proxy-node" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.402842 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerName="kube-rbac-proxy-node" Dec 06 06:05:13 crc kubenswrapper[4809]: E1206 06:05:13.402947 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64828a07-1605-4981-a9ca-9442ca370be5" containerName="extract" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.403051 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="64828a07-1605-4981-a9ca-9442ca370be5" containerName="extract" Dec 06 06:05:13 crc kubenswrapper[4809]: E1206 06:05:13.403133 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerName="northd" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.403202 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerName="northd" Dec 06 06:05:13 crc kubenswrapper[4809]: E1206 06:05:13.403274 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerName="ovn-controller" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.403351 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerName="ovn-controller" Dec 06 06:05:13 crc kubenswrapper[4809]: E1206 06:05:13.403442 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerName="kubecfg-setup" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.403529 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerName="kubecfg-setup" Dec 06 06:05:13 crc kubenswrapper[4809]: E1206 06:05:13.403608 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerName="sbdb" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.403676 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerName="sbdb" Dec 06 06:05:13 crc kubenswrapper[4809]: E1206 06:05:13.403745 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerName="ovnkube-controller" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.403812 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerName="ovnkube-controller" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.404086 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerName="ovnkube-controller" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.404194 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerName="nbdb" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.404269 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerName="sbdb" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.404349 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerName="ovn-controller" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.404421 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerName="northd" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.404506 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerName="ovn-acl-logging" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.404575 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerName="ovnkube-controller" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.404653 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="64828a07-1605-4981-a9ca-9442ca370be5" containerName="extract" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.404727 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerName="ovnkube-controller" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.404796 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerName="ovnkube-controller" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.404872 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerName="kube-rbac-proxy-node" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.404966 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerName="kube-rbac-proxy-ovn-metrics" Dec 06 06:05:13 crc kubenswrapper[4809]: E1206 06:05:13.405262 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerName="ovnkube-controller" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.405376 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerName="ovnkube-controller" Dec 06 06:05:13 crc kubenswrapper[4809]: E1206 06:05:13.405456 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerName="ovnkube-controller" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.405525 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerName="ovnkube-controller" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.405766 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" containerName="ovnkube-controller" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.409093 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.436078 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-systemd-units\") pod \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.436240 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" (UID: "ffb3a4f1-1e22-4220-ac80-3b2d69d2db99"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.436365 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.436410 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" (UID: "ffb3a4f1-1e22-4220-ac80-3b2d69d2db99"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.436533 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-etc-openvswitch\") pod \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.436606 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-host-cni-netd\") pod \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.436680 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-log-socket\") pod \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.436602 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" (UID: "ffb3a4f1-1e22-4220-ac80-3b2d69d2db99"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.436634 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" (UID: "ffb3a4f1-1e22-4220-ac80-3b2d69d2db99"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.436843 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-log-socket" (OuterVolumeSpecName: "log-socket") pod "ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" (UID: "ffb3a4f1-1e22-4220-ac80-3b2d69d2db99"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.436915 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-ovnkube-script-lib\") pod \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.437062 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-host-slash\") pod \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.437132 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-run-ovn\") pod \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.437196 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-host-run-ovn-kubernetes\") pod \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.437277 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-var-lib-openvswitch\") pod \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.437504 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-host-kubelet\") pod \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.437286 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" (UID: "ffb3a4f1-1e22-4220-ac80-3b2d69d2db99"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.437309 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-host-slash" (OuterVolumeSpecName: "host-slash") pod "ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" (UID: "ffb3a4f1-1e22-4220-ac80-3b2d69d2db99"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.437325 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" (UID: "ffb3a4f1-1e22-4220-ac80-3b2d69d2db99"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.437375 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" (UID: "ffb3a4f1-1e22-4220-ac80-3b2d69d2db99"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.437414 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" (UID: "ffb3a4f1-1e22-4220-ac80-3b2d69d2db99"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.437644 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" (UID: "ffb3a4f1-1e22-4220-ac80-3b2d69d2db99"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.437850 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-ovn-node-metrics-cert\") pod \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.437942 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-env-overrides\") pod \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.438022 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-node-log\") pod \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.438097 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-run-openvswitch\") pod \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.438342 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-host-cni-bin\") pod \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.438427 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-227vc\" (UniqueName: \"kubernetes.io/projected/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-kube-api-access-227vc\") pod \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.438532 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-ovnkube-config\") pod \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.438619 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-run-systemd\") pod \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.438713 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-host-run-netns\") pod \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\" (UID: \"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99\") " Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.438373 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-node-log" (OuterVolumeSpecName: "node-log") pod "ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" (UID: "ffb3a4f1-1e22-4220-ac80-3b2d69d2db99"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.438370 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" (UID: "ffb3a4f1-1e22-4220-ac80-3b2d69d2db99"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.438423 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" (UID: "ffb3a4f1-1e22-4220-ac80-3b2d69d2db99"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.438537 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" (UID: "ffb3a4f1-1e22-4220-ac80-3b2d69d2db99"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.438815 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" (UID: "ffb3a4f1-1e22-4220-ac80-3b2d69d2db99"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.439118 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" (UID: "ffb3a4f1-1e22-4220-ac80-3b2d69d2db99"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.439335 4809 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.439415 4809 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-node-log\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.439469 4809 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.439534 4809 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.440577 4809 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.440652 4809 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.440710 4809 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.440777 4809 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.440832 4809 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.440885 4809 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.440951 4809 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-log-socket\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.441009 4809 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.441061 4809 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-host-slash\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.441120 4809 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.441178 4809 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.441230 4809 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.441281 4809 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.442899 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" (UID: "ffb3a4f1-1e22-4220-ac80-3b2d69d2db99"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.444669 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-kube-api-access-227vc" (OuterVolumeSpecName: "kube-api-access-227vc") pod "ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" (UID: "ffb3a4f1-1e22-4220-ac80-3b2d69d2db99"). InnerVolumeSpecName "kube-api-access-227vc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.455579 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" (UID: "ffb3a4f1-1e22-4220-ac80-3b2d69d2db99"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.542722 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9024f80a-36d2-48c4-ba94-14032531ea73-ovnkube-script-lib\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.542793 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9024f80a-36d2-48c4-ba94-14032531ea73-run-systemd\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.542832 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9024f80a-36d2-48c4-ba94-14032531ea73-host-run-netns\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.542876 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9024f80a-36d2-48c4-ba94-14032531ea73-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.542917 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9024f80a-36d2-48c4-ba94-14032531ea73-run-ovn\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.542973 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6drb\" (UniqueName: \"kubernetes.io/projected/9024f80a-36d2-48c4-ba94-14032531ea73-kube-api-access-l6drb\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.543018 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9024f80a-36d2-48c4-ba94-14032531ea73-env-overrides\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.543066 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9024f80a-36d2-48c4-ba94-14032531ea73-systemd-units\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.543160 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9024f80a-36d2-48c4-ba94-14032531ea73-host-slash\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.543193 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9024f80a-36d2-48c4-ba94-14032531ea73-ovn-node-metrics-cert\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.543236 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9024f80a-36d2-48c4-ba94-14032531ea73-ovnkube-config\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.543314 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9024f80a-36d2-48c4-ba94-14032531ea73-node-log\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.543364 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9024f80a-36d2-48c4-ba94-14032531ea73-host-kubelet\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.543402 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9024f80a-36d2-48c4-ba94-14032531ea73-host-cni-bin\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.543436 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9024f80a-36d2-48c4-ba94-14032531ea73-run-openvswitch\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.543475 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9024f80a-36d2-48c4-ba94-14032531ea73-host-cni-netd\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.543515 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9024f80a-36d2-48c4-ba94-14032531ea73-var-lib-openvswitch\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.543668 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9024f80a-36d2-48c4-ba94-14032531ea73-log-socket\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.543811 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9024f80a-36d2-48c4-ba94-14032531ea73-etc-openvswitch\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.544075 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9024f80a-36d2-48c4-ba94-14032531ea73-host-run-ovn-kubernetes\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.544498 4809 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.544573 4809 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.544659 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-227vc\" (UniqueName: \"kubernetes.io/projected/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99-kube-api-access-227vc\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.645803 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9024f80a-36d2-48c4-ba94-14032531ea73-host-cni-bin\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.645846 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9024f80a-36d2-48c4-ba94-14032531ea73-run-openvswitch\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.645866 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9024f80a-36d2-48c4-ba94-14032531ea73-host-cni-netd\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.645886 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9024f80a-36d2-48c4-ba94-14032531ea73-var-lib-openvswitch\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.645906 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9024f80a-36d2-48c4-ba94-14032531ea73-log-socket\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.645911 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9024f80a-36d2-48c4-ba94-14032531ea73-host-cni-bin\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.645974 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9024f80a-36d2-48c4-ba94-14032531ea73-var-lib-openvswitch\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.645987 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9024f80a-36d2-48c4-ba94-14032531ea73-etc-openvswitch\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.645945 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9024f80a-36d2-48c4-ba94-14032531ea73-host-cni-netd\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.645945 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9024f80a-36d2-48c4-ba94-14032531ea73-etc-openvswitch\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.646021 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9024f80a-36d2-48c4-ba94-14032531ea73-log-socket\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.646039 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9024f80a-36d2-48c4-ba94-14032531ea73-run-openvswitch\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.646054 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9024f80a-36d2-48c4-ba94-14032531ea73-host-run-ovn-kubernetes\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.646092 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9024f80a-36d2-48c4-ba94-14032531ea73-ovnkube-script-lib\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.646111 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9024f80a-36d2-48c4-ba94-14032531ea73-run-systemd\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.646134 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9024f80a-36d2-48c4-ba94-14032531ea73-host-run-netns\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.646162 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9024f80a-36d2-48c4-ba94-14032531ea73-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.646187 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9024f80a-36d2-48c4-ba94-14032531ea73-run-ovn\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.646205 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6drb\" (UniqueName: \"kubernetes.io/projected/9024f80a-36d2-48c4-ba94-14032531ea73-kube-api-access-l6drb\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.646208 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9024f80a-36d2-48c4-ba94-14032531ea73-host-run-ovn-kubernetes\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.646232 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9024f80a-36d2-48c4-ba94-14032531ea73-env-overrides\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.646346 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9024f80a-36d2-48c4-ba94-14032531ea73-systemd-units\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.646406 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9024f80a-36d2-48c4-ba94-14032531ea73-host-slash\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.646436 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9024f80a-36d2-48c4-ba94-14032531ea73-ovn-node-metrics-cert\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.646480 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9024f80a-36d2-48c4-ba94-14032531ea73-ovnkube-config\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.646512 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9024f80a-36d2-48c4-ba94-14032531ea73-node-log\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.646551 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9024f80a-36d2-48c4-ba94-14032531ea73-host-kubelet\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.646661 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9024f80a-36d2-48c4-ba94-14032531ea73-host-kubelet\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.646689 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9024f80a-36d2-48c4-ba94-14032531ea73-systemd-units\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.646714 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9024f80a-36d2-48c4-ba94-14032531ea73-host-slash\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.646793 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9024f80a-36d2-48c4-ba94-14032531ea73-ovnkube-script-lib\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.646842 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9024f80a-36d2-48c4-ba94-14032531ea73-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.646877 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9024f80a-36d2-48c4-ba94-14032531ea73-run-systemd\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.646906 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9024f80a-36d2-48c4-ba94-14032531ea73-host-run-netns\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.647267 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9024f80a-36d2-48c4-ba94-14032531ea73-run-ovn\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.647278 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9024f80a-36d2-48c4-ba94-14032531ea73-node-log\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.647406 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9024f80a-36d2-48c4-ba94-14032531ea73-ovnkube-config\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.649965 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9024f80a-36d2-48c4-ba94-14032531ea73-ovn-node-metrics-cert\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.651536 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9024f80a-36d2-48c4-ba94-14032531ea73-env-overrides\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.664294 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6drb\" (UniqueName: \"kubernetes.io/projected/9024f80a-36d2-48c4-ba94-14032531ea73-kube-api-access-l6drb\") pod \"ovnkube-node-vlv76\" (UID: \"9024f80a-36d2-48c4-ba94-14032531ea73\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: I1206 06:05:13.727921 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:13 crc kubenswrapper[4809]: W1206 06:05:13.752561 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9024f80a_36d2_48c4_ba94_14032531ea73.slice/crio-341eaf51ec6ba5c97887f3265c77cf5cfb618cebc358a5b1c991c49e8cbf2086 WatchSource:0}: Error finding container 341eaf51ec6ba5c97887f3265c77cf5cfb618cebc358a5b1c991c49e8cbf2086: Status 404 returned error can't find the container with id 341eaf51ec6ba5c97887f3265c77cf5cfb618cebc358a5b1c991c49e8cbf2086 Dec 06 06:05:14 crc kubenswrapper[4809]: I1206 06:05:14.067281 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-j88qv" podUID="3c5e2355-685f-45fe-98d4-ee0e0406075d" containerName="registry-server" probeResult="failure" output=< Dec 06 06:05:14 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Dec 06 06:05:14 crc kubenswrapper[4809]: > Dec 06 06:05:14 crc kubenswrapper[4809]: I1206 06:05:14.167189 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5xvvb_ffb3a4f1-1e22-4220-ac80-3b2d69d2db99/ovn-acl-logging/0.log" Dec 06 06:05:14 crc kubenswrapper[4809]: I1206 06:05:14.167792 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5xvvb_ffb3a4f1-1e22-4220-ac80-3b2d69d2db99/ovn-controller/0.log" Dec 06 06:05:14 crc kubenswrapper[4809]: I1206 06:05:14.169087 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" event={"ID":"ffb3a4f1-1e22-4220-ac80-3b2d69d2db99","Type":"ContainerDied","Data":"91e57bd1e4bba7f733c10d4f92b20420e5365a2ab7815db7bba835c242b62f49"} Dec 06 06:05:14 crc kubenswrapper[4809]: I1206 06:05:14.169165 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-5xvvb" Dec 06 06:05:14 crc kubenswrapper[4809]: I1206 06:05:14.169197 4809 scope.go:117] "RemoveContainer" containerID="1cf68aa264056dd94a3e6687d35def2e9eb5580a4ee72a3acd46a061f4d06437" Dec 06 06:05:14 crc kubenswrapper[4809]: I1206 06:05:14.173002 4809 generic.go:334] "Generic (PLEG): container finished" podID="9024f80a-36d2-48c4-ba94-14032531ea73" containerID="d69af1b611cb80b8f07e97ace810ea3280ec3a9df3697dc16d998822e77674c3" exitCode=0 Dec 06 06:05:14 crc kubenswrapper[4809]: I1206 06:05:14.173734 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" event={"ID":"9024f80a-36d2-48c4-ba94-14032531ea73","Type":"ContainerDied","Data":"d69af1b611cb80b8f07e97ace810ea3280ec3a9df3697dc16d998822e77674c3"} Dec 06 06:05:14 crc kubenswrapper[4809]: I1206 06:05:14.173788 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" event={"ID":"9024f80a-36d2-48c4-ba94-14032531ea73","Type":"ContainerStarted","Data":"341eaf51ec6ba5c97887f3265c77cf5cfb618cebc358a5b1c991c49e8cbf2086"} Dec 06 06:05:14 crc kubenswrapper[4809]: I1206 06:05:14.224098 4809 scope.go:117] "RemoveContainer" containerID="dee8b965be3b2486f1f532f2bdfbbcbd84f9b3b711a11c25e86dae1284314d4c" Dec 06 06:05:14 crc kubenswrapper[4809]: I1206 06:05:14.243073 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-5xvvb"] Dec 06 06:05:14 crc kubenswrapper[4809]: I1206 06:05:14.248056 4809 scope.go:117] "RemoveContainer" containerID="3dcd421b38e5f9ba0c542a09f0047bde6c043942a096d11575afb7423ab2552c" Dec 06 06:05:14 crc kubenswrapper[4809]: I1206 06:05:14.248189 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-5xvvb"] Dec 06 06:05:14 crc kubenswrapper[4809]: I1206 06:05:14.273029 4809 scope.go:117] "RemoveContainer" containerID="ab9b3bd4c8a6d6af7939b06fe01b8a0a50d81bb8c669bc6a27f486953c234e11" Dec 06 06:05:14 crc kubenswrapper[4809]: I1206 06:05:14.291215 4809 scope.go:117] "RemoveContainer" containerID="83d58d489f532b78aca59f5b92584aad1dc292e6700c112c531cbeb3209fd90c" Dec 06 06:05:14 crc kubenswrapper[4809]: E1206 06:05:14.291773 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podffb3a4f1_1e22_4220_ac80_3b2d69d2db99.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9024f80a_36d2_48c4_ba94_14032531ea73.slice/crio-conmon-d69af1b611cb80b8f07e97ace810ea3280ec3a9df3697dc16d998822e77674c3.scope\": RecentStats: unable to find data in memory cache]" Dec 06 06:05:14 crc kubenswrapper[4809]: I1206 06:05:14.314400 4809 scope.go:117] "RemoveContainer" containerID="c981561d51870deffeb446e8ca769b7ac0a1a16edbe4e982e0f2c84a45043db9" Dec 06 06:05:14 crc kubenswrapper[4809]: I1206 06:05:14.335429 4809 scope.go:117] "RemoveContainer" containerID="0a3e4453935871bb84d68bc4a8b6c98f299b7c8094ba82291bc90e258e1fec55" Dec 06 06:05:14 crc kubenswrapper[4809]: I1206 06:05:14.352857 4809 scope.go:117] "RemoveContainer" containerID="5be86696b7a6518c5df930accb2c6890d58cfc37104fd34a359724b9e30231b1" Dec 06 06:05:14 crc kubenswrapper[4809]: I1206 06:05:14.370870 4809 scope.go:117] "RemoveContainer" containerID="3316dfe676913024c7e2fcc655fd9d5ee13d4ac415ae38c93cff13322be4339b" Dec 06 06:05:15 crc kubenswrapper[4809]: I1206 06:05:15.181720 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" event={"ID":"9024f80a-36d2-48c4-ba94-14032531ea73","Type":"ContainerStarted","Data":"60cf6051424247f3661db4e6243045217337273489de61b2f47cad893648b4fa"} Dec 06 06:05:15 crc kubenswrapper[4809]: I1206 06:05:15.394961 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ffb3a4f1-1e22-4220-ac80-3b2d69d2db99" path="/var/lib/kubelet/pods/ffb3a4f1-1e22-4220-ac80-3b2d69d2db99/volumes" Dec 06 06:05:17 crc kubenswrapper[4809]: I1206 06:05:17.197576 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" event={"ID":"9024f80a-36d2-48c4-ba94-14032531ea73","Type":"ContainerStarted","Data":"3cf8c46ab5864dd1d69321cd697004bd5619bf389ca4ef195e89e3ee6b849116"} Dec 06 06:05:17 crc kubenswrapper[4809]: I1206 06:05:17.198222 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" event={"ID":"9024f80a-36d2-48c4-ba94-14032531ea73","Type":"ContainerStarted","Data":"40ed18f53712695f6fe02565f77bd0b94642144547eba5c529753b2e24a59fe3"} Dec 06 06:05:17 crc kubenswrapper[4809]: I1206 06:05:17.198239 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" event={"ID":"9024f80a-36d2-48c4-ba94-14032531ea73","Type":"ContainerStarted","Data":"b9e59154245ecc177b4d99ffc8dccd9e1d2ea61d719eb9316640627064b562ce"} Dec 06 06:05:17 crc kubenswrapper[4809]: I1206 06:05:17.198252 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" event={"ID":"9024f80a-36d2-48c4-ba94-14032531ea73","Type":"ContainerStarted","Data":"a00688253293a5585c4f9d5d61773839a8eebe4af2131c470c14ff861eeccb34"} Dec 06 06:05:18 crc kubenswrapper[4809]: I1206 06:05:18.208863 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" event={"ID":"9024f80a-36d2-48c4-ba94-14032531ea73","Type":"ContainerStarted","Data":"4c318993f71801ad40901c8d7294b1048d2a48d0a6e8a9d95566ae859b008df0"} Dec 06 06:05:22 crc kubenswrapper[4809]: I1206 06:05:22.235757 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" event={"ID":"9024f80a-36d2-48c4-ba94-14032531ea73","Type":"ContainerStarted","Data":"6b8a2dc59d44f557eac2771a51d9765ef8afe22872654fd9e87ff78ee3832687"} Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.061039 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-j88qv" Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.109530 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-j88qv" Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.155560 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-h7k8w"] Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.156446 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-h7k8w" Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.168057 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.168099 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.168250 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-rvbjt" Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.259964 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" event={"ID":"9024f80a-36d2-48c4-ba94-14032531ea73","Type":"ContainerStarted","Data":"a79020f2f4a8bb89750d6188f6baec282228388e90e513d1703dd4c8de2924db"} Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.282876 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqfqd\" (UniqueName: \"kubernetes.io/projected/c034c52b-dd97-4722-84fe-b3389a008fbd-kube-api-access-zqfqd\") pod \"obo-prometheus-operator-668cf9dfbb-h7k8w\" (UID: \"c034c52b-dd97-4722-84fe-b3389a008fbd\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-h7k8w" Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.337468 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-t8tqd"] Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.338293 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-t8tqd" Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.344630 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-99254"] Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.345670 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-99254" Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.347139 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.347730 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-vq42s" Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.384207 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqfqd\" (UniqueName: \"kubernetes.io/projected/c034c52b-dd97-4722-84fe-b3389a008fbd-kube-api-access-zqfqd\") pod \"obo-prometheus-operator-668cf9dfbb-h7k8w\" (UID: \"c034c52b-dd97-4722-84fe-b3389a008fbd\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-h7k8w" Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.411820 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqfqd\" (UniqueName: \"kubernetes.io/projected/c034c52b-dd97-4722-84fe-b3389a008fbd-kube-api-access-zqfqd\") pod \"obo-prometheus-operator-668cf9dfbb-h7k8w\" (UID: \"c034c52b-dd97-4722-84fe-b3389a008fbd\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-h7k8w" Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.485730 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/35c30afa-953e-4808-b4ba-473ffb44c4a5-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-566bd9db68-99254\" (UID: \"35c30afa-953e-4808-b4ba-473ffb44c4a5\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-99254" Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.485784 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/51bf2206-9c20-4c9a-8dbd-1f74c0033236-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-566bd9db68-t8tqd\" (UID: \"51bf2206-9c20-4c9a-8dbd-1f74c0033236\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-t8tqd" Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.485865 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/35c30afa-953e-4808-b4ba-473ffb44c4a5-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-566bd9db68-99254\" (UID: \"35c30afa-953e-4808-b4ba-473ffb44c4a5\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-99254" Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.485888 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/51bf2206-9c20-4c9a-8dbd-1f74c0033236-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-566bd9db68-t8tqd\" (UID: \"51bf2206-9c20-4c9a-8dbd-1f74c0033236\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-t8tqd" Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.495923 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-bfc8s"] Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.496686 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-bfc8s" Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.501532 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-h7k8w" Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.503177 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-4d5vf" Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.504185 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Dec 06 06:05:23 crc kubenswrapper[4809]: E1206 06:05:23.537610 4809 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-h7k8w_openshift-operators_c034c52b-dd97-4722-84fe-b3389a008fbd_0(41168918e740c27c956ed6156d38e93a8ae305ee2ffe4cbefb5a7856f2f43e49): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 06 06:05:23 crc kubenswrapper[4809]: E1206 06:05:23.537692 4809 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-h7k8w_openshift-operators_c034c52b-dd97-4722-84fe-b3389a008fbd_0(41168918e740c27c956ed6156d38e93a8ae305ee2ffe4cbefb5a7856f2f43e49): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-h7k8w" Dec 06 06:05:23 crc kubenswrapper[4809]: E1206 06:05:23.537723 4809 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-h7k8w_openshift-operators_c034c52b-dd97-4722-84fe-b3389a008fbd_0(41168918e740c27c956ed6156d38e93a8ae305ee2ffe4cbefb5a7856f2f43e49): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-h7k8w" Dec 06 06:05:23 crc kubenswrapper[4809]: E1206 06:05:23.537774 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-668cf9dfbb-h7k8w_openshift-operators(c034c52b-dd97-4722-84fe-b3389a008fbd)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-668cf9dfbb-h7k8w_openshift-operators(c034c52b-dd97-4722-84fe-b3389a008fbd)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-h7k8w_openshift-operators_c034c52b-dd97-4722-84fe-b3389a008fbd_0(41168918e740c27c956ed6156d38e93a8ae305ee2ffe4cbefb5a7856f2f43e49): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-h7k8w" podUID="c034c52b-dd97-4722-84fe-b3389a008fbd" Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.586783 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/6d4f55df-5dc2-4e72-b7f8-cee9229da476-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-bfc8s\" (UID: \"6d4f55df-5dc2-4e72-b7f8-cee9229da476\") " pod="openshift-operators/observability-operator-d8bb48f5d-bfc8s" Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.586845 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/35c30afa-953e-4808-b4ba-473ffb44c4a5-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-566bd9db68-99254\" (UID: \"35c30afa-953e-4808-b4ba-473ffb44c4a5\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-99254" Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.586869 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/51bf2206-9c20-4c9a-8dbd-1f74c0033236-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-566bd9db68-t8tqd\" (UID: \"51bf2206-9c20-4c9a-8dbd-1f74c0033236\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-t8tqd" Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.586906 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/35c30afa-953e-4808-b4ba-473ffb44c4a5-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-566bd9db68-99254\" (UID: \"35c30afa-953e-4808-b4ba-473ffb44c4a5\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-99254" Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.586923 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/51bf2206-9c20-4c9a-8dbd-1f74c0033236-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-566bd9db68-t8tqd\" (UID: \"51bf2206-9c20-4c9a-8dbd-1f74c0033236\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-t8tqd" Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.586977 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lh545\" (UniqueName: \"kubernetes.io/projected/6d4f55df-5dc2-4e72-b7f8-cee9229da476-kube-api-access-lh545\") pod \"observability-operator-d8bb48f5d-bfc8s\" (UID: \"6d4f55df-5dc2-4e72-b7f8-cee9229da476\") " pod="openshift-operators/observability-operator-d8bb48f5d-bfc8s" Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.592184 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/35c30afa-953e-4808-b4ba-473ffb44c4a5-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-566bd9db68-99254\" (UID: \"35c30afa-953e-4808-b4ba-473ffb44c4a5\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-99254" Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.592184 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/51bf2206-9c20-4c9a-8dbd-1f74c0033236-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-566bd9db68-t8tqd\" (UID: \"51bf2206-9c20-4c9a-8dbd-1f74c0033236\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-t8tqd" Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.592733 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/51bf2206-9c20-4c9a-8dbd-1f74c0033236-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-566bd9db68-t8tqd\" (UID: \"51bf2206-9c20-4c9a-8dbd-1f74c0033236\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-t8tqd" Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.592926 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/35c30afa-953e-4808-b4ba-473ffb44c4a5-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-566bd9db68-99254\" (UID: \"35c30afa-953e-4808-b4ba-473ffb44c4a5\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-99254" Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.658070 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-t8tqd" Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.666084 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-99254" Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.688850 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lh545\" (UniqueName: \"kubernetes.io/projected/6d4f55df-5dc2-4e72-b7f8-cee9229da476-kube-api-access-lh545\") pod \"observability-operator-d8bb48f5d-bfc8s\" (UID: \"6d4f55df-5dc2-4e72-b7f8-cee9229da476\") " pod="openshift-operators/observability-operator-d8bb48f5d-bfc8s" Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.689036 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/6d4f55df-5dc2-4e72-b7f8-cee9229da476-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-bfc8s\" (UID: \"6d4f55df-5dc2-4e72-b7f8-cee9229da476\") " pod="openshift-operators/observability-operator-d8bb48f5d-bfc8s" Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.699119 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/6d4f55df-5dc2-4e72-b7f8-cee9229da476-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-bfc8s\" (UID: \"6d4f55df-5dc2-4e72-b7f8-cee9229da476\") " pod="openshift-operators/observability-operator-d8bb48f5d-bfc8s" Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.709685 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lh545\" (UniqueName: \"kubernetes.io/projected/6d4f55df-5dc2-4e72-b7f8-cee9229da476-kube-api-access-lh545\") pod \"observability-operator-d8bb48f5d-bfc8s\" (UID: \"6d4f55df-5dc2-4e72-b7f8-cee9229da476\") " pod="openshift-operators/observability-operator-d8bb48f5d-bfc8s" Dec 06 06:05:23 crc kubenswrapper[4809]: E1206 06:05:23.714109 4809 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-566bd9db68-t8tqd_openshift-operators_51bf2206-9c20-4c9a-8dbd-1f74c0033236_0(5d1606ccebde7de094e78f236010713bd34458d82bd36801359cd68eeadce4f9): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 06 06:05:23 crc kubenswrapper[4809]: E1206 06:05:23.714183 4809 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-566bd9db68-t8tqd_openshift-operators_51bf2206-9c20-4c9a-8dbd-1f74c0033236_0(5d1606ccebde7de094e78f236010713bd34458d82bd36801359cd68eeadce4f9): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-t8tqd" Dec 06 06:05:23 crc kubenswrapper[4809]: E1206 06:05:23.714214 4809 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-566bd9db68-t8tqd_openshift-operators_51bf2206-9c20-4c9a-8dbd-1f74c0033236_0(5d1606ccebde7de094e78f236010713bd34458d82bd36801359cd68eeadce4f9): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-t8tqd" Dec 06 06:05:23 crc kubenswrapper[4809]: E1206 06:05:23.714269 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-566bd9db68-t8tqd_openshift-operators(51bf2206-9c20-4c9a-8dbd-1f74c0033236)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-566bd9db68-t8tqd_openshift-operators(51bf2206-9c20-4c9a-8dbd-1f74c0033236)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-566bd9db68-t8tqd_openshift-operators_51bf2206-9c20-4c9a-8dbd-1f74c0033236_0(5d1606ccebde7de094e78f236010713bd34458d82bd36801359cd68eeadce4f9): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-t8tqd" podUID="51bf2206-9c20-4c9a-8dbd-1f74c0033236" Dec 06 06:05:23 crc kubenswrapper[4809]: E1206 06:05:23.714966 4809 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-566bd9db68-99254_openshift-operators_35c30afa-953e-4808-b4ba-473ffb44c4a5_0(cdcecaceb0b12d7af71f18564d025daf4d5b4a33830a1c160c6ad3290145c05e): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 06 06:05:23 crc kubenswrapper[4809]: E1206 06:05:23.715002 4809 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-566bd9db68-99254_openshift-operators_35c30afa-953e-4808-b4ba-473ffb44c4a5_0(cdcecaceb0b12d7af71f18564d025daf4d5b4a33830a1c160c6ad3290145c05e): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-99254" Dec 06 06:05:23 crc kubenswrapper[4809]: E1206 06:05:23.715022 4809 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-566bd9db68-99254_openshift-operators_35c30afa-953e-4808-b4ba-473ffb44c4a5_0(cdcecaceb0b12d7af71f18564d025daf4d5b4a33830a1c160c6ad3290145c05e): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-99254" Dec 06 06:05:23 crc kubenswrapper[4809]: E1206 06:05:23.715054 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-566bd9db68-99254_openshift-operators(35c30afa-953e-4808-b4ba-473ffb44c4a5)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-566bd9db68-99254_openshift-operators(35c30afa-953e-4808-b4ba-473ffb44c4a5)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-566bd9db68-99254_openshift-operators_35c30afa-953e-4808-b4ba-473ffb44c4a5_0(cdcecaceb0b12d7af71f18564d025daf4d5b4a33830a1c160c6ad3290145c05e): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-99254" podUID="35c30afa-953e-4808-b4ba-473ffb44c4a5" Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.719058 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5446b9c989-fh2sn"] Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.719955 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-fh2sn" Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.722296 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-7x28d" Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.805234 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j88qv"] Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.814663 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-bfc8s" Dec 06 06:05:23 crc kubenswrapper[4809]: E1206 06:05:23.853171 4809 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-bfc8s_openshift-operators_6d4f55df-5dc2-4e72-b7f8-cee9229da476_0(45ccc8f542246ca5de930d92523c37b8e2f219c0f4350b56d7d038cf204e777a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 06 06:05:23 crc kubenswrapper[4809]: E1206 06:05:23.853237 4809 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-bfc8s_openshift-operators_6d4f55df-5dc2-4e72-b7f8-cee9229da476_0(45ccc8f542246ca5de930d92523c37b8e2f219c0f4350b56d7d038cf204e777a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-bfc8s" Dec 06 06:05:23 crc kubenswrapper[4809]: E1206 06:05:23.853263 4809 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-bfc8s_openshift-operators_6d4f55df-5dc2-4e72-b7f8-cee9229da476_0(45ccc8f542246ca5de930d92523c37b8e2f219c0f4350b56d7d038cf204e777a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-bfc8s" Dec 06 06:05:23 crc kubenswrapper[4809]: E1206 06:05:23.853309 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-d8bb48f5d-bfc8s_openshift-operators(6d4f55df-5dc2-4e72-b7f8-cee9229da476)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-d8bb48f5d-bfc8s_openshift-operators(6d4f55df-5dc2-4e72-b7f8-cee9229da476)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-bfc8s_openshift-operators_6d4f55df-5dc2-4e72-b7f8-cee9229da476_0(45ccc8f542246ca5de930d92523c37b8e2f219c0f4350b56d7d038cf204e777a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-d8bb48f5d-bfc8s" podUID="6d4f55df-5dc2-4e72-b7f8-cee9229da476" Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.892613 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/bb9f4b83-1aa5-48ec-ac64-2d4c7835c5af-openshift-service-ca\") pod \"perses-operator-5446b9c989-fh2sn\" (UID: \"bb9f4b83-1aa5-48ec-ac64-2d4c7835c5af\") " pod="openshift-operators/perses-operator-5446b9c989-fh2sn" Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.892659 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lcnc\" (UniqueName: \"kubernetes.io/projected/bb9f4b83-1aa5-48ec-ac64-2d4c7835c5af-kube-api-access-6lcnc\") pod \"perses-operator-5446b9c989-fh2sn\" (UID: \"bb9f4b83-1aa5-48ec-ac64-2d4c7835c5af\") " pod="openshift-operators/perses-operator-5446b9c989-fh2sn" Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.993855 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/bb9f4b83-1aa5-48ec-ac64-2d4c7835c5af-openshift-service-ca\") pod \"perses-operator-5446b9c989-fh2sn\" (UID: \"bb9f4b83-1aa5-48ec-ac64-2d4c7835c5af\") " pod="openshift-operators/perses-operator-5446b9c989-fh2sn" Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.993896 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lcnc\" (UniqueName: \"kubernetes.io/projected/bb9f4b83-1aa5-48ec-ac64-2d4c7835c5af-kube-api-access-6lcnc\") pod \"perses-operator-5446b9c989-fh2sn\" (UID: \"bb9f4b83-1aa5-48ec-ac64-2d4c7835c5af\") " pod="openshift-operators/perses-operator-5446b9c989-fh2sn" Dec 06 06:05:23 crc kubenswrapper[4809]: I1206 06:05:23.994832 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/bb9f4b83-1aa5-48ec-ac64-2d4c7835c5af-openshift-service-ca\") pod \"perses-operator-5446b9c989-fh2sn\" (UID: \"bb9f4b83-1aa5-48ec-ac64-2d4c7835c5af\") " pod="openshift-operators/perses-operator-5446b9c989-fh2sn" Dec 06 06:05:24 crc kubenswrapper[4809]: I1206 06:05:24.012661 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lcnc\" (UniqueName: \"kubernetes.io/projected/bb9f4b83-1aa5-48ec-ac64-2d4c7835c5af-kube-api-access-6lcnc\") pod \"perses-operator-5446b9c989-fh2sn\" (UID: \"bb9f4b83-1aa5-48ec-ac64-2d4c7835c5af\") " pod="openshift-operators/perses-operator-5446b9c989-fh2sn" Dec 06 06:05:24 crc kubenswrapper[4809]: I1206 06:05:24.062339 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-fh2sn" Dec 06 06:05:24 crc kubenswrapper[4809]: E1206 06:05:24.084560 4809 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-fh2sn_openshift-operators_bb9f4b83-1aa5-48ec-ac64-2d4c7835c5af_0(e2f4eff8e9511ae660328b32f9f23185d968ed1fbed0ac0b0bb9d75207dbedfd): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 06 06:05:24 crc kubenswrapper[4809]: E1206 06:05:24.084627 4809 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-fh2sn_openshift-operators_bb9f4b83-1aa5-48ec-ac64-2d4c7835c5af_0(e2f4eff8e9511ae660328b32f9f23185d968ed1fbed0ac0b0bb9d75207dbedfd): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-fh2sn" Dec 06 06:05:24 crc kubenswrapper[4809]: E1206 06:05:24.084653 4809 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-fh2sn_openshift-operators_bb9f4b83-1aa5-48ec-ac64-2d4c7835c5af_0(e2f4eff8e9511ae660328b32f9f23185d968ed1fbed0ac0b0bb9d75207dbedfd): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-fh2sn" Dec 06 06:05:24 crc kubenswrapper[4809]: E1206 06:05:24.084693 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5446b9c989-fh2sn_openshift-operators(bb9f4b83-1aa5-48ec-ac64-2d4c7835c5af)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5446b9c989-fh2sn_openshift-operators(bb9f4b83-1aa5-48ec-ac64-2d4c7835c5af)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-fh2sn_openshift-operators_bb9f4b83-1aa5-48ec-ac64-2d4c7835c5af_0(e2f4eff8e9511ae660328b32f9f23185d968ed1fbed0ac0b0bb9d75207dbedfd): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5446b9c989-fh2sn" podUID="bb9f4b83-1aa5-48ec-ac64-2d4c7835c5af" Dec 06 06:05:24 crc kubenswrapper[4809]: I1206 06:05:24.264408 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-j88qv" podUID="3c5e2355-685f-45fe-98d4-ee0e0406075d" containerName="registry-server" containerID="cri-o://d86f9164d10d620c990222a1d697eef8646f314e472220d3c0405add4a8eb3fe" gracePeriod=2 Dec 06 06:05:24 crc kubenswrapper[4809]: I1206 06:05:24.313641 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" podStartSLOduration=11.313621652 podStartE2EDuration="11.313621652s" podCreationTimestamp="2025-12-06 06:05:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:05:24.311105614 +0000 UTC m=+849.200088576" watchObservedRunningTime="2025-12-06 06:05:24.313621652 +0000 UTC m=+849.202604594" Dec 06 06:05:24 crc kubenswrapper[4809]: I1206 06:05:24.591332 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-h7k8w"] Dec 06 06:05:24 crc kubenswrapper[4809]: I1206 06:05:24.591730 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-h7k8w" Dec 06 06:05:24 crc kubenswrapper[4809]: I1206 06:05:24.592228 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-h7k8w" Dec 06 06:05:24 crc kubenswrapper[4809]: I1206 06:05:24.610225 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-99254"] Dec 06 06:05:24 crc kubenswrapper[4809]: I1206 06:05:24.610315 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-99254" Dec 06 06:05:24 crc kubenswrapper[4809]: I1206 06:05:24.610681 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-99254" Dec 06 06:05:24 crc kubenswrapper[4809]: E1206 06:05:24.618866 4809 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-h7k8w_openshift-operators_c034c52b-dd97-4722-84fe-b3389a008fbd_0(cc5d69f2cab057de264469feac0dcb2af0de2a7db37db6d36f23521f99ff94a3): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 06 06:05:24 crc kubenswrapper[4809]: E1206 06:05:24.618921 4809 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-h7k8w_openshift-operators_c034c52b-dd97-4722-84fe-b3389a008fbd_0(cc5d69f2cab057de264469feac0dcb2af0de2a7db37db6d36f23521f99ff94a3): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-h7k8w" Dec 06 06:05:24 crc kubenswrapper[4809]: E1206 06:05:24.618957 4809 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-h7k8w_openshift-operators_c034c52b-dd97-4722-84fe-b3389a008fbd_0(cc5d69f2cab057de264469feac0dcb2af0de2a7db37db6d36f23521f99ff94a3): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-h7k8w" Dec 06 06:05:24 crc kubenswrapper[4809]: E1206 06:05:24.618999 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-668cf9dfbb-h7k8w_openshift-operators(c034c52b-dd97-4722-84fe-b3389a008fbd)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-668cf9dfbb-h7k8w_openshift-operators(c034c52b-dd97-4722-84fe-b3389a008fbd)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-h7k8w_openshift-operators_c034c52b-dd97-4722-84fe-b3389a008fbd_0(cc5d69f2cab057de264469feac0dcb2af0de2a7db37db6d36f23521f99ff94a3): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-h7k8w" podUID="c034c52b-dd97-4722-84fe-b3389a008fbd" Dec 06 06:05:24 crc kubenswrapper[4809]: E1206 06:05:24.641611 4809 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-566bd9db68-99254_openshift-operators_35c30afa-953e-4808-b4ba-473ffb44c4a5_0(78124c8f4b7cec7544545d3ca6acc4b5b1f91b3ea0ab6a7d1bf60b5dba3373af): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 06 06:05:24 crc kubenswrapper[4809]: E1206 06:05:24.641684 4809 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-566bd9db68-99254_openshift-operators_35c30afa-953e-4808-b4ba-473ffb44c4a5_0(78124c8f4b7cec7544545d3ca6acc4b5b1f91b3ea0ab6a7d1bf60b5dba3373af): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-99254" Dec 06 06:05:24 crc kubenswrapper[4809]: E1206 06:05:24.641706 4809 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-566bd9db68-99254_openshift-operators_35c30afa-953e-4808-b4ba-473ffb44c4a5_0(78124c8f4b7cec7544545d3ca6acc4b5b1f91b3ea0ab6a7d1bf60b5dba3373af): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-99254" Dec 06 06:05:24 crc kubenswrapper[4809]: E1206 06:05:24.641750 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-566bd9db68-99254_openshift-operators(35c30afa-953e-4808-b4ba-473ffb44c4a5)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-566bd9db68-99254_openshift-operators(35c30afa-953e-4808-b4ba-473ffb44c4a5)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-566bd9db68-99254_openshift-operators_35c30afa-953e-4808-b4ba-473ffb44c4a5_0(78124c8f4b7cec7544545d3ca6acc4b5b1f91b3ea0ab6a7d1bf60b5dba3373af): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-99254" podUID="35c30afa-953e-4808-b4ba-473ffb44c4a5" Dec 06 06:05:24 crc kubenswrapper[4809]: I1206 06:05:24.665334 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-t8tqd"] Dec 06 06:05:24 crc kubenswrapper[4809]: I1206 06:05:24.665459 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-t8tqd" Dec 06 06:05:24 crc kubenswrapper[4809]: I1206 06:05:24.665927 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-t8tqd" Dec 06 06:05:24 crc kubenswrapper[4809]: I1206 06:05:24.688257 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-fh2sn"] Dec 06 06:05:24 crc kubenswrapper[4809]: I1206 06:05:24.688663 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-fh2sn" Dec 06 06:05:24 crc kubenswrapper[4809]: I1206 06:05:24.689295 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-fh2sn" Dec 06 06:05:24 crc kubenswrapper[4809]: I1206 06:05:24.708994 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-bfc8s"] Dec 06 06:05:24 crc kubenswrapper[4809]: I1206 06:05:24.709116 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-bfc8s" Dec 06 06:05:24 crc kubenswrapper[4809]: I1206 06:05:24.709488 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-bfc8s" Dec 06 06:05:24 crc kubenswrapper[4809]: E1206 06:05:24.743446 4809 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-566bd9db68-t8tqd_openshift-operators_51bf2206-9c20-4c9a-8dbd-1f74c0033236_0(1cf14dfa3aa09e697a8129623b3098013e75715de0ba72af7d3980dac7b2495b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 06 06:05:24 crc kubenswrapper[4809]: E1206 06:05:24.743521 4809 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-566bd9db68-t8tqd_openshift-operators_51bf2206-9c20-4c9a-8dbd-1f74c0033236_0(1cf14dfa3aa09e697a8129623b3098013e75715de0ba72af7d3980dac7b2495b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-t8tqd" Dec 06 06:05:24 crc kubenswrapper[4809]: E1206 06:05:24.743548 4809 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-566bd9db68-t8tqd_openshift-operators_51bf2206-9c20-4c9a-8dbd-1f74c0033236_0(1cf14dfa3aa09e697a8129623b3098013e75715de0ba72af7d3980dac7b2495b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-t8tqd" Dec 06 06:05:24 crc kubenswrapper[4809]: E1206 06:05:24.743601 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-566bd9db68-t8tqd_openshift-operators(51bf2206-9c20-4c9a-8dbd-1f74c0033236)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-566bd9db68-t8tqd_openshift-operators(51bf2206-9c20-4c9a-8dbd-1f74c0033236)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-566bd9db68-t8tqd_openshift-operators_51bf2206-9c20-4c9a-8dbd-1f74c0033236_0(1cf14dfa3aa09e697a8129623b3098013e75715de0ba72af7d3980dac7b2495b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-t8tqd" podUID="51bf2206-9c20-4c9a-8dbd-1f74c0033236" Dec 06 06:05:24 crc kubenswrapper[4809]: E1206 06:05:24.783285 4809 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-fh2sn_openshift-operators_bb9f4b83-1aa5-48ec-ac64-2d4c7835c5af_0(fd796642c1ff4de4cf9da882ac9d776fc36c5b80964d6a55eb756783fb4e3df1): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 06 06:05:24 crc kubenswrapper[4809]: E1206 06:05:24.783351 4809 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-fh2sn_openshift-operators_bb9f4b83-1aa5-48ec-ac64-2d4c7835c5af_0(fd796642c1ff4de4cf9da882ac9d776fc36c5b80964d6a55eb756783fb4e3df1): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-fh2sn" Dec 06 06:05:24 crc kubenswrapper[4809]: E1206 06:05:24.783375 4809 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-fh2sn_openshift-operators_bb9f4b83-1aa5-48ec-ac64-2d4c7835c5af_0(fd796642c1ff4de4cf9da882ac9d776fc36c5b80964d6a55eb756783fb4e3df1): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-fh2sn" Dec 06 06:05:24 crc kubenswrapper[4809]: E1206 06:05:24.783423 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5446b9c989-fh2sn_openshift-operators(bb9f4b83-1aa5-48ec-ac64-2d4c7835c5af)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5446b9c989-fh2sn_openshift-operators(bb9f4b83-1aa5-48ec-ac64-2d4c7835c5af)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-fh2sn_openshift-operators_bb9f4b83-1aa5-48ec-ac64-2d4c7835c5af_0(fd796642c1ff4de4cf9da882ac9d776fc36c5b80964d6a55eb756783fb4e3df1): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5446b9c989-fh2sn" podUID="bb9f4b83-1aa5-48ec-ac64-2d4c7835c5af" Dec 06 06:05:24 crc kubenswrapper[4809]: E1206 06:05:24.800871 4809 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-bfc8s_openshift-operators_6d4f55df-5dc2-4e72-b7f8-cee9229da476_0(e66513678dc2203078503420f57fe9d0d90878584c89d044f241d050c7363b6d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 06 06:05:24 crc kubenswrapper[4809]: E1206 06:05:24.800978 4809 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-bfc8s_openshift-operators_6d4f55df-5dc2-4e72-b7f8-cee9229da476_0(e66513678dc2203078503420f57fe9d0d90878584c89d044f241d050c7363b6d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-bfc8s" Dec 06 06:05:24 crc kubenswrapper[4809]: E1206 06:05:24.801011 4809 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-bfc8s_openshift-operators_6d4f55df-5dc2-4e72-b7f8-cee9229da476_0(e66513678dc2203078503420f57fe9d0d90878584c89d044f241d050c7363b6d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-bfc8s" Dec 06 06:05:24 crc kubenswrapper[4809]: E1206 06:05:24.801059 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-d8bb48f5d-bfc8s_openshift-operators(6d4f55df-5dc2-4e72-b7f8-cee9229da476)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-d8bb48f5d-bfc8s_openshift-operators(6d4f55df-5dc2-4e72-b7f8-cee9229da476)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-bfc8s_openshift-operators_6d4f55df-5dc2-4e72-b7f8-cee9229da476_0(e66513678dc2203078503420f57fe9d0d90878584c89d044f241d050c7363b6d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-d8bb48f5d-bfc8s" podUID="6d4f55df-5dc2-4e72-b7f8-cee9229da476" Dec 06 06:05:26 crc kubenswrapper[4809]: I1206 06:05:26.278603 4809 generic.go:334] "Generic (PLEG): container finished" podID="3c5e2355-685f-45fe-98d4-ee0e0406075d" containerID="d86f9164d10d620c990222a1d697eef8646f314e472220d3c0405add4a8eb3fe" exitCode=0 Dec 06 06:05:26 crc kubenswrapper[4809]: I1206 06:05:26.279022 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j88qv" event={"ID":"3c5e2355-685f-45fe-98d4-ee0e0406075d","Type":"ContainerDied","Data":"d86f9164d10d620c990222a1d697eef8646f314e472220d3c0405add4a8eb3fe"} Dec 06 06:05:26 crc kubenswrapper[4809]: I1206 06:05:26.279056 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j88qv" event={"ID":"3c5e2355-685f-45fe-98d4-ee0e0406075d","Type":"ContainerDied","Data":"a1b7e312ad90ffbadb566f9fdcdd9eebe5f2803a5e7d2932b68c5bb3f6fd7781"} Dec 06 06:05:26 crc kubenswrapper[4809]: I1206 06:05:26.279071 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a1b7e312ad90ffbadb566f9fdcdd9eebe5f2803a5e7d2932b68c5bb3f6fd7781" Dec 06 06:05:26 crc kubenswrapper[4809]: I1206 06:05:26.305514 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j88qv" Dec 06 06:05:26 crc kubenswrapper[4809]: I1206 06:05:26.433160 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qqjk\" (UniqueName: \"kubernetes.io/projected/3c5e2355-685f-45fe-98d4-ee0e0406075d-kube-api-access-9qqjk\") pod \"3c5e2355-685f-45fe-98d4-ee0e0406075d\" (UID: \"3c5e2355-685f-45fe-98d4-ee0e0406075d\") " Dec 06 06:05:26 crc kubenswrapper[4809]: I1206 06:05:26.433252 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c5e2355-685f-45fe-98d4-ee0e0406075d-catalog-content\") pod \"3c5e2355-685f-45fe-98d4-ee0e0406075d\" (UID: \"3c5e2355-685f-45fe-98d4-ee0e0406075d\") " Dec 06 06:05:26 crc kubenswrapper[4809]: I1206 06:05:26.433296 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c5e2355-685f-45fe-98d4-ee0e0406075d-utilities\") pod \"3c5e2355-685f-45fe-98d4-ee0e0406075d\" (UID: \"3c5e2355-685f-45fe-98d4-ee0e0406075d\") " Dec 06 06:05:26 crc kubenswrapper[4809]: I1206 06:05:26.434240 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c5e2355-685f-45fe-98d4-ee0e0406075d-utilities" (OuterVolumeSpecName: "utilities") pod "3c5e2355-685f-45fe-98d4-ee0e0406075d" (UID: "3c5e2355-685f-45fe-98d4-ee0e0406075d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:05:26 crc kubenswrapper[4809]: I1206 06:05:26.441286 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c5e2355-685f-45fe-98d4-ee0e0406075d-kube-api-access-9qqjk" (OuterVolumeSpecName: "kube-api-access-9qqjk") pod "3c5e2355-685f-45fe-98d4-ee0e0406075d" (UID: "3c5e2355-685f-45fe-98d4-ee0e0406075d"). InnerVolumeSpecName "kube-api-access-9qqjk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:05:26 crc kubenswrapper[4809]: I1206 06:05:26.536301 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qqjk\" (UniqueName: \"kubernetes.io/projected/3c5e2355-685f-45fe-98d4-ee0e0406075d-kube-api-access-9qqjk\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:26 crc kubenswrapper[4809]: I1206 06:05:26.536347 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c5e2355-685f-45fe-98d4-ee0e0406075d-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:26 crc kubenswrapper[4809]: I1206 06:05:26.538915 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c5e2355-685f-45fe-98d4-ee0e0406075d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3c5e2355-685f-45fe-98d4-ee0e0406075d" (UID: "3c5e2355-685f-45fe-98d4-ee0e0406075d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:05:26 crc kubenswrapper[4809]: I1206 06:05:26.637948 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c5e2355-685f-45fe-98d4-ee0e0406075d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:27 crc kubenswrapper[4809]: I1206 06:05:27.284160 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j88qv" Dec 06 06:05:27 crc kubenswrapper[4809]: I1206 06:05:27.324651 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j88qv"] Dec 06 06:05:27 crc kubenswrapper[4809]: I1206 06:05:27.331908 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-j88qv"] Dec 06 06:05:27 crc kubenswrapper[4809]: I1206 06:05:27.398402 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c5e2355-685f-45fe-98d4-ee0e0406075d" path="/var/lib/kubelet/pods/3c5e2355-685f-45fe-98d4-ee0e0406075d/volumes" Dec 06 06:05:33 crc kubenswrapper[4809]: I1206 06:05:33.728857 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:33 crc kubenswrapper[4809]: I1206 06:05:33.729393 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:33 crc kubenswrapper[4809]: I1206 06:05:33.730246 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:33 crc kubenswrapper[4809]: I1206 06:05:33.752228 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:33 crc kubenswrapper[4809]: I1206 06:05:33.761144 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:33 crc kubenswrapper[4809]: I1206 06:05:33.773082 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" Dec 06 06:05:34 crc kubenswrapper[4809]: I1206 06:05:34.496161 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:05:34 crc kubenswrapper[4809]: I1206 06:05:34.497050 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:05:34 crc kubenswrapper[4809]: I1206 06:05:34.497216 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-npms2" Dec 06 06:05:34 crc kubenswrapper[4809]: I1206 06:05:34.498486 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"49ef8bff82d74abdc9d792d5696488d1189b4c9562bc3eb566b2639019bb757d"} pod="openshift-machine-config-operator/machine-config-daemon-npms2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 06:05:34 crc kubenswrapper[4809]: I1206 06:05:34.498691 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" containerID="cri-o://49ef8bff82d74abdc9d792d5696488d1189b4c9562bc3eb566b2639019bb757d" gracePeriod=600 Dec 06 06:05:37 crc kubenswrapper[4809]: I1206 06:05:37.388000 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-99254" Dec 06 06:05:37 crc kubenswrapper[4809]: I1206 06:05:37.389108 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-99254" Dec 06 06:05:37 crc kubenswrapper[4809]: I1206 06:05:37.983800 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-99254"] Dec 06 06:05:38 crc kubenswrapper[4809]: I1206 06:05:38.379163 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-99254" event={"ID":"35c30afa-953e-4808-b4ba-473ffb44c4a5","Type":"ContainerStarted","Data":"21cc0e7ecfcf78dee94f6806efeb8a55c3a92c7cc7e2f5ded79f84c0b40d856a"} Dec 06 06:05:38 crc kubenswrapper[4809]: I1206 06:05:38.381660 4809 generic.go:334] "Generic (PLEG): container finished" podID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerID="49ef8bff82d74abdc9d792d5696488d1189b4c9562bc3eb566b2639019bb757d" exitCode=0 Dec 06 06:05:38 crc kubenswrapper[4809]: I1206 06:05:38.381712 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" event={"ID":"cbf4cf62-024e-4703-a8b8-9aecda9cd26a","Type":"ContainerDied","Data":"49ef8bff82d74abdc9d792d5696488d1189b4c9562bc3eb566b2639019bb757d"} Dec 06 06:05:38 crc kubenswrapper[4809]: I1206 06:05:38.381771 4809 scope.go:117] "RemoveContainer" containerID="e6b027954a4e3b35818e27fb0ec5b9b81b52eb65e6a41e8dc2459f7e54dc0c30" Dec 06 06:05:38 crc kubenswrapper[4809]: I1206 06:05:38.388666 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-bfc8s" Dec 06 06:05:38 crc kubenswrapper[4809]: I1206 06:05:38.388717 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-t8tqd" Dec 06 06:05:38 crc kubenswrapper[4809]: I1206 06:05:38.388878 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-h7k8w" Dec 06 06:05:38 crc kubenswrapper[4809]: I1206 06:05:38.389709 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-bfc8s" Dec 06 06:05:38 crc kubenswrapper[4809]: I1206 06:05:38.389718 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-h7k8w" Dec 06 06:05:38 crc kubenswrapper[4809]: I1206 06:05:38.390021 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-t8tqd" Dec 06 06:05:38 crc kubenswrapper[4809]: I1206 06:05:38.787676 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-h7k8w"] Dec 06 06:05:38 crc kubenswrapper[4809]: W1206 06:05:38.798827 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc034c52b_dd97_4722_84fe_b3389a008fbd.slice/crio-f6f97d42b0e2951bd0e5f176cc13bc2574de419460774be6fdc52ba65c4fe4e7 WatchSource:0}: Error finding container f6f97d42b0e2951bd0e5f176cc13bc2574de419460774be6fdc52ba65c4fe4e7: Status 404 returned error can't find the container with id f6f97d42b0e2951bd0e5f176cc13bc2574de419460774be6fdc52ba65c4fe4e7 Dec 06 06:05:38 crc kubenswrapper[4809]: I1206 06:05:38.915179 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-t8tqd"] Dec 06 06:05:38 crc kubenswrapper[4809]: I1206 06:05:38.919536 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-bfc8s"] Dec 06 06:05:38 crc kubenswrapper[4809]: W1206 06:05:38.934405 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod51bf2206_9c20_4c9a_8dbd_1f74c0033236.slice/crio-5351c4f335e0e4ac58252af59a14cef95815a196c95177663b86df17e7eb32af WatchSource:0}: Error finding container 5351c4f335e0e4ac58252af59a14cef95815a196c95177663b86df17e7eb32af: Status 404 returned error can't find the container with id 5351c4f335e0e4ac58252af59a14cef95815a196c95177663b86df17e7eb32af Dec 06 06:05:39 crc kubenswrapper[4809]: I1206 06:05:39.387557 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-fh2sn" Dec 06 06:05:39 crc kubenswrapper[4809]: I1206 06:05:39.388451 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-fh2sn" Dec 06 06:05:39 crc kubenswrapper[4809]: I1206 06:05:39.393899 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-t8tqd" event={"ID":"51bf2206-9c20-4c9a-8dbd-1f74c0033236","Type":"ContainerStarted","Data":"5351c4f335e0e4ac58252af59a14cef95815a196c95177663b86df17e7eb32af"} Dec 06 06:05:39 crc kubenswrapper[4809]: I1206 06:05:39.393952 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-bfc8s" event={"ID":"6d4f55df-5dc2-4e72-b7f8-cee9229da476","Type":"ContainerStarted","Data":"93e347ac5e4789b4ddd484f6c63d3193d8ea12a1f1319015648d1bd2c19d7e6a"} Dec 06 06:05:39 crc kubenswrapper[4809]: I1206 06:05:39.393965 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-h7k8w" event={"ID":"c034c52b-dd97-4722-84fe-b3389a008fbd","Type":"ContainerStarted","Data":"f6f97d42b0e2951bd0e5f176cc13bc2574de419460774be6fdc52ba65c4fe4e7"} Dec 06 06:05:39 crc kubenswrapper[4809]: I1206 06:05:39.396374 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" event={"ID":"cbf4cf62-024e-4703-a8b8-9aecda9cd26a","Type":"ContainerStarted","Data":"bb59482414fdfbb9268b139562a808d2a5e04bbf56acc96dd715da16e5bd5913"} Dec 06 06:05:39 crc kubenswrapper[4809]: I1206 06:05:39.626972 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-fh2sn"] Dec 06 06:05:39 crc kubenswrapper[4809]: W1206 06:05:39.640539 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb9f4b83_1aa5_48ec_ac64_2d4c7835c5af.slice/crio-4637dd5e53bc0aae668681c71a35ecb4ac35eb5de5758d5b97d577146ec0dcc6 WatchSource:0}: Error finding container 4637dd5e53bc0aae668681c71a35ecb4ac35eb5de5758d5b97d577146ec0dcc6: Status 404 returned error can't find the container with id 4637dd5e53bc0aae668681c71a35ecb4ac35eb5de5758d5b97d577146ec0dcc6 Dec 06 06:05:40 crc kubenswrapper[4809]: I1206 06:05:40.404392 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-fh2sn" event={"ID":"bb9f4b83-1aa5-48ec-ac64-2d4c7835c5af","Type":"ContainerStarted","Data":"4637dd5e53bc0aae668681c71a35ecb4ac35eb5de5758d5b97d577146ec0dcc6"} Dec 06 06:05:52 crc kubenswrapper[4809]: I1206 06:05:52.514105 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-h7k8w" event={"ID":"c034c52b-dd97-4722-84fe-b3389a008fbd","Type":"ContainerStarted","Data":"ce71cd004a055c79461ae51f0bb8cd83440dfcde9f8902222fc6af5a0ffa41c2"} Dec 06 06:05:52 crc kubenswrapper[4809]: I1206 06:05:52.516596 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-fh2sn" event={"ID":"bb9f4b83-1aa5-48ec-ac64-2d4c7835c5af","Type":"ContainerStarted","Data":"382df1099302a09091bc48ab800cdba3edae0c94395bc7251ae4190ca5217dc7"} Dec 06 06:05:52 crc kubenswrapper[4809]: I1206 06:05:52.517099 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5446b9c989-fh2sn" Dec 06 06:05:52 crc kubenswrapper[4809]: I1206 06:05:52.519026 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-99254" event={"ID":"35c30afa-953e-4808-b4ba-473ffb44c4a5","Type":"ContainerStarted","Data":"b4daa8e574498423eaf148c84d700cf11b70c97a59837441102a02a3f731a78c"} Dec 06 06:05:52 crc kubenswrapper[4809]: I1206 06:05:52.520893 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-t8tqd" event={"ID":"51bf2206-9c20-4c9a-8dbd-1f74c0033236","Type":"ContainerStarted","Data":"c9b49ef9216cf2b9fa6f8653172fd417281cf105c33c29d94e4d2b203bd3931e"} Dec 06 06:05:52 crc kubenswrapper[4809]: I1206 06:05:52.536777 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-h7k8w" podStartSLOduration=17.409890222 podStartE2EDuration="29.536756362s" podCreationTimestamp="2025-12-06 06:05:23 +0000 UTC" firstStartedPulling="2025-12-06 06:05:38.80059156 +0000 UTC m=+863.689574502" lastFinishedPulling="2025-12-06 06:05:50.9274577 +0000 UTC m=+875.816440642" observedRunningTime="2025-12-06 06:05:52.531972942 +0000 UTC m=+877.420955894" watchObservedRunningTime="2025-12-06 06:05:52.536756362 +0000 UTC m=+877.425739304" Dec 06 06:05:52 crc kubenswrapper[4809]: I1206 06:05:52.559139 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5446b9c989-fh2sn" podStartSLOduration=18.174523784 podStartE2EDuration="29.559122908s" podCreationTimestamp="2025-12-06 06:05:23 +0000 UTC" firstStartedPulling="2025-12-06 06:05:39.643384459 +0000 UTC m=+864.532367401" lastFinishedPulling="2025-12-06 06:05:51.027983583 +0000 UTC m=+875.916966525" observedRunningTime="2025-12-06 06:05:52.558535391 +0000 UTC m=+877.447518333" watchObservedRunningTime="2025-12-06 06:05:52.559122908 +0000 UTC m=+877.448105840" Dec 06 06:05:52 crc kubenswrapper[4809]: I1206 06:05:52.587405 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-t8tqd" podStartSLOduration=17.513108188 podStartE2EDuration="29.587379963s" podCreationTimestamp="2025-12-06 06:05:23 +0000 UTC" firstStartedPulling="2025-12-06 06:05:38.937454028 +0000 UTC m=+863.826436960" lastFinishedPulling="2025-12-06 06:05:51.011725793 +0000 UTC m=+875.900708735" observedRunningTime="2025-12-06 06:05:52.579705604 +0000 UTC m=+877.468688576" watchObservedRunningTime="2025-12-06 06:05:52.587379963 +0000 UTC m=+877.476362905" Dec 06 06:05:54 crc kubenswrapper[4809]: I1206 06:05:54.537430 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-bfc8s" event={"ID":"6d4f55df-5dc2-4e72-b7f8-cee9229da476","Type":"ContainerStarted","Data":"29e0d9710a020c6208a7684bac57c9ff194335d91be10a73aadd51aa3f289272"} Dec 06 06:05:54 crc kubenswrapper[4809]: I1206 06:05:54.539380 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-d8bb48f5d-bfc8s" Dec 06 06:05:54 crc kubenswrapper[4809]: I1206 06:05:54.556958 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-d8bb48f5d-bfc8s" podStartSLOduration=16.472705636 podStartE2EDuration="31.556923942s" podCreationTimestamp="2025-12-06 06:05:23 +0000 UTC" firstStartedPulling="2025-12-06 06:05:38.928599998 +0000 UTC m=+863.817582940" lastFinishedPulling="2025-12-06 06:05:54.012818304 +0000 UTC m=+878.901801246" observedRunningTime="2025-12-06 06:05:54.554700251 +0000 UTC m=+879.443683203" watchObservedRunningTime="2025-12-06 06:05:54.556923942 +0000 UTC m=+879.445906884" Dec 06 06:05:54 crc kubenswrapper[4809]: I1206 06:05:54.558649 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-566bd9db68-99254" podStartSLOduration=18.612314651 podStartE2EDuration="31.558641438s" podCreationTimestamp="2025-12-06 06:05:23 +0000 UTC" firstStartedPulling="2025-12-06 06:05:37.996750397 +0000 UTC m=+862.885733339" lastFinishedPulling="2025-12-06 06:05:50.943077184 +0000 UTC m=+875.832060126" observedRunningTime="2025-12-06 06:05:52.601451844 +0000 UTC m=+877.490434776" watchObservedRunningTime="2025-12-06 06:05:54.558641438 +0000 UTC m=+879.447624380" Dec 06 06:05:54 crc kubenswrapper[4809]: I1206 06:05:54.591761 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-d8bb48f5d-bfc8s" Dec 06 06:06:04 crc kubenswrapper[4809]: I1206 06:06:04.065185 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5446b9c989-fh2sn" Dec 06 06:06:04 crc kubenswrapper[4809]: I1206 06:06:04.069765 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-s5sjr"] Dec 06 06:06:04 crc kubenswrapper[4809]: E1206 06:06:04.070205 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c5e2355-685f-45fe-98d4-ee0e0406075d" containerName="registry-server" Dec 06 06:06:04 crc kubenswrapper[4809]: I1206 06:06:04.070284 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c5e2355-685f-45fe-98d4-ee0e0406075d" containerName="registry-server" Dec 06 06:06:04 crc kubenswrapper[4809]: E1206 06:06:04.070345 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c5e2355-685f-45fe-98d4-ee0e0406075d" containerName="extract-utilities" Dec 06 06:06:04 crc kubenswrapper[4809]: I1206 06:06:04.070411 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c5e2355-685f-45fe-98d4-ee0e0406075d" containerName="extract-utilities" Dec 06 06:06:04 crc kubenswrapper[4809]: E1206 06:06:04.070488 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c5e2355-685f-45fe-98d4-ee0e0406075d" containerName="extract-content" Dec 06 06:06:04 crc kubenswrapper[4809]: I1206 06:06:04.070547 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c5e2355-685f-45fe-98d4-ee0e0406075d" containerName="extract-content" Dec 06 06:06:04 crc kubenswrapper[4809]: I1206 06:06:04.070740 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c5e2355-685f-45fe-98d4-ee0e0406075d" containerName="registry-server" Dec 06 06:06:04 crc kubenswrapper[4809]: I1206 06:06:04.071302 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-s5sjr" Dec 06 06:06:04 crc kubenswrapper[4809]: I1206 06:06:04.074218 4809 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-ms2r6" Dec 06 06:06:04 crc kubenswrapper[4809]: I1206 06:06:04.074282 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 06 06:06:04 crc kubenswrapper[4809]: I1206 06:06:04.074985 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 06 06:06:04 crc kubenswrapper[4809]: I1206 06:06:04.102310 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-s5sjr"] Dec 06 06:06:04 crc kubenswrapper[4809]: I1206 06:06:04.114615 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-wgc72"] Dec 06 06:06:04 crc kubenswrapper[4809]: I1206 06:06:04.116847 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-wgc72" Dec 06 06:06:04 crc kubenswrapper[4809]: I1206 06:06:04.123794 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-7vvw8"] Dec 06 06:06:04 crc kubenswrapper[4809]: I1206 06:06:04.128320 4809 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-rtfzv" Dec 06 06:06:04 crc kubenswrapper[4809]: I1206 06:06:04.137359 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-7vvw8" Dec 06 06:06:04 crc kubenswrapper[4809]: I1206 06:06:04.144573 4809 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-c5r4d" Dec 06 06:06:04 crc kubenswrapper[4809]: I1206 06:06:04.146078 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xb5tx\" (UniqueName: \"kubernetes.io/projected/79eae299-f38e-4441-954b-833293821894-kube-api-access-xb5tx\") pod \"cert-manager-cainjector-7f985d654d-s5sjr\" (UID: \"79eae299-f38e-4441-954b-833293821894\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-s5sjr" Dec 06 06:06:04 crc kubenswrapper[4809]: I1206 06:06:04.166089 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-wgc72"] Dec 06 06:06:04 crc kubenswrapper[4809]: I1206 06:06:04.174065 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-7vvw8"] Dec 06 06:06:04 crc kubenswrapper[4809]: I1206 06:06:04.247166 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xb5tx\" (UniqueName: \"kubernetes.io/projected/79eae299-f38e-4441-954b-833293821894-kube-api-access-xb5tx\") pod \"cert-manager-cainjector-7f985d654d-s5sjr\" (UID: \"79eae299-f38e-4441-954b-833293821894\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-s5sjr" Dec 06 06:06:04 crc kubenswrapper[4809]: I1206 06:06:04.247224 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5pdw\" (UniqueName: \"kubernetes.io/projected/886a998b-a62e-412c-ae50-fd1ba3a4a098-kube-api-access-v5pdw\") pod \"cert-manager-webhook-5655c58dd6-7vvw8\" (UID: \"886a998b-a62e-412c-ae50-fd1ba3a4a098\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-7vvw8" Dec 06 06:06:04 crc kubenswrapper[4809]: I1206 06:06:04.247308 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcjjp\" (UniqueName: \"kubernetes.io/projected/8bc6fec5-e0e7-416f-b7ac-0a671cac3c0b-kube-api-access-rcjjp\") pod \"cert-manager-5b446d88c5-wgc72\" (UID: \"8bc6fec5-e0e7-416f-b7ac-0a671cac3c0b\") " pod="cert-manager/cert-manager-5b446d88c5-wgc72" Dec 06 06:06:04 crc kubenswrapper[4809]: I1206 06:06:04.269995 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xb5tx\" (UniqueName: \"kubernetes.io/projected/79eae299-f38e-4441-954b-833293821894-kube-api-access-xb5tx\") pod \"cert-manager-cainjector-7f985d654d-s5sjr\" (UID: \"79eae299-f38e-4441-954b-833293821894\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-s5sjr" Dec 06 06:06:04 crc kubenswrapper[4809]: I1206 06:06:04.348396 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcjjp\" (UniqueName: \"kubernetes.io/projected/8bc6fec5-e0e7-416f-b7ac-0a671cac3c0b-kube-api-access-rcjjp\") pod \"cert-manager-5b446d88c5-wgc72\" (UID: \"8bc6fec5-e0e7-416f-b7ac-0a671cac3c0b\") " pod="cert-manager/cert-manager-5b446d88c5-wgc72" Dec 06 06:06:04 crc kubenswrapper[4809]: I1206 06:06:04.348836 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5pdw\" (UniqueName: \"kubernetes.io/projected/886a998b-a62e-412c-ae50-fd1ba3a4a098-kube-api-access-v5pdw\") pod \"cert-manager-webhook-5655c58dd6-7vvw8\" (UID: \"886a998b-a62e-412c-ae50-fd1ba3a4a098\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-7vvw8" Dec 06 06:06:04 crc kubenswrapper[4809]: I1206 06:06:04.366551 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5pdw\" (UniqueName: \"kubernetes.io/projected/886a998b-a62e-412c-ae50-fd1ba3a4a098-kube-api-access-v5pdw\") pod \"cert-manager-webhook-5655c58dd6-7vvw8\" (UID: \"886a998b-a62e-412c-ae50-fd1ba3a4a098\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-7vvw8" Dec 06 06:06:04 crc kubenswrapper[4809]: I1206 06:06:04.367303 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcjjp\" (UniqueName: \"kubernetes.io/projected/8bc6fec5-e0e7-416f-b7ac-0a671cac3c0b-kube-api-access-rcjjp\") pod \"cert-manager-5b446d88c5-wgc72\" (UID: \"8bc6fec5-e0e7-416f-b7ac-0a671cac3c0b\") " pod="cert-manager/cert-manager-5b446d88c5-wgc72" Dec 06 06:06:04 crc kubenswrapper[4809]: I1206 06:06:04.394135 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-s5sjr" Dec 06 06:06:04 crc kubenswrapper[4809]: I1206 06:06:04.441066 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-wgc72" Dec 06 06:06:04 crc kubenswrapper[4809]: I1206 06:06:04.466958 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-7vvw8" Dec 06 06:06:04 crc kubenswrapper[4809]: I1206 06:06:04.863997 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-s5sjr"] Dec 06 06:06:04 crc kubenswrapper[4809]: I1206 06:06:04.929452 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-wgc72"] Dec 06 06:06:04 crc kubenswrapper[4809]: W1206 06:06:04.936462 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8bc6fec5_e0e7_416f_b7ac_0a671cac3c0b.slice/crio-0ded1aafcdb38771082f82ffd70b28f8bf2d23c0543f29ce245d4b82438ba085 WatchSource:0}: Error finding container 0ded1aafcdb38771082f82ffd70b28f8bf2d23c0543f29ce245d4b82438ba085: Status 404 returned error can't find the container with id 0ded1aafcdb38771082f82ffd70b28f8bf2d23c0543f29ce245d4b82438ba085 Dec 06 06:06:05 crc kubenswrapper[4809]: I1206 06:06:05.186151 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-7vvw8"] Dec 06 06:06:05 crc kubenswrapper[4809]: W1206 06:06:05.194329 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod886a998b_a62e_412c_ae50_fd1ba3a4a098.slice/crio-67b2c581a675c35f6eee186e4dd8a794c5d9962041ad201b153567f5d8c25e65 WatchSource:0}: Error finding container 67b2c581a675c35f6eee186e4dd8a794c5d9962041ad201b153567f5d8c25e65: Status 404 returned error can't find the container with id 67b2c581a675c35f6eee186e4dd8a794c5d9962041ad201b153567f5d8c25e65 Dec 06 06:06:05 crc kubenswrapper[4809]: I1206 06:06:05.636165 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-s5sjr" event={"ID":"79eae299-f38e-4441-954b-833293821894","Type":"ContainerStarted","Data":"b0dc830fb1778ef22e0000b73550ec854a3eca0b96fbfc6852ffe051e9a8ad28"} Dec 06 06:06:05 crc kubenswrapper[4809]: I1206 06:06:05.637775 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-wgc72" event={"ID":"8bc6fec5-e0e7-416f-b7ac-0a671cac3c0b","Type":"ContainerStarted","Data":"0ded1aafcdb38771082f82ffd70b28f8bf2d23c0543f29ce245d4b82438ba085"} Dec 06 06:06:05 crc kubenswrapper[4809]: I1206 06:06:05.638750 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-7vvw8" event={"ID":"886a998b-a62e-412c-ae50-fd1ba3a4a098","Type":"ContainerStarted","Data":"67b2c581a675c35f6eee186e4dd8a794c5d9962041ad201b153567f5d8c25e65"} Dec 06 06:06:08 crc kubenswrapper[4809]: I1206 06:06:08.687146 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-s5sjr" event={"ID":"79eae299-f38e-4441-954b-833293821894","Type":"ContainerStarted","Data":"a50e9b04ebcc9599766c21bffce62f126a566386adc3a727cf70a8c4d0d6f63d"} Dec 06 06:06:14 crc kubenswrapper[4809]: I1206 06:06:14.733878 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-7vvw8" event={"ID":"886a998b-a62e-412c-ae50-fd1ba3a4a098","Type":"ContainerStarted","Data":"caa5249f1c766f453ea37b5bdf161e48771bfe1e4908e8b92d09d672ec0baaef"} Dec 06 06:06:14 crc kubenswrapper[4809]: I1206 06:06:14.734365 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-7vvw8" Dec 06 06:06:14 crc kubenswrapper[4809]: I1206 06:06:14.735482 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-wgc72" event={"ID":"8bc6fec5-e0e7-416f-b7ac-0a671cac3c0b","Type":"ContainerStarted","Data":"17ef2f9fe54d58797256797c9fc796d48f0a00c211940016e1c40963010132f8"} Dec 06 06:06:14 crc kubenswrapper[4809]: I1206 06:06:14.748298 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-7vvw8" podStartSLOduration=1.576663393 podStartE2EDuration="10.748278015s" podCreationTimestamp="2025-12-06 06:06:04 +0000 UTC" firstStartedPulling="2025-12-06 06:06:05.196497166 +0000 UTC m=+890.085480108" lastFinishedPulling="2025-12-06 06:06:14.368111788 +0000 UTC m=+899.257094730" observedRunningTime="2025-12-06 06:06:14.74694848 +0000 UTC m=+899.635931422" watchObservedRunningTime="2025-12-06 06:06:14.748278015 +0000 UTC m=+899.637260957" Dec 06 06:06:14 crc kubenswrapper[4809]: I1206 06:06:14.751457 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-s5sjr" podStartSLOduration=7.5446648 podStartE2EDuration="10.751436771s" podCreationTimestamp="2025-12-06 06:06:04 +0000 UTC" firstStartedPulling="2025-12-06 06:06:04.870736623 +0000 UTC m=+889.759719555" lastFinishedPulling="2025-12-06 06:06:08.077508584 +0000 UTC m=+892.966491526" observedRunningTime="2025-12-06 06:06:08.717699205 +0000 UTC m=+893.606682167" watchObservedRunningTime="2025-12-06 06:06:14.751436771 +0000 UTC m=+899.640419713" Dec 06 06:06:19 crc kubenswrapper[4809]: I1206 06:06:19.470169 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-7vvw8" Dec 06 06:06:19 crc kubenswrapper[4809]: I1206 06:06:19.494597 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-wgc72" podStartSLOduration=6.059665336 podStartE2EDuration="15.494578999s" podCreationTimestamp="2025-12-06 06:06:04 +0000 UTC" firstStartedPulling="2025-12-06 06:06:04.939356411 +0000 UTC m=+889.828339353" lastFinishedPulling="2025-12-06 06:06:14.374270074 +0000 UTC m=+899.263253016" observedRunningTime="2025-12-06 06:06:14.763549299 +0000 UTC m=+899.652532251" watchObservedRunningTime="2025-12-06 06:06:19.494578999 +0000 UTC m=+904.383561941" Dec 06 06:06:48 crc kubenswrapper[4809]: I1206 06:06:48.056404 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpd5xh"] Dec 06 06:06:48 crc kubenswrapper[4809]: I1206 06:06:48.058688 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpd5xh" Dec 06 06:06:48 crc kubenswrapper[4809]: I1206 06:06:48.063853 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 06 06:06:48 crc kubenswrapper[4809]: I1206 06:06:48.066959 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpd5xh"] Dec 06 06:06:48 crc kubenswrapper[4809]: I1206 06:06:48.075758 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e12ceeab-a1f8-4a5b-aefc-77d7cf057f40-bundle\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpd5xh\" (UID: \"e12ceeab-a1f8-4a5b-aefc-77d7cf057f40\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpd5xh" Dec 06 06:06:48 crc kubenswrapper[4809]: I1206 06:06:48.075834 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e12ceeab-a1f8-4a5b-aefc-77d7cf057f40-util\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpd5xh\" (UID: \"e12ceeab-a1f8-4a5b-aefc-77d7cf057f40\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpd5xh" Dec 06 06:06:48 crc kubenswrapper[4809]: I1206 06:06:48.075889 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5sqd\" (UniqueName: \"kubernetes.io/projected/e12ceeab-a1f8-4a5b-aefc-77d7cf057f40-kube-api-access-s5sqd\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpd5xh\" (UID: \"e12ceeab-a1f8-4a5b-aefc-77d7cf057f40\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpd5xh" Dec 06 06:06:48 crc kubenswrapper[4809]: I1206 06:06:48.176632 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e12ceeab-a1f8-4a5b-aefc-77d7cf057f40-util\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpd5xh\" (UID: \"e12ceeab-a1f8-4a5b-aefc-77d7cf057f40\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpd5xh" Dec 06 06:06:48 crc kubenswrapper[4809]: I1206 06:06:48.176739 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5sqd\" (UniqueName: \"kubernetes.io/projected/e12ceeab-a1f8-4a5b-aefc-77d7cf057f40-kube-api-access-s5sqd\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpd5xh\" (UID: \"e12ceeab-a1f8-4a5b-aefc-77d7cf057f40\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpd5xh" Dec 06 06:06:48 crc kubenswrapper[4809]: I1206 06:06:48.176771 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e12ceeab-a1f8-4a5b-aefc-77d7cf057f40-bundle\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpd5xh\" (UID: \"e12ceeab-a1f8-4a5b-aefc-77d7cf057f40\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpd5xh" Dec 06 06:06:48 crc kubenswrapper[4809]: I1206 06:06:48.177299 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e12ceeab-a1f8-4a5b-aefc-77d7cf057f40-bundle\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpd5xh\" (UID: \"e12ceeab-a1f8-4a5b-aefc-77d7cf057f40\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpd5xh" Dec 06 06:06:48 crc kubenswrapper[4809]: I1206 06:06:48.178109 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e12ceeab-a1f8-4a5b-aefc-77d7cf057f40-util\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpd5xh\" (UID: \"e12ceeab-a1f8-4a5b-aefc-77d7cf057f40\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpd5xh" Dec 06 06:06:48 crc kubenswrapper[4809]: I1206 06:06:48.218123 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5sqd\" (UniqueName: \"kubernetes.io/projected/e12ceeab-a1f8-4a5b-aefc-77d7cf057f40-kube-api-access-s5sqd\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpd5xh\" (UID: \"e12ceeab-a1f8-4a5b-aefc-77d7cf057f40\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpd5xh" Dec 06 06:06:48 crc kubenswrapper[4809]: I1206 06:06:48.291952 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8pfj4q"] Dec 06 06:06:48 crc kubenswrapper[4809]: I1206 06:06:48.293286 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8pfj4q" Dec 06 06:06:48 crc kubenswrapper[4809]: I1206 06:06:48.309879 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8pfj4q"] Dec 06 06:06:48 crc kubenswrapper[4809]: I1206 06:06:48.380133 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpd5xh" Dec 06 06:06:48 crc kubenswrapper[4809]: I1206 06:06:48.479965 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/de7c1cd3-87d5-431f-a971-5474ac3d56ec-bundle\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8pfj4q\" (UID: \"de7c1cd3-87d5-431f-a971-5474ac3d56ec\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8pfj4q" Dec 06 06:06:48 crc kubenswrapper[4809]: I1206 06:06:48.480306 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/de7c1cd3-87d5-431f-a971-5474ac3d56ec-util\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8pfj4q\" (UID: \"de7c1cd3-87d5-431f-a971-5474ac3d56ec\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8pfj4q" Dec 06 06:06:48 crc kubenswrapper[4809]: I1206 06:06:48.480398 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fr749\" (UniqueName: \"kubernetes.io/projected/de7c1cd3-87d5-431f-a971-5474ac3d56ec-kube-api-access-fr749\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8pfj4q\" (UID: \"de7c1cd3-87d5-431f-a971-5474ac3d56ec\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8pfj4q" Dec 06 06:06:48 crc kubenswrapper[4809]: I1206 06:06:48.582404 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fr749\" (UniqueName: \"kubernetes.io/projected/de7c1cd3-87d5-431f-a971-5474ac3d56ec-kube-api-access-fr749\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8pfj4q\" (UID: \"de7c1cd3-87d5-431f-a971-5474ac3d56ec\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8pfj4q" Dec 06 06:06:48 crc kubenswrapper[4809]: I1206 06:06:48.582459 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/de7c1cd3-87d5-431f-a971-5474ac3d56ec-bundle\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8pfj4q\" (UID: \"de7c1cd3-87d5-431f-a971-5474ac3d56ec\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8pfj4q" Dec 06 06:06:48 crc kubenswrapper[4809]: I1206 06:06:48.582524 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/de7c1cd3-87d5-431f-a971-5474ac3d56ec-util\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8pfj4q\" (UID: \"de7c1cd3-87d5-431f-a971-5474ac3d56ec\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8pfj4q" Dec 06 06:06:48 crc kubenswrapper[4809]: I1206 06:06:48.582987 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/de7c1cd3-87d5-431f-a971-5474ac3d56ec-util\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8pfj4q\" (UID: \"de7c1cd3-87d5-431f-a971-5474ac3d56ec\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8pfj4q" Dec 06 06:06:48 crc kubenswrapper[4809]: I1206 06:06:48.583196 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/de7c1cd3-87d5-431f-a971-5474ac3d56ec-bundle\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8pfj4q\" (UID: \"de7c1cd3-87d5-431f-a971-5474ac3d56ec\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8pfj4q" Dec 06 06:06:48 crc kubenswrapper[4809]: I1206 06:06:48.610229 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fr749\" (UniqueName: \"kubernetes.io/projected/de7c1cd3-87d5-431f-a971-5474ac3d56ec-kube-api-access-fr749\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8pfj4q\" (UID: \"de7c1cd3-87d5-431f-a971-5474ac3d56ec\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8pfj4q" Dec 06 06:06:48 crc kubenswrapper[4809]: I1206 06:06:48.610578 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8pfj4q" Dec 06 06:06:48 crc kubenswrapper[4809]: I1206 06:06:48.632546 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpd5xh"] Dec 06 06:06:48 crc kubenswrapper[4809]: I1206 06:06:48.809479 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8pfj4q"] Dec 06 06:06:48 crc kubenswrapper[4809]: W1206 06:06:48.814299 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podde7c1cd3_87d5_431f_a971_5474ac3d56ec.slice/crio-85c5c41e678c9e349f37991b084bc738b0955faaa0d3ac115231c384ebe3c432 WatchSource:0}: Error finding container 85c5c41e678c9e349f37991b084bc738b0955faaa0d3ac115231c384ebe3c432: Status 404 returned error can't find the container with id 85c5c41e678c9e349f37991b084bc738b0955faaa0d3ac115231c384ebe3c432 Dec 06 06:06:48 crc kubenswrapper[4809]: I1206 06:06:48.984527 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8pfj4q" event={"ID":"de7c1cd3-87d5-431f-a971-5474ac3d56ec","Type":"ContainerStarted","Data":"85c5c41e678c9e349f37991b084bc738b0955faaa0d3ac115231c384ebe3c432"} Dec 06 06:06:48 crc kubenswrapper[4809]: I1206 06:06:48.985429 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpd5xh" event={"ID":"e12ceeab-a1f8-4a5b-aefc-77d7cf057f40","Type":"ContainerStarted","Data":"4c890a7cc903b7a1cdd70d2441d42740b095cea90c510ff5a70b9d81676f6a52"} Dec 06 06:06:49 crc kubenswrapper[4809]: I1206 06:06:49.997562 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpd5xh" event={"ID":"e12ceeab-a1f8-4a5b-aefc-77d7cf057f40","Type":"ContainerStarted","Data":"8db4d26694ce33da3784ca874a1d0c42ac443dbd30950142583950209867d91d"} Dec 06 06:06:50 crc kubenswrapper[4809]: I1206 06:06:50.001115 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8pfj4q" event={"ID":"de7c1cd3-87d5-431f-a971-5474ac3d56ec","Type":"ContainerStarted","Data":"8301274676bbbf20507bd99751a3c2d24a26c0c2973149f8641286ca5dde5a6f"} Dec 06 06:06:51 crc kubenswrapper[4809]: I1206 06:06:51.009994 4809 generic.go:334] "Generic (PLEG): container finished" podID="de7c1cd3-87d5-431f-a971-5474ac3d56ec" containerID="8301274676bbbf20507bd99751a3c2d24a26c0c2973149f8641286ca5dde5a6f" exitCode=0 Dec 06 06:06:51 crc kubenswrapper[4809]: I1206 06:06:51.010045 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8pfj4q" event={"ID":"de7c1cd3-87d5-431f-a971-5474ac3d56ec","Type":"ContainerDied","Data":"8301274676bbbf20507bd99751a3c2d24a26c0c2973149f8641286ca5dde5a6f"} Dec 06 06:06:51 crc kubenswrapper[4809]: I1206 06:06:51.012265 4809 generic.go:334] "Generic (PLEG): container finished" podID="e12ceeab-a1f8-4a5b-aefc-77d7cf057f40" containerID="8db4d26694ce33da3784ca874a1d0c42ac443dbd30950142583950209867d91d" exitCode=0 Dec 06 06:06:51 crc kubenswrapper[4809]: I1206 06:06:51.012296 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpd5xh" event={"ID":"e12ceeab-a1f8-4a5b-aefc-77d7cf057f40","Type":"ContainerDied","Data":"8db4d26694ce33da3784ca874a1d0c42ac443dbd30950142583950209867d91d"} Dec 06 06:06:54 crc kubenswrapper[4809]: I1206 06:06:54.035326 4809 generic.go:334] "Generic (PLEG): container finished" podID="e12ceeab-a1f8-4a5b-aefc-77d7cf057f40" containerID="d9a2a2aa6644a5d92a56ada3990c3d814d6e6d95d1c9999d619c7edcf1a39975" exitCode=0 Dec 06 06:06:54 crc kubenswrapper[4809]: I1206 06:06:54.035411 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpd5xh" event={"ID":"e12ceeab-a1f8-4a5b-aefc-77d7cf057f40","Type":"ContainerDied","Data":"d9a2a2aa6644a5d92a56ada3990c3d814d6e6d95d1c9999d619c7edcf1a39975"} Dec 06 06:06:55 crc kubenswrapper[4809]: I1206 06:06:55.043758 4809 generic.go:334] "Generic (PLEG): container finished" podID="e12ceeab-a1f8-4a5b-aefc-77d7cf057f40" containerID="8556235a14b9d938aec1a087e265e5c0ea023769b20399c2453c12d8b35911cc" exitCode=0 Dec 06 06:06:55 crc kubenswrapper[4809]: I1206 06:06:55.043812 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpd5xh" event={"ID":"e12ceeab-a1f8-4a5b-aefc-77d7cf057f40","Type":"ContainerDied","Data":"8556235a14b9d938aec1a087e265e5c0ea023769b20399c2453c12d8b35911cc"} Dec 06 06:06:56 crc kubenswrapper[4809]: I1206 06:06:56.052600 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8pfj4q" event={"ID":"de7c1cd3-87d5-431f-a971-5474ac3d56ec","Type":"ContainerStarted","Data":"5edf64932d864bf470a9161f1617a6192139373c1803c190bf1ca16b03c64c44"} Dec 06 06:06:56 crc kubenswrapper[4809]: I1206 06:06:56.287882 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpd5xh" Dec 06 06:06:56 crc kubenswrapper[4809]: I1206 06:06:56.337020 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e12ceeab-a1f8-4a5b-aefc-77d7cf057f40-util\") pod \"e12ceeab-a1f8-4a5b-aefc-77d7cf057f40\" (UID: \"e12ceeab-a1f8-4a5b-aefc-77d7cf057f40\") " Dec 06 06:06:56 crc kubenswrapper[4809]: I1206 06:06:56.337075 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s5sqd\" (UniqueName: \"kubernetes.io/projected/e12ceeab-a1f8-4a5b-aefc-77d7cf057f40-kube-api-access-s5sqd\") pod \"e12ceeab-a1f8-4a5b-aefc-77d7cf057f40\" (UID: \"e12ceeab-a1f8-4a5b-aefc-77d7cf057f40\") " Dec 06 06:06:56 crc kubenswrapper[4809]: I1206 06:06:56.337164 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e12ceeab-a1f8-4a5b-aefc-77d7cf057f40-bundle\") pod \"e12ceeab-a1f8-4a5b-aefc-77d7cf057f40\" (UID: \"e12ceeab-a1f8-4a5b-aefc-77d7cf057f40\") " Dec 06 06:06:56 crc kubenswrapper[4809]: I1206 06:06:56.338109 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e12ceeab-a1f8-4a5b-aefc-77d7cf057f40-bundle" (OuterVolumeSpecName: "bundle") pod "e12ceeab-a1f8-4a5b-aefc-77d7cf057f40" (UID: "e12ceeab-a1f8-4a5b-aefc-77d7cf057f40"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:06:56 crc kubenswrapper[4809]: I1206 06:06:56.342209 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e12ceeab-a1f8-4a5b-aefc-77d7cf057f40-kube-api-access-s5sqd" (OuterVolumeSpecName: "kube-api-access-s5sqd") pod "e12ceeab-a1f8-4a5b-aefc-77d7cf057f40" (UID: "e12ceeab-a1f8-4a5b-aefc-77d7cf057f40"). InnerVolumeSpecName "kube-api-access-s5sqd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:06:56 crc kubenswrapper[4809]: I1206 06:06:56.383004 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e12ceeab-a1f8-4a5b-aefc-77d7cf057f40-util" (OuterVolumeSpecName: "util") pod "e12ceeab-a1f8-4a5b-aefc-77d7cf057f40" (UID: "e12ceeab-a1f8-4a5b-aefc-77d7cf057f40"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:06:56 crc kubenswrapper[4809]: I1206 06:06:56.438349 4809 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e12ceeab-a1f8-4a5b-aefc-77d7cf057f40-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:06:56 crc kubenswrapper[4809]: I1206 06:06:56.438381 4809 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e12ceeab-a1f8-4a5b-aefc-77d7cf057f40-util\") on node \"crc\" DevicePath \"\"" Dec 06 06:06:56 crc kubenswrapper[4809]: I1206 06:06:56.438390 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s5sqd\" (UniqueName: \"kubernetes.io/projected/e12ceeab-a1f8-4a5b-aefc-77d7cf057f40-kube-api-access-s5sqd\") on node \"crc\" DevicePath \"\"" Dec 06 06:06:57 crc kubenswrapper[4809]: I1206 06:06:57.060745 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpd5xh" event={"ID":"e12ceeab-a1f8-4a5b-aefc-77d7cf057f40","Type":"ContainerDied","Data":"4c890a7cc903b7a1cdd70d2441d42740b095cea90c510ff5a70b9d81676f6a52"} Dec 06 06:06:57 crc kubenswrapper[4809]: I1206 06:06:57.060776 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpd5xh" Dec 06 06:06:57 crc kubenswrapper[4809]: I1206 06:06:57.060790 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4c890a7cc903b7a1cdd70d2441d42740b095cea90c510ff5a70b9d81676f6a52" Dec 06 06:06:57 crc kubenswrapper[4809]: I1206 06:06:57.062503 4809 generic.go:334] "Generic (PLEG): container finished" podID="de7c1cd3-87d5-431f-a971-5474ac3d56ec" containerID="5edf64932d864bf470a9161f1617a6192139373c1803c190bf1ca16b03c64c44" exitCode=0 Dec 06 06:06:57 crc kubenswrapper[4809]: I1206 06:06:57.062537 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8pfj4q" event={"ID":"de7c1cd3-87d5-431f-a971-5474ac3d56ec","Type":"ContainerDied","Data":"5edf64932d864bf470a9161f1617a6192139373c1803c190bf1ca16b03c64c44"} Dec 06 06:06:58 crc kubenswrapper[4809]: I1206 06:06:58.073723 4809 generic.go:334] "Generic (PLEG): container finished" podID="de7c1cd3-87d5-431f-a971-5474ac3d56ec" containerID="0bed582cba829b832981c042919fcdccd2b99c8e4054be57bd9d4c9c1e1a152f" exitCode=0 Dec 06 06:06:58 crc kubenswrapper[4809]: I1206 06:06:58.074052 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8pfj4q" event={"ID":"de7c1cd3-87d5-431f-a971-5474ac3d56ec","Type":"ContainerDied","Data":"0bed582cba829b832981c042919fcdccd2b99c8e4054be57bd9d4c9c1e1a152f"} Dec 06 06:06:59 crc kubenswrapper[4809]: I1206 06:06:59.395101 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8pfj4q" Dec 06 06:06:59 crc kubenswrapper[4809]: I1206 06:06:59.591987 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fr749\" (UniqueName: \"kubernetes.io/projected/de7c1cd3-87d5-431f-a971-5474ac3d56ec-kube-api-access-fr749\") pod \"de7c1cd3-87d5-431f-a971-5474ac3d56ec\" (UID: \"de7c1cd3-87d5-431f-a971-5474ac3d56ec\") " Dec 06 06:06:59 crc kubenswrapper[4809]: I1206 06:06:59.592046 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/de7c1cd3-87d5-431f-a971-5474ac3d56ec-bundle\") pod \"de7c1cd3-87d5-431f-a971-5474ac3d56ec\" (UID: \"de7c1cd3-87d5-431f-a971-5474ac3d56ec\") " Dec 06 06:06:59 crc kubenswrapper[4809]: I1206 06:06:59.592134 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/de7c1cd3-87d5-431f-a971-5474ac3d56ec-util\") pod \"de7c1cd3-87d5-431f-a971-5474ac3d56ec\" (UID: \"de7c1cd3-87d5-431f-a971-5474ac3d56ec\") " Dec 06 06:06:59 crc kubenswrapper[4809]: I1206 06:06:59.593870 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de7c1cd3-87d5-431f-a971-5474ac3d56ec-bundle" (OuterVolumeSpecName: "bundle") pod "de7c1cd3-87d5-431f-a971-5474ac3d56ec" (UID: "de7c1cd3-87d5-431f-a971-5474ac3d56ec"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:06:59 crc kubenswrapper[4809]: I1206 06:06:59.599631 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de7c1cd3-87d5-431f-a971-5474ac3d56ec-kube-api-access-fr749" (OuterVolumeSpecName: "kube-api-access-fr749") pod "de7c1cd3-87d5-431f-a971-5474ac3d56ec" (UID: "de7c1cd3-87d5-431f-a971-5474ac3d56ec"). InnerVolumeSpecName "kube-api-access-fr749". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:06:59 crc kubenswrapper[4809]: I1206 06:06:59.611105 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de7c1cd3-87d5-431f-a971-5474ac3d56ec-util" (OuterVolumeSpecName: "util") pod "de7c1cd3-87d5-431f-a971-5474ac3d56ec" (UID: "de7c1cd3-87d5-431f-a971-5474ac3d56ec"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:06:59 crc kubenswrapper[4809]: I1206 06:06:59.694209 4809 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/de7c1cd3-87d5-431f-a971-5474ac3d56ec-util\") on node \"crc\" DevicePath \"\"" Dec 06 06:06:59 crc kubenswrapper[4809]: I1206 06:06:59.694246 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fr749\" (UniqueName: \"kubernetes.io/projected/de7c1cd3-87d5-431f-a971-5474ac3d56ec-kube-api-access-fr749\") on node \"crc\" DevicePath \"\"" Dec 06 06:06:59 crc kubenswrapper[4809]: I1206 06:06:59.694264 4809 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/de7c1cd3-87d5-431f-a971-5474ac3d56ec-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:07:00 crc kubenswrapper[4809]: I1206 06:07:00.087953 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8pfj4q" event={"ID":"de7c1cd3-87d5-431f-a971-5474ac3d56ec","Type":"ContainerDied","Data":"85c5c41e678c9e349f37991b084bc738b0955faaa0d3ac115231c384ebe3c432"} Dec 06 06:07:00 crc kubenswrapper[4809]: I1206 06:07:00.088263 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85c5c41e678c9e349f37991b084bc738b0955faaa0d3ac115231c384ebe3c432" Dec 06 06:07:00 crc kubenswrapper[4809]: I1206 06:07:00.088018 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8pfj4q" Dec 06 06:07:04 crc kubenswrapper[4809]: I1206 06:07:04.717752 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-697657f76c-5stmg"] Dec 06 06:07:04 crc kubenswrapper[4809]: E1206 06:07:04.718402 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de7c1cd3-87d5-431f-a971-5474ac3d56ec" containerName="util" Dec 06 06:07:04 crc kubenswrapper[4809]: I1206 06:07:04.718417 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="de7c1cd3-87d5-431f-a971-5474ac3d56ec" containerName="util" Dec 06 06:07:04 crc kubenswrapper[4809]: E1206 06:07:04.718433 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e12ceeab-a1f8-4a5b-aefc-77d7cf057f40" containerName="pull" Dec 06 06:07:04 crc kubenswrapper[4809]: I1206 06:07:04.718441 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e12ceeab-a1f8-4a5b-aefc-77d7cf057f40" containerName="pull" Dec 06 06:07:04 crc kubenswrapper[4809]: E1206 06:07:04.718456 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de7c1cd3-87d5-431f-a971-5474ac3d56ec" containerName="pull" Dec 06 06:07:04 crc kubenswrapper[4809]: I1206 06:07:04.718463 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="de7c1cd3-87d5-431f-a971-5474ac3d56ec" containerName="pull" Dec 06 06:07:04 crc kubenswrapper[4809]: E1206 06:07:04.718472 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e12ceeab-a1f8-4a5b-aefc-77d7cf057f40" containerName="util" Dec 06 06:07:04 crc kubenswrapper[4809]: I1206 06:07:04.718479 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e12ceeab-a1f8-4a5b-aefc-77d7cf057f40" containerName="util" Dec 06 06:07:04 crc kubenswrapper[4809]: E1206 06:07:04.718489 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e12ceeab-a1f8-4a5b-aefc-77d7cf057f40" containerName="extract" Dec 06 06:07:04 crc kubenswrapper[4809]: I1206 06:07:04.718496 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e12ceeab-a1f8-4a5b-aefc-77d7cf057f40" containerName="extract" Dec 06 06:07:04 crc kubenswrapper[4809]: E1206 06:07:04.718509 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de7c1cd3-87d5-431f-a971-5474ac3d56ec" containerName="extract" Dec 06 06:07:04 crc kubenswrapper[4809]: I1206 06:07:04.718516 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="de7c1cd3-87d5-431f-a971-5474ac3d56ec" containerName="extract" Dec 06 06:07:04 crc kubenswrapper[4809]: I1206 06:07:04.718647 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="e12ceeab-a1f8-4a5b-aefc-77d7cf057f40" containerName="extract" Dec 06 06:07:04 crc kubenswrapper[4809]: I1206 06:07:04.718672 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="de7c1cd3-87d5-431f-a971-5474ac3d56ec" containerName="extract" Dec 06 06:07:04 crc kubenswrapper[4809]: I1206 06:07:04.719495 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-697657f76c-5stmg" Dec 06 06:07:04 crc kubenswrapper[4809]: I1206 06:07:04.723704 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-metrics" Dec 06 06:07:04 crc kubenswrapper[4809]: I1206 06:07:04.723805 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-dockercfg-bk6bd" Dec 06 06:07:04 crc kubenswrapper[4809]: I1206 06:07:04.723981 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"loki-operator-manager-config" Dec 06 06:07:04 crc kubenswrapper[4809]: I1206 06:07:04.724042 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-service-cert" Dec 06 06:07:04 crc kubenswrapper[4809]: I1206 06:07:04.724066 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"kube-root-ca.crt" Dec 06 06:07:04 crc kubenswrapper[4809]: I1206 06:07:04.724166 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"openshift-service-ca.crt" Dec 06 06:07:04 crc kubenswrapper[4809]: I1206 06:07:04.730150 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-697657f76c-5stmg"] Dec 06 06:07:04 crc kubenswrapper[4809]: I1206 06:07:04.774072 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/19fc8500-fddb-4650-839a-82166e98a567-webhook-cert\") pod \"loki-operator-controller-manager-697657f76c-5stmg\" (UID: \"19fc8500-fddb-4650-839a-82166e98a567\") " pod="openshift-operators-redhat/loki-operator-controller-manager-697657f76c-5stmg" Dec 06 06:07:04 crc kubenswrapper[4809]: I1206 06:07:04.774123 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjvdb\" (UniqueName: \"kubernetes.io/projected/19fc8500-fddb-4650-839a-82166e98a567-kube-api-access-bjvdb\") pod \"loki-operator-controller-manager-697657f76c-5stmg\" (UID: \"19fc8500-fddb-4650-839a-82166e98a567\") " pod="openshift-operators-redhat/loki-operator-controller-manager-697657f76c-5stmg" Dec 06 06:07:04 crc kubenswrapper[4809]: I1206 06:07:04.774168 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/19fc8500-fddb-4650-839a-82166e98a567-apiservice-cert\") pod \"loki-operator-controller-manager-697657f76c-5stmg\" (UID: \"19fc8500-fddb-4650-839a-82166e98a567\") " pod="openshift-operators-redhat/loki-operator-controller-manager-697657f76c-5stmg" Dec 06 06:07:04 crc kubenswrapper[4809]: I1206 06:07:04.774200 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/19fc8500-fddb-4650-839a-82166e98a567-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-697657f76c-5stmg\" (UID: \"19fc8500-fddb-4650-839a-82166e98a567\") " pod="openshift-operators-redhat/loki-operator-controller-manager-697657f76c-5stmg" Dec 06 06:07:04 crc kubenswrapper[4809]: I1206 06:07:04.774242 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/19fc8500-fddb-4650-839a-82166e98a567-manager-config\") pod \"loki-operator-controller-manager-697657f76c-5stmg\" (UID: \"19fc8500-fddb-4650-839a-82166e98a567\") " pod="openshift-operators-redhat/loki-operator-controller-manager-697657f76c-5stmg" Dec 06 06:07:04 crc kubenswrapper[4809]: I1206 06:07:04.875393 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/19fc8500-fddb-4650-839a-82166e98a567-apiservice-cert\") pod \"loki-operator-controller-manager-697657f76c-5stmg\" (UID: \"19fc8500-fddb-4650-839a-82166e98a567\") " pod="openshift-operators-redhat/loki-operator-controller-manager-697657f76c-5stmg" Dec 06 06:07:04 crc kubenswrapper[4809]: I1206 06:07:04.875460 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/19fc8500-fddb-4650-839a-82166e98a567-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-697657f76c-5stmg\" (UID: \"19fc8500-fddb-4650-839a-82166e98a567\") " pod="openshift-operators-redhat/loki-operator-controller-manager-697657f76c-5stmg" Dec 06 06:07:04 crc kubenswrapper[4809]: I1206 06:07:04.875504 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/19fc8500-fddb-4650-839a-82166e98a567-manager-config\") pod \"loki-operator-controller-manager-697657f76c-5stmg\" (UID: \"19fc8500-fddb-4650-839a-82166e98a567\") " pod="openshift-operators-redhat/loki-operator-controller-manager-697657f76c-5stmg" Dec 06 06:07:04 crc kubenswrapper[4809]: I1206 06:07:04.875577 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/19fc8500-fddb-4650-839a-82166e98a567-webhook-cert\") pod \"loki-operator-controller-manager-697657f76c-5stmg\" (UID: \"19fc8500-fddb-4650-839a-82166e98a567\") " pod="openshift-operators-redhat/loki-operator-controller-manager-697657f76c-5stmg" Dec 06 06:07:04 crc kubenswrapper[4809]: I1206 06:07:04.875607 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjvdb\" (UniqueName: \"kubernetes.io/projected/19fc8500-fddb-4650-839a-82166e98a567-kube-api-access-bjvdb\") pod \"loki-operator-controller-manager-697657f76c-5stmg\" (UID: \"19fc8500-fddb-4650-839a-82166e98a567\") " pod="openshift-operators-redhat/loki-operator-controller-manager-697657f76c-5stmg" Dec 06 06:07:04 crc kubenswrapper[4809]: I1206 06:07:04.876752 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/19fc8500-fddb-4650-839a-82166e98a567-manager-config\") pod \"loki-operator-controller-manager-697657f76c-5stmg\" (UID: \"19fc8500-fddb-4650-839a-82166e98a567\") " pod="openshift-operators-redhat/loki-operator-controller-manager-697657f76c-5stmg" Dec 06 06:07:04 crc kubenswrapper[4809]: I1206 06:07:04.880842 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/19fc8500-fddb-4650-839a-82166e98a567-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-697657f76c-5stmg\" (UID: \"19fc8500-fddb-4650-839a-82166e98a567\") " pod="openshift-operators-redhat/loki-operator-controller-manager-697657f76c-5stmg" Dec 06 06:07:04 crc kubenswrapper[4809]: I1206 06:07:04.880969 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/19fc8500-fddb-4650-839a-82166e98a567-apiservice-cert\") pod \"loki-operator-controller-manager-697657f76c-5stmg\" (UID: \"19fc8500-fddb-4650-839a-82166e98a567\") " pod="openshift-operators-redhat/loki-operator-controller-manager-697657f76c-5stmg" Dec 06 06:07:04 crc kubenswrapper[4809]: I1206 06:07:04.881130 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/19fc8500-fddb-4650-839a-82166e98a567-webhook-cert\") pod \"loki-operator-controller-manager-697657f76c-5stmg\" (UID: \"19fc8500-fddb-4650-839a-82166e98a567\") " pod="openshift-operators-redhat/loki-operator-controller-manager-697657f76c-5stmg" Dec 06 06:07:04 crc kubenswrapper[4809]: I1206 06:07:04.899041 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjvdb\" (UniqueName: \"kubernetes.io/projected/19fc8500-fddb-4650-839a-82166e98a567-kube-api-access-bjvdb\") pod \"loki-operator-controller-manager-697657f76c-5stmg\" (UID: \"19fc8500-fddb-4650-839a-82166e98a567\") " pod="openshift-operators-redhat/loki-operator-controller-manager-697657f76c-5stmg" Dec 06 06:07:05 crc kubenswrapper[4809]: I1206 06:07:05.037712 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-697657f76c-5stmg" Dec 06 06:07:05 crc kubenswrapper[4809]: I1206 06:07:05.468070 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-697657f76c-5stmg"] Dec 06 06:07:05 crc kubenswrapper[4809]: W1206 06:07:05.480174 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod19fc8500_fddb_4650_839a_82166e98a567.slice/crio-0b39e73d3b4fab310322bd8fdf85556084c815c681e40c81d43d050be29bd248 WatchSource:0}: Error finding container 0b39e73d3b4fab310322bd8fdf85556084c815c681e40c81d43d050be29bd248: Status 404 returned error can't find the container with id 0b39e73d3b4fab310322bd8fdf85556084c815c681e40c81d43d050be29bd248 Dec 06 06:07:06 crc kubenswrapper[4809]: I1206 06:07:06.132967 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-697657f76c-5stmg" event={"ID":"19fc8500-fddb-4650-839a-82166e98a567","Type":"ContainerStarted","Data":"0b39e73d3b4fab310322bd8fdf85556084c815c681e40c81d43d050be29bd248"} Dec 06 06:07:10 crc kubenswrapper[4809]: I1206 06:07:10.016104 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/cluster-logging-operator-ff9846bd-lcbr4"] Dec 06 06:07:10 crc kubenswrapper[4809]: I1206 06:07:10.017116 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/cluster-logging-operator-ff9846bd-lcbr4" Dec 06 06:07:10 crc kubenswrapper[4809]: I1206 06:07:10.020986 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"cluster-logging-operator-dockercfg-x22td" Dec 06 06:07:10 crc kubenswrapper[4809]: I1206 06:07:10.022203 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"kube-root-ca.crt" Dec 06 06:07:10 crc kubenswrapper[4809]: I1206 06:07:10.022637 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"openshift-service-ca.crt" Dec 06 06:07:10 crc kubenswrapper[4809]: I1206 06:07:10.029188 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/cluster-logging-operator-ff9846bd-lcbr4"] Dec 06 06:07:10 crc kubenswrapper[4809]: I1206 06:07:10.130188 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2k9df\" (UniqueName: \"kubernetes.io/projected/167bcfaa-27d5-4604-92f4-502351d91eba-kube-api-access-2k9df\") pod \"cluster-logging-operator-ff9846bd-lcbr4\" (UID: \"167bcfaa-27d5-4604-92f4-502351d91eba\") " pod="openshift-logging/cluster-logging-operator-ff9846bd-lcbr4" Dec 06 06:07:10 crc kubenswrapper[4809]: I1206 06:07:10.231594 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2k9df\" (UniqueName: \"kubernetes.io/projected/167bcfaa-27d5-4604-92f4-502351d91eba-kube-api-access-2k9df\") pod \"cluster-logging-operator-ff9846bd-lcbr4\" (UID: \"167bcfaa-27d5-4604-92f4-502351d91eba\") " pod="openshift-logging/cluster-logging-operator-ff9846bd-lcbr4" Dec 06 06:07:10 crc kubenswrapper[4809]: I1206 06:07:10.272471 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2k9df\" (UniqueName: \"kubernetes.io/projected/167bcfaa-27d5-4604-92f4-502351d91eba-kube-api-access-2k9df\") pod \"cluster-logging-operator-ff9846bd-lcbr4\" (UID: \"167bcfaa-27d5-4604-92f4-502351d91eba\") " pod="openshift-logging/cluster-logging-operator-ff9846bd-lcbr4" Dec 06 06:07:10 crc kubenswrapper[4809]: I1206 06:07:10.333128 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/cluster-logging-operator-ff9846bd-lcbr4" Dec 06 06:07:11 crc kubenswrapper[4809]: I1206 06:07:11.888678 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/cluster-logging-operator-ff9846bd-lcbr4"] Dec 06 06:07:13 crc kubenswrapper[4809]: I1206 06:07:13.196041 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/cluster-logging-operator-ff9846bd-lcbr4" event={"ID":"167bcfaa-27d5-4604-92f4-502351d91eba","Type":"ContainerStarted","Data":"107fa27aa067901b7f5ef0ad74f30e723544205ceb540706b95307f810d268cb"} Dec 06 06:07:13 crc kubenswrapper[4809]: I1206 06:07:13.199208 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-697657f76c-5stmg" event={"ID":"19fc8500-fddb-4650-839a-82166e98a567","Type":"ContainerStarted","Data":"fb4c1d3b58b0b0436f44913e8e182a2234dd3fde8bf8f57bff9bc4c4237b6157"} Dec 06 06:07:23 crc kubenswrapper[4809]: I1206 06:07:23.281961 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/cluster-logging-operator-ff9846bd-lcbr4" event={"ID":"167bcfaa-27d5-4604-92f4-502351d91eba","Type":"ContainerStarted","Data":"f82e026ff47aa3a0ad760051eb27169cd91e76903058c34e580145dd62059e1a"} Dec 06 06:07:23 crc kubenswrapper[4809]: I1206 06:07:23.284547 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-697657f76c-5stmg" event={"ID":"19fc8500-fddb-4650-839a-82166e98a567","Type":"ContainerStarted","Data":"ad5e735abfde2eb060e1a50285e136559ac1fe0ef645570d812ca1c620e1031f"} Dec 06 06:07:23 crc kubenswrapper[4809]: I1206 06:07:23.284898 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators-redhat/loki-operator-controller-manager-697657f76c-5stmg" Dec 06 06:07:23 crc kubenswrapper[4809]: I1206 06:07:23.289549 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators-redhat/loki-operator-controller-manager-697657f76c-5stmg" Dec 06 06:07:23 crc kubenswrapper[4809]: I1206 06:07:23.306635 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/cluster-logging-operator-ff9846bd-lcbr4" podStartSLOduration=3.11566478 podStartE2EDuration="13.306613192s" podCreationTimestamp="2025-12-06 06:07:10 +0000 UTC" firstStartedPulling="2025-12-06 06:07:12.529501216 +0000 UTC m=+957.418484168" lastFinishedPulling="2025-12-06 06:07:22.720449638 +0000 UTC m=+967.609432580" observedRunningTime="2025-12-06 06:07:23.304405382 +0000 UTC m=+968.193388324" watchObservedRunningTime="2025-12-06 06:07:23.306613192 +0000 UTC m=+968.195596134" Dec 06 06:07:23 crc kubenswrapper[4809]: I1206 06:07:23.344833 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators-redhat/loki-operator-controller-manager-697657f76c-5stmg" podStartSLOduration=1.93868944 podStartE2EDuration="19.344798045s" podCreationTimestamp="2025-12-06 06:07:04 +0000 UTC" firstStartedPulling="2025-12-06 06:07:05.482565733 +0000 UTC m=+950.371548665" lastFinishedPulling="2025-12-06 06:07:22.888674328 +0000 UTC m=+967.777657270" observedRunningTime="2025-12-06 06:07:23.332282216 +0000 UTC m=+968.221265168" watchObservedRunningTime="2025-12-06 06:07:23.344798045 +0000 UTC m=+968.233780987" Dec 06 06:07:28 crc kubenswrapper[4809]: I1206 06:07:28.528241 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["minio-dev/minio"] Dec 06 06:07:28 crc kubenswrapper[4809]: I1206 06:07:28.529540 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Dec 06 06:07:28 crc kubenswrapper[4809]: I1206 06:07:28.531598 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"openshift-service-ca.crt" Dec 06 06:07:28 crc kubenswrapper[4809]: I1206 06:07:28.532214 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"kube-root-ca.crt" Dec 06 06:07:28 crc kubenswrapper[4809]: I1206 06:07:28.537911 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Dec 06 06:07:28 crc kubenswrapper[4809]: I1206 06:07:28.639567 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8mm5\" (UniqueName: \"kubernetes.io/projected/fcf522a3-2c78-4e63-a25f-dde1a660e605-kube-api-access-f8mm5\") pod \"minio\" (UID: \"fcf522a3-2c78-4e63-a25f-dde1a660e605\") " pod="minio-dev/minio" Dec 06 06:07:28 crc kubenswrapper[4809]: I1206 06:07:28.639665 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-08467ed1-72e3-4853-9138-4d8d42bb95fc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-08467ed1-72e3-4853-9138-4d8d42bb95fc\") pod \"minio\" (UID: \"fcf522a3-2c78-4e63-a25f-dde1a660e605\") " pod="minio-dev/minio" Dec 06 06:07:28 crc kubenswrapper[4809]: I1206 06:07:28.741208 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-08467ed1-72e3-4853-9138-4d8d42bb95fc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-08467ed1-72e3-4853-9138-4d8d42bb95fc\") pod \"minio\" (UID: \"fcf522a3-2c78-4e63-a25f-dde1a660e605\") " pod="minio-dev/minio" Dec 06 06:07:28 crc kubenswrapper[4809]: I1206 06:07:28.741293 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8mm5\" (UniqueName: \"kubernetes.io/projected/fcf522a3-2c78-4e63-a25f-dde1a660e605-kube-api-access-f8mm5\") pod \"minio\" (UID: \"fcf522a3-2c78-4e63-a25f-dde1a660e605\") " pod="minio-dev/minio" Dec 06 06:07:28 crc kubenswrapper[4809]: I1206 06:07:28.745269 4809 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 06 06:07:28 crc kubenswrapper[4809]: I1206 06:07:28.745315 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-08467ed1-72e3-4853-9138-4d8d42bb95fc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-08467ed1-72e3-4853-9138-4d8d42bb95fc\") pod \"minio\" (UID: \"fcf522a3-2c78-4e63-a25f-dde1a660e605\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/9994e3bbab25cb55c04941709a963ce4aca5630716502a35d0a4b60649a99ff5/globalmount\"" pod="minio-dev/minio" Dec 06 06:07:28 crc kubenswrapper[4809]: I1206 06:07:28.767487 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8mm5\" (UniqueName: \"kubernetes.io/projected/fcf522a3-2c78-4e63-a25f-dde1a660e605-kube-api-access-f8mm5\") pod \"minio\" (UID: \"fcf522a3-2c78-4e63-a25f-dde1a660e605\") " pod="minio-dev/minio" Dec 06 06:07:28 crc kubenswrapper[4809]: I1206 06:07:28.790381 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-08467ed1-72e3-4853-9138-4d8d42bb95fc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-08467ed1-72e3-4853-9138-4d8d42bb95fc\") pod \"minio\" (UID: \"fcf522a3-2c78-4e63-a25f-dde1a660e605\") " pod="minio-dev/minio" Dec 06 06:07:28 crc kubenswrapper[4809]: I1206 06:07:28.846262 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Dec 06 06:07:29 crc kubenswrapper[4809]: I1206 06:07:29.322115 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Dec 06 06:07:29 crc kubenswrapper[4809]: W1206 06:07:29.341331 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfcf522a3_2c78_4e63_a25f_dde1a660e605.slice/crio-f68ab3a6245795c3a1d33104d1d42312b1354dde2f0cc97d5d88f8b8467acf80 WatchSource:0}: Error finding container f68ab3a6245795c3a1d33104d1d42312b1354dde2f0cc97d5d88f8b8467acf80: Status 404 returned error can't find the container with id f68ab3a6245795c3a1d33104d1d42312b1354dde2f0cc97d5d88f8b8467acf80 Dec 06 06:07:30 crc kubenswrapper[4809]: I1206 06:07:30.331905 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"fcf522a3-2c78-4e63-a25f-dde1a660e605","Type":"ContainerStarted","Data":"f68ab3a6245795c3a1d33104d1d42312b1354dde2f0cc97d5d88f8b8467acf80"} Dec 06 06:07:36 crc kubenswrapper[4809]: I1206 06:07:36.377671 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"fcf522a3-2c78-4e63-a25f-dde1a660e605","Type":"ContainerStarted","Data":"9551cdc45639b9d9613f7940945cd496a4380c0217d9b02481301eb120889eab"} Dec 06 06:07:36 crc kubenswrapper[4809]: I1206 06:07:36.397792 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="minio-dev/minio" podStartSLOduration=4.349493432 podStartE2EDuration="10.397772609s" podCreationTimestamp="2025-12-06 06:07:26 +0000 UTC" firstStartedPulling="2025-12-06 06:07:29.343417703 +0000 UTC m=+974.232400645" lastFinishedPulling="2025-12-06 06:07:35.39169687 +0000 UTC m=+980.280679822" observedRunningTime="2025-12-06 06:07:36.393849943 +0000 UTC m=+981.282832885" watchObservedRunningTime="2025-12-06 06:07:36.397772609 +0000 UTC m=+981.286755551" Dec 06 06:07:40 crc kubenswrapper[4809]: I1206 06:07:40.966055 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-distributor-76cc67bf56-r5p25"] Dec 06 06:07:40 crc kubenswrapper[4809]: I1206 06:07:40.967301 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-distributor-76cc67bf56-r5p25" Dec 06 06:07:40 crc kubenswrapper[4809]: I1206 06:07:40.971748 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-dockercfg-bvxgg" Dec 06 06:07:40 crc kubenswrapper[4809]: I1206 06:07:40.971752 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-distributor-grpc" Dec 06 06:07:40 crc kubenswrapper[4809]: I1206 06:07:40.972270 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-ca-bundle" Dec 06 06:07:40 crc kubenswrapper[4809]: I1206 06:07:40.973085 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-config" Dec 06 06:07:40 crc kubenswrapper[4809]: I1206 06:07:40.973269 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-distributor-http" Dec 06 06:07:40 crc kubenswrapper[4809]: I1206 06:07:40.994035 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-distributor-76cc67bf56-r5p25"] Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.098539 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ae7975b-9830-47bf-af4f-6f7fe5e47894-config\") pod \"logging-loki-distributor-76cc67bf56-r5p25\" (UID: \"2ae7975b-9830-47bf-af4f-6f7fe5e47894\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-r5p25" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.098609 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pqj7\" (UniqueName: \"kubernetes.io/projected/2ae7975b-9830-47bf-af4f-6f7fe5e47894-kube-api-access-5pqj7\") pod \"logging-loki-distributor-76cc67bf56-r5p25\" (UID: \"2ae7975b-9830-47bf-af4f-6f7fe5e47894\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-r5p25" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.098677 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2ae7975b-9830-47bf-af4f-6f7fe5e47894-logging-loki-ca-bundle\") pod \"logging-loki-distributor-76cc67bf56-r5p25\" (UID: \"2ae7975b-9830-47bf-af4f-6f7fe5e47894\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-r5p25" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.098816 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/2ae7975b-9830-47bf-af4f-6f7fe5e47894-logging-loki-distributor-http\") pod \"logging-loki-distributor-76cc67bf56-r5p25\" (UID: \"2ae7975b-9830-47bf-af4f-6f7fe5e47894\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-r5p25" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.098860 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/2ae7975b-9830-47bf-af4f-6f7fe5e47894-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-76cc67bf56-r5p25\" (UID: \"2ae7975b-9830-47bf-af4f-6f7fe5e47894\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-r5p25" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.120531 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-querier-5895d59bb8-q5rsl"] Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.121523 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-querier-5895d59bb8-q5rsl" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.123905 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-querier-grpc" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.124715 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-querier-http" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.124909 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-s3" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.146041 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-querier-5895d59bb8-q5rsl"] Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.200125 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2ae7975b-9830-47bf-af4f-6f7fe5e47894-logging-loki-ca-bundle\") pod \"logging-loki-distributor-76cc67bf56-r5p25\" (UID: \"2ae7975b-9830-47bf-af4f-6f7fe5e47894\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-r5p25" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.200214 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/2ae7975b-9830-47bf-af4f-6f7fe5e47894-logging-loki-distributor-http\") pod \"logging-loki-distributor-76cc67bf56-r5p25\" (UID: \"2ae7975b-9830-47bf-af4f-6f7fe5e47894\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-r5p25" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.200263 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/2ae7975b-9830-47bf-af4f-6f7fe5e47894-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-76cc67bf56-r5p25\" (UID: \"2ae7975b-9830-47bf-af4f-6f7fe5e47894\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-r5p25" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.200305 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ae7975b-9830-47bf-af4f-6f7fe5e47894-config\") pod \"logging-loki-distributor-76cc67bf56-r5p25\" (UID: \"2ae7975b-9830-47bf-af4f-6f7fe5e47894\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-r5p25" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.200339 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pqj7\" (UniqueName: \"kubernetes.io/projected/2ae7975b-9830-47bf-af4f-6f7fe5e47894-kube-api-access-5pqj7\") pod \"logging-loki-distributor-76cc67bf56-r5p25\" (UID: \"2ae7975b-9830-47bf-af4f-6f7fe5e47894\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-r5p25" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.201332 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2ae7975b-9830-47bf-af4f-6f7fe5e47894-logging-loki-ca-bundle\") pod \"logging-loki-distributor-76cc67bf56-r5p25\" (UID: \"2ae7975b-9830-47bf-af4f-6f7fe5e47894\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-r5p25" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.201623 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ae7975b-9830-47bf-af4f-6f7fe5e47894-config\") pod \"logging-loki-distributor-76cc67bf56-r5p25\" (UID: \"2ae7975b-9830-47bf-af4f-6f7fe5e47894\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-r5p25" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.207691 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/2ae7975b-9830-47bf-af4f-6f7fe5e47894-logging-loki-distributor-http\") pod \"logging-loki-distributor-76cc67bf56-r5p25\" (UID: \"2ae7975b-9830-47bf-af4f-6f7fe5e47894\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-r5p25" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.208450 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/2ae7975b-9830-47bf-af4f-6f7fe5e47894-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-76cc67bf56-r5p25\" (UID: \"2ae7975b-9830-47bf-af4f-6f7fe5e47894\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-r5p25" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.227050 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pqj7\" (UniqueName: \"kubernetes.io/projected/2ae7975b-9830-47bf-af4f-6f7fe5e47894-kube-api-access-5pqj7\") pod \"logging-loki-distributor-76cc67bf56-r5p25\" (UID: \"2ae7975b-9830-47bf-af4f-6f7fe5e47894\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-r5p25" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.240499 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-query-frontend-84558f7c9f-9mw6j"] Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.241325 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9mw6j" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.244324 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-query-frontend-grpc" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.244585 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-query-frontend-http" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.258663 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-query-frontend-84558f7c9f-9mw6j"] Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.296768 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-distributor-76cc67bf56-r5p25" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.301404 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4ce3880-869a-4fef-b76c-e669273700e1-config\") pod \"logging-loki-querier-5895d59bb8-q5rsl\" (UID: \"c4ce3880-869a-4fef-b76c-e669273700e1\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-q5rsl" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.301639 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/c4ce3880-869a-4fef-b76c-e669273700e1-logging-loki-s3\") pod \"logging-loki-querier-5895d59bb8-q5rsl\" (UID: \"c4ce3880-869a-4fef-b76c-e669273700e1\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-q5rsl" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.301813 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c4ce3880-869a-4fef-b76c-e669273700e1-logging-loki-ca-bundle\") pod \"logging-loki-querier-5895d59bb8-q5rsl\" (UID: \"c4ce3880-869a-4fef-b76c-e669273700e1\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-q5rsl" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.301965 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbdp8\" (UniqueName: \"kubernetes.io/projected/c4ce3880-869a-4fef-b76c-e669273700e1-kube-api-access-xbdp8\") pod \"logging-loki-querier-5895d59bb8-q5rsl\" (UID: \"c4ce3880-869a-4fef-b76c-e669273700e1\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-q5rsl" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.302095 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/c4ce3880-869a-4fef-b76c-e669273700e1-logging-loki-querier-grpc\") pod \"logging-loki-querier-5895d59bb8-q5rsl\" (UID: \"c4ce3880-869a-4fef-b76c-e669273700e1\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-q5rsl" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.302220 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/c4ce3880-869a-4fef-b76c-e669273700e1-logging-loki-querier-http\") pod \"logging-loki-querier-5895d59bb8-q5rsl\" (UID: \"c4ce3880-869a-4fef-b76c-e669273700e1\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-q5rsl" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.385963 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-gateway-65498c4f8f-48rjq"] Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.389098 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-65498c4f8f-48rjq" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.395894 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-http" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.396229 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-gateway-ca-bundle" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.396385 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-gateway" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.396510 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-dockercfg-q2phz" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.411098 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.411177 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-client-http" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.446248 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq"] Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.448043 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.452780 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/142a1d98-f469-4d1c-8541-d4ceb151f74e-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-84558f7c9f-9mw6j\" (UID: \"142a1d98-f469-4d1c-8541-d4ceb151f74e\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9mw6j" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.452846 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwb8x\" (UniqueName: \"kubernetes.io/projected/142a1d98-f469-4d1c-8541-d4ceb151f74e-kube-api-access-rwb8x\") pod \"logging-loki-query-frontend-84558f7c9f-9mw6j\" (UID: \"142a1d98-f469-4d1c-8541-d4ceb151f74e\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9mw6j" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.452906 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/72a3ed95-3e3d-4faf-88e4-ad1731902910-logging-loki-ca-bundle\") pod \"logging-loki-gateway-65498c4f8f-48rjq\" (UID: \"72a3ed95-3e3d-4faf-88e4-ad1731902910\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-48rjq" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.452961 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/72a3ed95-3e3d-4faf-88e4-ad1731902910-lokistack-gateway\") pod \"logging-loki-gateway-65498c4f8f-48rjq\" (UID: \"72a3ed95-3e3d-4faf-88e4-ad1731902910\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-48rjq" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.453229 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/142a1d98-f469-4d1c-8541-d4ceb151f74e-config\") pod \"logging-loki-query-frontend-84558f7c9f-9mw6j\" (UID: \"142a1d98-f469-4d1c-8541-d4ceb151f74e\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9mw6j" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.453262 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/72a3ed95-3e3d-4faf-88e4-ad1731902910-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-65498c4f8f-48rjq\" (UID: \"72a3ed95-3e3d-4faf-88e4-ad1731902910\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-48rjq" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.453289 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72zch\" (UniqueName: \"kubernetes.io/projected/72a3ed95-3e3d-4faf-88e4-ad1731902910-kube-api-access-72zch\") pod \"logging-loki-gateway-65498c4f8f-48rjq\" (UID: \"72a3ed95-3e3d-4faf-88e4-ad1731902910\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-48rjq" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.453360 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4ce3880-869a-4fef-b76c-e669273700e1-config\") pod \"logging-loki-querier-5895d59bb8-q5rsl\" (UID: \"c4ce3880-869a-4fef-b76c-e669273700e1\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-q5rsl" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.456978 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/72a3ed95-3e3d-4faf-88e4-ad1731902910-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-65498c4f8f-48rjq\" (UID: \"72a3ed95-3e3d-4faf-88e4-ad1731902910\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-48rjq" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.457106 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/c4ce3880-869a-4fef-b76c-e669273700e1-logging-loki-s3\") pod \"logging-loki-querier-5895d59bb8-q5rsl\" (UID: \"c4ce3880-869a-4fef-b76c-e669273700e1\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-q5rsl" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.457224 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/142a1d98-f469-4d1c-8541-d4ceb151f74e-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-84558f7c9f-9mw6j\" (UID: \"142a1d98-f469-4d1c-8541-d4ceb151f74e\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9mw6j" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.457403 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c4ce3880-869a-4fef-b76c-e669273700e1-logging-loki-ca-bundle\") pod \"logging-loki-querier-5895d59bb8-q5rsl\" (UID: \"c4ce3880-869a-4fef-b76c-e669273700e1\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-q5rsl" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.457483 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbdp8\" (UniqueName: \"kubernetes.io/projected/c4ce3880-869a-4fef-b76c-e669273700e1-kube-api-access-xbdp8\") pod \"logging-loki-querier-5895d59bb8-q5rsl\" (UID: \"c4ce3880-869a-4fef-b76c-e669273700e1\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-q5rsl" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.457562 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/c4ce3880-869a-4fef-b76c-e669273700e1-logging-loki-querier-grpc\") pod \"logging-loki-querier-5895d59bb8-q5rsl\" (UID: \"c4ce3880-869a-4fef-b76c-e669273700e1\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-q5rsl" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.457638 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/72a3ed95-3e3d-4faf-88e4-ad1731902910-tls-secret\") pod \"logging-loki-gateway-65498c4f8f-48rjq\" (UID: \"72a3ed95-3e3d-4faf-88e4-ad1731902910\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-48rjq" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.457665 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4ce3880-869a-4fef-b76c-e669273700e1-config\") pod \"logging-loki-querier-5895d59bb8-q5rsl\" (UID: \"c4ce3880-869a-4fef-b76c-e669273700e1\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-q5rsl" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.457719 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/142a1d98-f469-4d1c-8541-d4ceb151f74e-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-84558f7c9f-9mw6j\" (UID: \"142a1d98-f469-4d1c-8541-d4ceb151f74e\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9mw6j" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.457755 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/72a3ed95-3e3d-4faf-88e4-ad1731902910-rbac\") pod \"logging-loki-gateway-65498c4f8f-48rjq\" (UID: \"72a3ed95-3e3d-4faf-88e4-ad1731902910\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-48rjq" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.457825 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/c4ce3880-869a-4fef-b76c-e669273700e1-logging-loki-querier-http\") pod \"logging-loki-querier-5895d59bb8-q5rsl\" (UID: \"c4ce3880-869a-4fef-b76c-e669273700e1\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-q5rsl" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.457974 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/72a3ed95-3e3d-4faf-88e4-ad1731902910-tenants\") pod \"logging-loki-gateway-65498c4f8f-48rjq\" (UID: \"72a3ed95-3e3d-4faf-88e4-ad1731902910\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-48rjq" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.458244 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c4ce3880-869a-4fef-b76c-e669273700e1-logging-loki-ca-bundle\") pod \"logging-loki-querier-5895d59bb8-q5rsl\" (UID: \"c4ce3880-869a-4fef-b76c-e669273700e1\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-q5rsl" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.465548 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/c4ce3880-869a-4fef-b76c-e669273700e1-logging-loki-s3\") pod \"logging-loki-querier-5895d59bb8-q5rsl\" (UID: \"c4ce3880-869a-4fef-b76c-e669273700e1\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-q5rsl" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.465622 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-65498c4f8f-48rjq"] Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.466125 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/c4ce3880-869a-4fef-b76c-e669273700e1-logging-loki-querier-grpc\") pod \"logging-loki-querier-5895d59bb8-q5rsl\" (UID: \"c4ce3880-869a-4fef-b76c-e669273700e1\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-q5rsl" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.466746 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/c4ce3880-869a-4fef-b76c-e669273700e1-logging-loki-querier-http\") pod \"logging-loki-querier-5895d59bb8-q5rsl\" (UID: \"c4ce3880-869a-4fef-b76c-e669273700e1\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-q5rsl" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.505591 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbdp8\" (UniqueName: \"kubernetes.io/projected/c4ce3880-869a-4fef-b76c-e669273700e1-kube-api-access-xbdp8\") pod \"logging-loki-querier-5895d59bb8-q5rsl\" (UID: \"c4ce3880-869a-4fef-b76c-e669273700e1\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-q5rsl" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.517027 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq"] Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.578493 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/72a3ed95-3e3d-4faf-88e4-ad1731902910-tenants\") pod \"logging-loki-gateway-65498c4f8f-48rjq\" (UID: \"72a3ed95-3e3d-4faf-88e4-ad1731902910\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-48rjq" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.578547 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/142a1d98-f469-4d1c-8541-d4ceb151f74e-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-84558f7c9f-9mw6j\" (UID: \"142a1d98-f469-4d1c-8541-d4ceb151f74e\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9mw6j" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.578573 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/ef8f3e7e-57ea-42b8-a777-1778e5ed975b-rbac\") pod \"logging-loki-gateway-65498c4f8f-9s8qq\" (UID: \"ef8f3e7e-57ea-42b8-a777-1778e5ed975b\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.578593 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwb8x\" (UniqueName: \"kubernetes.io/projected/142a1d98-f469-4d1c-8541-d4ceb151f74e-kube-api-access-rwb8x\") pod \"logging-loki-query-frontend-84558f7c9f-9mw6j\" (UID: \"142a1d98-f469-4d1c-8541-d4ceb151f74e\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9mw6j" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.578614 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/ef8f3e7e-57ea-42b8-a777-1778e5ed975b-tenants\") pod \"logging-loki-gateway-65498c4f8f-9s8qq\" (UID: \"ef8f3e7e-57ea-42b8-a777-1778e5ed975b\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.578633 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/ef8f3e7e-57ea-42b8-a777-1778e5ed975b-tls-secret\") pod \"logging-loki-gateway-65498c4f8f-9s8qq\" (UID: \"ef8f3e7e-57ea-42b8-a777-1778e5ed975b\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.578653 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/72a3ed95-3e3d-4faf-88e4-ad1731902910-logging-loki-ca-bundle\") pod \"logging-loki-gateway-65498c4f8f-48rjq\" (UID: \"72a3ed95-3e3d-4faf-88e4-ad1731902910\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-48rjq" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.578670 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/72a3ed95-3e3d-4faf-88e4-ad1731902910-lokistack-gateway\") pod \"logging-loki-gateway-65498c4f8f-48rjq\" (UID: \"72a3ed95-3e3d-4faf-88e4-ad1731902910\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-48rjq" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.578688 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ef8f3e7e-57ea-42b8-a777-1778e5ed975b-logging-loki-ca-bundle\") pod \"logging-loki-gateway-65498c4f8f-9s8qq\" (UID: \"ef8f3e7e-57ea-42b8-a777-1778e5ed975b\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.578704 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/142a1d98-f469-4d1c-8541-d4ceb151f74e-config\") pod \"logging-loki-query-frontend-84558f7c9f-9mw6j\" (UID: \"142a1d98-f469-4d1c-8541-d4ceb151f74e\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9mw6j" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.578720 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/72a3ed95-3e3d-4faf-88e4-ad1731902910-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-65498c4f8f-48rjq\" (UID: \"72a3ed95-3e3d-4faf-88e4-ad1731902910\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-48rjq" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.578736 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72zch\" (UniqueName: \"kubernetes.io/projected/72a3ed95-3e3d-4faf-88e4-ad1731902910-kube-api-access-72zch\") pod \"logging-loki-gateway-65498c4f8f-48rjq\" (UID: \"72a3ed95-3e3d-4faf-88e4-ad1731902910\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-48rjq" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.578773 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/ef8f3e7e-57ea-42b8-a777-1778e5ed975b-lokistack-gateway\") pod \"logging-loki-gateway-65498c4f8f-9s8qq\" (UID: \"ef8f3e7e-57ea-42b8-a777-1778e5ed975b\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.578797 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/72a3ed95-3e3d-4faf-88e4-ad1731902910-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-65498c4f8f-48rjq\" (UID: \"72a3ed95-3e3d-4faf-88e4-ad1731902910\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-48rjq" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.578820 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/142a1d98-f469-4d1c-8541-d4ceb151f74e-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-84558f7c9f-9mw6j\" (UID: \"142a1d98-f469-4d1c-8541-d4ceb151f74e\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9mw6j" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.578842 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/72a3ed95-3e3d-4faf-88e4-ad1731902910-tls-secret\") pod \"logging-loki-gateway-65498c4f8f-48rjq\" (UID: \"72a3ed95-3e3d-4faf-88e4-ad1731902910\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-48rjq" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.578860 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/142a1d98-f469-4d1c-8541-d4ceb151f74e-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-84558f7c9f-9mw6j\" (UID: \"142a1d98-f469-4d1c-8541-d4ceb151f74e\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9mw6j" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.578876 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/72a3ed95-3e3d-4faf-88e4-ad1731902910-rbac\") pod \"logging-loki-gateway-65498c4f8f-48rjq\" (UID: \"72a3ed95-3e3d-4faf-88e4-ad1731902910\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-48rjq" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.578894 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wc7kt\" (UniqueName: \"kubernetes.io/projected/ef8f3e7e-57ea-42b8-a777-1778e5ed975b-kube-api-access-wc7kt\") pod \"logging-loki-gateway-65498c4f8f-9s8qq\" (UID: \"ef8f3e7e-57ea-42b8-a777-1778e5ed975b\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.578914 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/ef8f3e7e-57ea-42b8-a777-1778e5ed975b-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-65498c4f8f-9s8qq\" (UID: \"ef8f3e7e-57ea-42b8-a777-1778e5ed975b\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.578950 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ef8f3e7e-57ea-42b8-a777-1778e5ed975b-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-65498c4f8f-9s8qq\" (UID: \"ef8f3e7e-57ea-42b8-a777-1778e5ed975b\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.580080 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/72a3ed95-3e3d-4faf-88e4-ad1731902910-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-65498c4f8f-48rjq\" (UID: \"72a3ed95-3e3d-4faf-88e4-ad1731902910\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-48rjq" Dec 06 06:07:41 crc kubenswrapper[4809]: E1206 06:07:41.583400 4809 secret.go:188] Couldn't get secret openshift-logging/logging-loki-gateway-http: secret "logging-loki-gateway-http" not found Dec 06 06:07:41 crc kubenswrapper[4809]: E1206 06:07:41.583462 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/72a3ed95-3e3d-4faf-88e4-ad1731902910-tls-secret podName:72a3ed95-3e3d-4faf-88e4-ad1731902910 nodeName:}" failed. No retries permitted until 2025-12-06 06:07:42.083445708 +0000 UTC m=+986.972428640 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-secret" (UniqueName: "kubernetes.io/secret/72a3ed95-3e3d-4faf-88e4-ad1731902910-tls-secret") pod "logging-loki-gateway-65498c4f8f-48rjq" (UID: "72a3ed95-3e3d-4faf-88e4-ad1731902910") : secret "logging-loki-gateway-http" not found Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.586217 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/142a1d98-f469-4d1c-8541-d4ceb151f74e-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-84558f7c9f-9mw6j\" (UID: \"142a1d98-f469-4d1c-8541-d4ceb151f74e\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9mw6j" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.587006 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/72a3ed95-3e3d-4faf-88e4-ad1731902910-rbac\") pod \"logging-loki-gateway-65498c4f8f-48rjq\" (UID: \"72a3ed95-3e3d-4faf-88e4-ad1731902910\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-48rjq" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.589093 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/72a3ed95-3e3d-4faf-88e4-ad1731902910-lokistack-gateway\") pod \"logging-loki-gateway-65498c4f8f-48rjq\" (UID: \"72a3ed95-3e3d-4faf-88e4-ad1731902910\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-48rjq" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.591488 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/142a1d98-f469-4d1c-8541-d4ceb151f74e-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-84558f7c9f-9mw6j\" (UID: \"142a1d98-f469-4d1c-8541-d4ceb151f74e\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9mw6j" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.592044 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/142a1d98-f469-4d1c-8541-d4ceb151f74e-config\") pod \"logging-loki-query-frontend-84558f7c9f-9mw6j\" (UID: \"142a1d98-f469-4d1c-8541-d4ceb151f74e\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9mw6j" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.593026 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/72a3ed95-3e3d-4faf-88e4-ad1731902910-logging-loki-ca-bundle\") pod \"logging-loki-gateway-65498c4f8f-48rjq\" (UID: \"72a3ed95-3e3d-4faf-88e4-ad1731902910\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-48rjq" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.593458 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/72a3ed95-3e3d-4faf-88e4-ad1731902910-tenants\") pod \"logging-loki-gateway-65498c4f8f-48rjq\" (UID: \"72a3ed95-3e3d-4faf-88e4-ad1731902910\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-48rjq" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.593544 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/142a1d98-f469-4d1c-8541-d4ceb151f74e-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-84558f7c9f-9mw6j\" (UID: \"142a1d98-f469-4d1c-8541-d4ceb151f74e\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9mw6j" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.597753 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/72a3ed95-3e3d-4faf-88e4-ad1731902910-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-65498c4f8f-48rjq\" (UID: \"72a3ed95-3e3d-4faf-88e4-ad1731902910\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-48rjq" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.636054 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwb8x\" (UniqueName: \"kubernetes.io/projected/142a1d98-f469-4d1c-8541-d4ceb151f74e-kube-api-access-rwb8x\") pod \"logging-loki-query-frontend-84558f7c9f-9mw6j\" (UID: \"142a1d98-f469-4d1c-8541-d4ceb151f74e\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9mw6j" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.636247 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72zch\" (UniqueName: \"kubernetes.io/projected/72a3ed95-3e3d-4faf-88e4-ad1731902910-kube-api-access-72zch\") pod \"logging-loki-gateway-65498c4f8f-48rjq\" (UID: \"72a3ed95-3e3d-4faf-88e4-ad1731902910\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-48rjq" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.680246 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ef8f3e7e-57ea-42b8-a777-1778e5ed975b-logging-loki-ca-bundle\") pod \"logging-loki-gateway-65498c4f8f-9s8qq\" (UID: \"ef8f3e7e-57ea-42b8-a777-1778e5ed975b\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.680304 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/ef8f3e7e-57ea-42b8-a777-1778e5ed975b-lokistack-gateway\") pod \"logging-loki-gateway-65498c4f8f-9s8qq\" (UID: \"ef8f3e7e-57ea-42b8-a777-1778e5ed975b\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.680363 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wc7kt\" (UniqueName: \"kubernetes.io/projected/ef8f3e7e-57ea-42b8-a777-1778e5ed975b-kube-api-access-wc7kt\") pod \"logging-loki-gateway-65498c4f8f-9s8qq\" (UID: \"ef8f3e7e-57ea-42b8-a777-1778e5ed975b\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.680384 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/ef8f3e7e-57ea-42b8-a777-1778e5ed975b-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-65498c4f8f-9s8qq\" (UID: \"ef8f3e7e-57ea-42b8-a777-1778e5ed975b\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.680401 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ef8f3e7e-57ea-42b8-a777-1778e5ed975b-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-65498c4f8f-9s8qq\" (UID: \"ef8f3e7e-57ea-42b8-a777-1778e5ed975b\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.680432 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/ef8f3e7e-57ea-42b8-a777-1778e5ed975b-rbac\") pod \"logging-loki-gateway-65498c4f8f-9s8qq\" (UID: \"ef8f3e7e-57ea-42b8-a777-1778e5ed975b\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.680450 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/ef8f3e7e-57ea-42b8-a777-1778e5ed975b-tenants\") pod \"logging-loki-gateway-65498c4f8f-9s8qq\" (UID: \"ef8f3e7e-57ea-42b8-a777-1778e5ed975b\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.680470 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/ef8f3e7e-57ea-42b8-a777-1778e5ed975b-tls-secret\") pod \"logging-loki-gateway-65498c4f8f-9s8qq\" (UID: \"ef8f3e7e-57ea-42b8-a777-1778e5ed975b\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq" Dec 06 06:07:41 crc kubenswrapper[4809]: E1206 06:07:41.680579 4809 secret.go:188] Couldn't get secret openshift-logging/logging-loki-gateway-http: secret "logging-loki-gateway-http" not found Dec 06 06:07:41 crc kubenswrapper[4809]: E1206 06:07:41.680625 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ef8f3e7e-57ea-42b8-a777-1778e5ed975b-tls-secret podName:ef8f3e7e-57ea-42b8-a777-1778e5ed975b nodeName:}" failed. No retries permitted until 2025-12-06 06:07:42.180612113 +0000 UTC m=+987.069595055 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-secret" (UniqueName: "kubernetes.io/secret/ef8f3e7e-57ea-42b8-a777-1778e5ed975b-tls-secret") pod "logging-loki-gateway-65498c4f8f-9s8qq" (UID: "ef8f3e7e-57ea-42b8-a777-1778e5ed975b") : secret "logging-loki-gateway-http" not found Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.681515 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ef8f3e7e-57ea-42b8-a777-1778e5ed975b-logging-loki-ca-bundle\") pod \"logging-loki-gateway-65498c4f8f-9s8qq\" (UID: \"ef8f3e7e-57ea-42b8-a777-1778e5ed975b\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.682195 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/ef8f3e7e-57ea-42b8-a777-1778e5ed975b-lokistack-gateway\") pod \"logging-loki-gateway-65498c4f8f-9s8qq\" (UID: \"ef8f3e7e-57ea-42b8-a777-1778e5ed975b\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.686399 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/ef8f3e7e-57ea-42b8-a777-1778e5ed975b-rbac\") pod \"logging-loki-gateway-65498c4f8f-9s8qq\" (UID: \"ef8f3e7e-57ea-42b8-a777-1778e5ed975b\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.686692 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ef8f3e7e-57ea-42b8-a777-1778e5ed975b-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-65498c4f8f-9s8qq\" (UID: \"ef8f3e7e-57ea-42b8-a777-1778e5ed975b\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.688360 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/ef8f3e7e-57ea-42b8-a777-1778e5ed975b-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-65498c4f8f-9s8qq\" (UID: \"ef8f3e7e-57ea-42b8-a777-1778e5ed975b\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.690147 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/ef8f3e7e-57ea-42b8-a777-1778e5ed975b-tenants\") pod \"logging-loki-gateway-65498c4f8f-9s8qq\" (UID: \"ef8f3e7e-57ea-42b8-a777-1778e5ed975b\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.705751 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wc7kt\" (UniqueName: \"kubernetes.io/projected/ef8f3e7e-57ea-42b8-a777-1778e5ed975b-kube-api-access-wc7kt\") pod \"logging-loki-gateway-65498c4f8f-9s8qq\" (UID: \"ef8f3e7e-57ea-42b8-a777-1778e5ed975b\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.737291 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-querier-5895d59bb8-q5rsl" Dec 06 06:07:41 crc kubenswrapper[4809]: I1206 06:07:41.870337 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9mw6j" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.033187 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-distributor-76cc67bf56-r5p25"] Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.086042 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/72a3ed95-3e3d-4faf-88e4-ad1731902910-tls-secret\") pod \"logging-loki-gateway-65498c4f8f-48rjq\" (UID: \"72a3ed95-3e3d-4faf-88e4-ad1731902910\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-48rjq" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.092051 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/72a3ed95-3e3d-4faf-88e4-ad1731902910-tls-secret\") pod \"logging-loki-gateway-65498c4f8f-48rjq\" (UID: \"72a3ed95-3e3d-4faf-88e4-ad1731902910\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-48rjq" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.120724 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.121503 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-ingester-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.124656 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-ingester-http" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.125832 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-ingester-grpc" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.152719 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.173545 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-query-frontend-84558f7c9f-9mw6j"] Dec 06 06:07:42 crc kubenswrapper[4809]: W1206 06:07:42.174366 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod142a1d98_f469_4d1c_8541_d4ceb151f74e.slice/crio-bb2a56007fdd15a2567dd6619e79f13a420f0954d57ce83486e24e30a4f134e0 WatchSource:0}: Error finding container bb2a56007fdd15a2567dd6619e79f13a420f0954d57ce83486e24e30a4f134e0: Status 404 returned error can't find the container with id bb2a56007fdd15a2567dd6619e79f13a420f0954d57ce83486e24e30a4f134e0 Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.187892 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f3f468b-668a-4b15-af5b-c519c4ef03d7-config\") pod \"logging-loki-ingester-0\" (UID: \"0f3f468b-668a-4b15-af5b-c519c4ef03d7\") " pod="openshift-logging/logging-loki-ingester-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.187979 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-7f7a085c-8f20-42b4-8648-ae03a2e87db8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7f7a085c-8f20-42b4-8648-ae03a2e87db8\") pod \"logging-loki-ingester-0\" (UID: \"0f3f468b-668a-4b15-af5b-c519c4ef03d7\") " pod="openshift-logging/logging-loki-ingester-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.188031 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0f3f468b-668a-4b15-af5b-c519c4ef03d7-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"0f3f468b-668a-4b15-af5b-c519c4ef03d7\") " pod="openshift-logging/logging-loki-ingester-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.188054 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/0f3f468b-668a-4b15-af5b-c519c4ef03d7-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"0f3f468b-668a-4b15-af5b-c519c4ef03d7\") " pod="openshift-logging/logging-loki-ingester-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.188195 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2f9tr\" (UniqueName: \"kubernetes.io/projected/0f3f468b-668a-4b15-af5b-c519c4ef03d7-kube-api-access-2f9tr\") pod \"logging-loki-ingester-0\" (UID: \"0f3f468b-668a-4b15-af5b-c519c4ef03d7\") " pod="openshift-logging/logging-loki-ingester-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.188228 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/0f3f468b-668a-4b15-af5b-c519c4ef03d7-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"0f3f468b-668a-4b15-af5b-c519c4ef03d7\") " pod="openshift-logging/logging-loki-ingester-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.188242 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/0f3f468b-668a-4b15-af5b-c519c4ef03d7-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"0f3f468b-668a-4b15-af5b-c519c4ef03d7\") " pod="openshift-logging/logging-loki-ingester-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.188562 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-433abc69-1a9a-4b4c-aa00-4772672d0fe2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-433abc69-1a9a-4b4c-aa00-4772672d0fe2\") pod \"logging-loki-ingester-0\" (UID: \"0f3f468b-668a-4b15-af5b-c519c4ef03d7\") " pod="openshift-logging/logging-loki-ingester-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.188661 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/ef8f3e7e-57ea-42b8-a777-1778e5ed975b-tls-secret\") pod \"logging-loki-gateway-65498c4f8f-9s8qq\" (UID: \"ef8f3e7e-57ea-42b8-a777-1778e5ed975b\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.192966 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/ef8f3e7e-57ea-42b8-a777-1778e5ed975b-tls-secret\") pod \"logging-loki-gateway-65498c4f8f-9s8qq\" (UID: \"ef8f3e7e-57ea-42b8-a777-1778e5ed975b\") " pod="openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.236137 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.236982 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-compactor-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.240022 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-compactor-http" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.242302 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-compactor-grpc" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.251195 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.291021 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2f9tr\" (UniqueName: \"kubernetes.io/projected/0f3f468b-668a-4b15-af5b-c519c4ef03d7-kube-api-access-2f9tr\") pod \"logging-loki-ingester-0\" (UID: \"0f3f468b-668a-4b15-af5b-c519c4ef03d7\") " pod="openshift-logging/logging-loki-ingester-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.291087 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/0f3f468b-668a-4b15-af5b-c519c4ef03d7-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"0f3f468b-668a-4b15-af5b-c519c4ef03d7\") " pod="openshift-logging/logging-loki-ingester-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.291145 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/0f3f468b-668a-4b15-af5b-c519c4ef03d7-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"0f3f468b-668a-4b15-af5b-c519c4ef03d7\") " pod="openshift-logging/logging-loki-ingester-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.291173 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-433abc69-1a9a-4b4c-aa00-4772672d0fe2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-433abc69-1a9a-4b4c-aa00-4772672d0fe2\") pod \"logging-loki-ingester-0\" (UID: \"0f3f468b-668a-4b15-af5b-c519c4ef03d7\") " pod="openshift-logging/logging-loki-ingester-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.291240 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f3f468b-668a-4b15-af5b-c519c4ef03d7-config\") pod \"logging-loki-ingester-0\" (UID: \"0f3f468b-668a-4b15-af5b-c519c4ef03d7\") " pod="openshift-logging/logging-loki-ingester-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.291276 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-7f7a085c-8f20-42b4-8648-ae03a2e87db8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7f7a085c-8f20-42b4-8648-ae03a2e87db8\") pod \"logging-loki-ingester-0\" (UID: \"0f3f468b-668a-4b15-af5b-c519c4ef03d7\") " pod="openshift-logging/logging-loki-ingester-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.291314 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0f3f468b-668a-4b15-af5b-c519c4ef03d7-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"0f3f468b-668a-4b15-af5b-c519c4ef03d7\") " pod="openshift-logging/logging-loki-ingester-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.291339 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/0f3f468b-668a-4b15-af5b-c519c4ef03d7-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"0f3f468b-668a-4b15-af5b-c519c4ef03d7\") " pod="openshift-logging/logging-loki-ingester-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.292586 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0f3f468b-668a-4b15-af5b-c519c4ef03d7-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"0f3f468b-668a-4b15-af5b-c519c4ef03d7\") " pod="openshift-logging/logging-loki-ingester-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.292634 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f3f468b-668a-4b15-af5b-c519c4ef03d7-config\") pod \"logging-loki-ingester-0\" (UID: \"0f3f468b-668a-4b15-af5b-c519c4ef03d7\") " pod="openshift-logging/logging-loki-ingester-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.298315 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/0f3f468b-668a-4b15-af5b-c519c4ef03d7-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"0f3f468b-668a-4b15-af5b-c519c4ef03d7\") " pod="openshift-logging/logging-loki-ingester-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.298584 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/0f3f468b-668a-4b15-af5b-c519c4ef03d7-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"0f3f468b-668a-4b15-af5b-c519c4ef03d7\") " pod="openshift-logging/logging-loki-ingester-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.299422 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-querier-5895d59bb8-q5rsl"] Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.305519 4809 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.305564 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-7f7a085c-8f20-42b4-8648-ae03a2e87db8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7f7a085c-8f20-42b4-8648-ae03a2e87db8\") pod \"logging-loki-ingester-0\" (UID: \"0f3f468b-668a-4b15-af5b-c519c4ef03d7\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/f392416f4fa458a5beb72f81ef6e678f202450047ee18a421fa456a97a5f965d/globalmount\"" pod="openshift-logging/logging-loki-ingester-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.305592 4809 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.305631 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-433abc69-1a9a-4b4c-aa00-4772672d0fe2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-433abc69-1a9a-4b4c-aa00-4772672d0fe2\") pod \"logging-loki-ingester-0\" (UID: \"0f3f468b-668a-4b15-af5b-c519c4ef03d7\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/343086bdc7b554a646bb771abfae62168629c37755fd5acaaf8e13e7a9f8afad/globalmount\"" pod="openshift-logging/logging-loki-ingester-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.311400 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2f9tr\" (UniqueName: \"kubernetes.io/projected/0f3f468b-668a-4b15-af5b-c519c4ef03d7-kube-api-access-2f9tr\") pod \"logging-loki-ingester-0\" (UID: \"0f3f468b-668a-4b15-af5b-c519c4ef03d7\") " pod="openshift-logging/logging-loki-ingester-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.317027 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/0f3f468b-668a-4b15-af5b-c519c4ef03d7-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"0f3f468b-668a-4b15-af5b-c519c4ef03d7\") " pod="openshift-logging/logging-loki-ingester-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.321512 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.322405 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-index-gateway-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.325251 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-index-gateway-http" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.326023 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-index-gateway-grpc" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.333153 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-7f7a085c-8f20-42b4-8648-ae03a2e87db8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7f7a085c-8f20-42b4-8648-ae03a2e87db8\") pod \"logging-loki-ingester-0\" (UID: \"0f3f468b-668a-4b15-af5b-c519c4ef03d7\") " pod="openshift-logging/logging-loki-ingester-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.333560 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-65498c4f8f-48rjq" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.349765 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-433abc69-1a9a-4b4c-aa00-4772672d0fe2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-433abc69-1a9a-4b4c-aa00-4772672d0fe2\") pod \"logging-loki-ingester-0\" (UID: \"0f3f468b-668a-4b15-af5b-c519c4ef03d7\") " pod="openshift-logging/logging-loki-ingester-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.351618 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.393554 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.393947 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/2365b183-7913-4ce1-94bc-b30705d653b2-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"2365b183-7913-4ce1-94bc-b30705d653b2\") " pod="openshift-logging/logging-loki-compactor-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.394053 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tjzx\" (UniqueName: \"kubernetes.io/projected/2365b183-7913-4ce1-94bc-b30705d653b2-kube-api-access-7tjzx\") pod \"logging-loki-compactor-0\" (UID: \"2365b183-7913-4ce1-94bc-b30705d653b2\") " pod="openshift-logging/logging-loki-compactor-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.394339 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-e38be995-7327-4672-8a76-c8e59ba329af\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e38be995-7327-4672-8a76-c8e59ba329af\") pod \"logging-loki-compactor-0\" (UID: \"2365b183-7913-4ce1-94bc-b30705d653b2\") " pod="openshift-logging/logging-loki-compactor-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.394376 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/2365b183-7913-4ce1-94bc-b30705d653b2-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"2365b183-7913-4ce1-94bc-b30705d653b2\") " pod="openshift-logging/logging-loki-compactor-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.394420 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/2365b183-7913-4ce1-94bc-b30705d653b2-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"2365b183-7913-4ce1-94bc-b30705d653b2\") " pod="openshift-logging/logging-loki-compactor-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.394480 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2365b183-7913-4ce1-94bc-b30705d653b2-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"2365b183-7913-4ce1-94bc-b30705d653b2\") " pod="openshift-logging/logging-loki-compactor-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.394543 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2365b183-7913-4ce1-94bc-b30705d653b2-config\") pod \"logging-loki-compactor-0\" (UID: \"2365b183-7913-4ce1-94bc-b30705d653b2\") " pod="openshift-logging/logging-loki-compactor-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.445983 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-querier-5895d59bb8-q5rsl" event={"ID":"c4ce3880-869a-4fef-b76c-e669273700e1","Type":"ContainerStarted","Data":"21e4d82aed5d52bec81a77ba65c5ee6aa75e5190de6f29720100d704bc36b7c9"} Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.447895 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9mw6j" event={"ID":"142a1d98-f469-4d1c-8541-d4ceb151f74e","Type":"ContainerStarted","Data":"bb2a56007fdd15a2567dd6619e79f13a420f0954d57ce83486e24e30a4f134e0"} Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.450773 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-distributor-76cc67bf56-r5p25" event={"ID":"2ae7975b-9830-47bf-af4f-6f7fe5e47894","Type":"ContainerStarted","Data":"3ae8d870a1f49b28bd4e9d4c448ae1bbd1638c37be45bd08b6f1ef00397ab5dc"} Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.460269 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-ingester-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.496369 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/535f6f4f-93e7-4483-9c80-598c21bd1e6e-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"535f6f4f-93e7-4483-9c80-598c21bd1e6e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.496435 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/2365b183-7913-4ce1-94bc-b30705d653b2-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"2365b183-7913-4ce1-94bc-b30705d653b2\") " pod="openshift-logging/logging-loki-compactor-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.496480 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2365b183-7913-4ce1-94bc-b30705d653b2-config\") pod \"logging-loki-compactor-0\" (UID: \"2365b183-7913-4ce1-94bc-b30705d653b2\") " pod="openshift-logging/logging-loki-compactor-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.496521 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/2365b183-7913-4ce1-94bc-b30705d653b2-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"2365b183-7913-4ce1-94bc-b30705d653b2\") " pod="openshift-logging/logging-loki-compactor-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.496589 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tjzx\" (UniqueName: \"kubernetes.io/projected/2365b183-7913-4ce1-94bc-b30705d653b2-kube-api-access-7tjzx\") pod \"logging-loki-compactor-0\" (UID: \"2365b183-7913-4ce1-94bc-b30705d653b2\") " pod="openshift-logging/logging-loki-compactor-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.496622 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/535f6f4f-93e7-4483-9c80-598c21bd1e6e-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"535f6f4f-93e7-4483-9c80-598c21bd1e6e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.496649 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-e38be995-7327-4672-8a76-c8e59ba329af\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e38be995-7327-4672-8a76-c8e59ba329af\") pod \"logging-loki-compactor-0\" (UID: \"2365b183-7913-4ce1-94bc-b30705d653b2\") " pod="openshift-logging/logging-loki-compactor-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.496680 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/2365b183-7913-4ce1-94bc-b30705d653b2-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"2365b183-7913-4ce1-94bc-b30705d653b2\") " pod="openshift-logging/logging-loki-compactor-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.496702 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2365b183-7913-4ce1-94bc-b30705d653b2-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"2365b183-7913-4ce1-94bc-b30705d653b2\") " pod="openshift-logging/logging-loki-compactor-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.496726 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/535f6f4f-93e7-4483-9c80-598c21bd1e6e-config\") pod \"logging-loki-index-gateway-0\" (UID: \"535f6f4f-93e7-4483-9c80-598c21bd1e6e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.496748 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lz8dr\" (UniqueName: \"kubernetes.io/projected/535f6f4f-93e7-4483-9c80-598c21bd1e6e-kube-api-access-lz8dr\") pod \"logging-loki-index-gateway-0\" (UID: \"535f6f4f-93e7-4483-9c80-598c21bd1e6e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.496779 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-54d45cb9-c651-485e-b10f-9c05397a59f6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-54d45cb9-c651-485e-b10f-9c05397a59f6\") pod \"logging-loki-index-gateway-0\" (UID: \"535f6f4f-93e7-4483-9c80-598c21bd1e6e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.496826 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/535f6f4f-93e7-4483-9c80-598c21bd1e6e-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"535f6f4f-93e7-4483-9c80-598c21bd1e6e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.496860 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/535f6f4f-93e7-4483-9c80-598c21bd1e6e-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"535f6f4f-93e7-4483-9c80-598c21bd1e6e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.498856 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2365b183-7913-4ce1-94bc-b30705d653b2-config\") pod \"logging-loki-compactor-0\" (UID: \"2365b183-7913-4ce1-94bc-b30705d653b2\") " pod="openshift-logging/logging-loki-compactor-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.501499 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/2365b183-7913-4ce1-94bc-b30705d653b2-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"2365b183-7913-4ce1-94bc-b30705d653b2\") " pod="openshift-logging/logging-loki-compactor-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.502846 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2365b183-7913-4ce1-94bc-b30705d653b2-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"2365b183-7913-4ce1-94bc-b30705d653b2\") " pod="openshift-logging/logging-loki-compactor-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.504458 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/2365b183-7913-4ce1-94bc-b30705d653b2-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"2365b183-7913-4ce1-94bc-b30705d653b2\") " pod="openshift-logging/logging-loki-compactor-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.507858 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/2365b183-7913-4ce1-94bc-b30705d653b2-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"2365b183-7913-4ce1-94bc-b30705d653b2\") " pod="openshift-logging/logging-loki-compactor-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.508889 4809 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.509006 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-e38be995-7327-4672-8a76-c8e59ba329af\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e38be995-7327-4672-8a76-c8e59ba329af\") pod \"logging-loki-compactor-0\" (UID: \"2365b183-7913-4ce1-94bc-b30705d653b2\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/8bc70feef2d7ecd4224458339e125a9fb62fb52bdd0ea671b53b8d799d481dd2/globalmount\"" pod="openshift-logging/logging-loki-compactor-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.522657 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tjzx\" (UniqueName: \"kubernetes.io/projected/2365b183-7913-4ce1-94bc-b30705d653b2-kube-api-access-7tjzx\") pod \"logging-loki-compactor-0\" (UID: \"2365b183-7913-4ce1-94bc-b30705d653b2\") " pod="openshift-logging/logging-loki-compactor-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.539695 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-e38be995-7327-4672-8a76-c8e59ba329af\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e38be995-7327-4672-8a76-c8e59ba329af\") pod \"logging-loki-compactor-0\" (UID: \"2365b183-7913-4ce1-94bc-b30705d653b2\") " pod="openshift-logging/logging-loki-compactor-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.561446 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-compactor-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.587979 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-65498c4f8f-48rjq"] Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.598239 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/535f6f4f-93e7-4483-9c80-598c21bd1e6e-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"535f6f4f-93e7-4483-9c80-598c21bd1e6e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.598314 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/535f6f4f-93e7-4483-9c80-598c21bd1e6e-config\") pod \"logging-loki-index-gateway-0\" (UID: \"535f6f4f-93e7-4483-9c80-598c21bd1e6e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.598342 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lz8dr\" (UniqueName: \"kubernetes.io/projected/535f6f4f-93e7-4483-9c80-598c21bd1e6e-kube-api-access-lz8dr\") pod \"logging-loki-index-gateway-0\" (UID: \"535f6f4f-93e7-4483-9c80-598c21bd1e6e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.598372 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-54d45cb9-c651-485e-b10f-9c05397a59f6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-54d45cb9-c651-485e-b10f-9c05397a59f6\") pod \"logging-loki-index-gateway-0\" (UID: \"535f6f4f-93e7-4483-9c80-598c21bd1e6e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.598403 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/535f6f4f-93e7-4483-9c80-598c21bd1e6e-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"535f6f4f-93e7-4483-9c80-598c21bd1e6e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.598442 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/535f6f4f-93e7-4483-9c80-598c21bd1e6e-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"535f6f4f-93e7-4483-9c80-598c21bd1e6e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.598480 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/535f6f4f-93e7-4483-9c80-598c21bd1e6e-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"535f6f4f-93e7-4483-9c80-598c21bd1e6e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.600082 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/535f6f4f-93e7-4483-9c80-598c21bd1e6e-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"535f6f4f-93e7-4483-9c80-598c21bd1e6e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.600919 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/535f6f4f-93e7-4483-9c80-598c21bd1e6e-config\") pod \"logging-loki-index-gateway-0\" (UID: \"535f6f4f-93e7-4483-9c80-598c21bd1e6e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.602676 4809 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.602712 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-54d45cb9-c651-485e-b10f-9c05397a59f6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-54d45cb9-c651-485e-b10f-9c05397a59f6\") pod \"logging-loki-index-gateway-0\" (UID: \"535f6f4f-93e7-4483-9c80-598c21bd1e6e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/dec5590cb61e186ad0149ec7fdd190ee7b03ab5931b332820fd6aa662e9ee3ba/globalmount\"" pod="openshift-logging/logging-loki-index-gateway-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.608465 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/535f6f4f-93e7-4483-9c80-598c21bd1e6e-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"535f6f4f-93e7-4483-9c80-598c21bd1e6e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.612566 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/535f6f4f-93e7-4483-9c80-598c21bd1e6e-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"535f6f4f-93e7-4483-9c80-598c21bd1e6e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.616298 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/535f6f4f-93e7-4483-9c80-598c21bd1e6e-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"535f6f4f-93e7-4483-9c80-598c21bd1e6e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.625841 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lz8dr\" (UniqueName: \"kubernetes.io/projected/535f6f4f-93e7-4483-9c80-598c21bd1e6e-kube-api-access-lz8dr\") pod \"logging-loki-index-gateway-0\" (UID: \"535f6f4f-93e7-4483-9c80-598c21bd1e6e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.690836 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-54d45cb9-c651-485e-b10f-9c05397a59f6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-54d45cb9-c651-485e-b10f-9c05397a59f6\") pod \"logging-loki-index-gateway-0\" (UID: \"535f6f4f-93e7-4483-9c80-598c21bd1e6e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.712051 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-index-gateway-0" Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.847921 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Dec 06 06:07:42 crc kubenswrapper[4809]: W1206 06:07:42.859839 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0f3f468b_668a_4b15_af5b_c519c4ef03d7.slice/crio-518fd2b4fae927c52cf34e963d8f3e45567d7e96f98e75760d2b04b5b44d390d WatchSource:0}: Error finding container 518fd2b4fae927c52cf34e963d8f3e45567d7e96f98e75760d2b04b5b44d390d: Status 404 returned error can't find the container with id 518fd2b4fae927c52cf34e963d8f3e45567d7e96f98e75760d2b04b5b44d390d Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.896976 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq"] Dec 06 06:07:42 crc kubenswrapper[4809]: I1206 06:07:42.917513 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Dec 06 06:07:43 crc kubenswrapper[4809]: I1206 06:07:43.257623 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Dec 06 06:07:43 crc kubenswrapper[4809]: W1206 06:07:43.267437 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod535f6f4f_93e7_4483_9c80_598c21bd1e6e.slice/crio-92bf977af793fae60d45a5895432396b4fb808975b80ac839d40f7c22592c4f5 WatchSource:0}: Error finding container 92bf977af793fae60d45a5895432396b4fb808975b80ac839d40f7c22592c4f5: Status 404 returned error can't find the container with id 92bf977af793fae60d45a5895432396b4fb808975b80ac839d40f7c22592c4f5 Dec 06 06:07:43 crc kubenswrapper[4809]: I1206 06:07:43.461129 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-index-gateway-0" event={"ID":"535f6f4f-93e7-4483-9c80-598c21bd1e6e","Type":"ContainerStarted","Data":"92bf977af793fae60d45a5895432396b4fb808975b80ac839d40f7c22592c4f5"} Dec 06 06:07:43 crc kubenswrapper[4809]: I1206 06:07:43.462254 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-compactor-0" event={"ID":"2365b183-7913-4ce1-94bc-b30705d653b2","Type":"ContainerStarted","Data":"8ef031a432bd9eed1655c2848d24722f6038598e5d8c77cfe8c5737fb49202c5"} Dec 06 06:07:43 crc kubenswrapper[4809]: I1206 06:07:43.463340 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq" event={"ID":"ef8f3e7e-57ea-42b8-a777-1778e5ed975b","Type":"ContainerStarted","Data":"180ee0bc8151c18b8706f15c2fb9147b2d6b3fff89c7f0737e0179e512f0a245"} Dec 06 06:07:43 crc kubenswrapper[4809]: I1206 06:07:43.464400 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-ingester-0" event={"ID":"0f3f468b-668a-4b15-af5b-c519c4ef03d7","Type":"ContainerStarted","Data":"518fd2b4fae927c52cf34e963d8f3e45567d7e96f98e75760d2b04b5b44d390d"} Dec 06 06:07:43 crc kubenswrapper[4809]: I1206 06:07:43.465411 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-65498c4f8f-48rjq" event={"ID":"72a3ed95-3e3d-4faf-88e4-ad1731902910","Type":"ContainerStarted","Data":"bf20873b77e958dc5498ee9d26896686b1dd140b85b55864c5d146e9eff3f425"} Dec 06 06:07:45 crc kubenswrapper[4809]: I1206 06:07:45.822176 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-d5twx"] Dec 06 06:07:45 crc kubenswrapper[4809]: I1206 06:07:45.827252 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d5twx" Dec 06 06:07:45 crc kubenswrapper[4809]: I1206 06:07:45.841336 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-d5twx"] Dec 06 06:07:45 crc kubenswrapper[4809]: I1206 06:07:45.881625 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00-utilities\") pod \"redhat-marketplace-d5twx\" (UID: \"2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00\") " pod="openshift-marketplace/redhat-marketplace-d5twx" Dec 06 06:07:45 crc kubenswrapper[4809]: I1206 06:07:45.881722 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7bv7\" (UniqueName: \"kubernetes.io/projected/2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00-kube-api-access-g7bv7\") pod \"redhat-marketplace-d5twx\" (UID: \"2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00\") " pod="openshift-marketplace/redhat-marketplace-d5twx" Dec 06 06:07:45 crc kubenswrapper[4809]: I1206 06:07:45.881767 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00-catalog-content\") pod \"redhat-marketplace-d5twx\" (UID: \"2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00\") " pod="openshift-marketplace/redhat-marketplace-d5twx" Dec 06 06:07:45 crc kubenswrapper[4809]: I1206 06:07:45.982726 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00-catalog-content\") pod \"redhat-marketplace-d5twx\" (UID: \"2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00\") " pod="openshift-marketplace/redhat-marketplace-d5twx" Dec 06 06:07:45 crc kubenswrapper[4809]: I1206 06:07:45.982800 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00-utilities\") pod \"redhat-marketplace-d5twx\" (UID: \"2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00\") " pod="openshift-marketplace/redhat-marketplace-d5twx" Dec 06 06:07:45 crc kubenswrapper[4809]: I1206 06:07:45.982880 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7bv7\" (UniqueName: \"kubernetes.io/projected/2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00-kube-api-access-g7bv7\") pod \"redhat-marketplace-d5twx\" (UID: \"2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00\") " pod="openshift-marketplace/redhat-marketplace-d5twx" Dec 06 06:07:45 crc kubenswrapper[4809]: I1206 06:07:45.986683 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00-utilities\") pod \"redhat-marketplace-d5twx\" (UID: \"2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00\") " pod="openshift-marketplace/redhat-marketplace-d5twx" Dec 06 06:07:45 crc kubenswrapper[4809]: I1206 06:07:45.986769 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00-catalog-content\") pod \"redhat-marketplace-d5twx\" (UID: \"2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00\") " pod="openshift-marketplace/redhat-marketplace-d5twx" Dec 06 06:07:46 crc kubenswrapper[4809]: I1206 06:07:46.005048 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7bv7\" (UniqueName: \"kubernetes.io/projected/2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00-kube-api-access-g7bv7\") pod \"redhat-marketplace-d5twx\" (UID: \"2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00\") " pod="openshift-marketplace/redhat-marketplace-d5twx" Dec 06 06:07:46 crc kubenswrapper[4809]: I1206 06:07:46.150691 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d5twx" Dec 06 06:07:48 crc kubenswrapper[4809]: I1206 06:07:48.146368 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-d5twx"] Dec 06 06:07:48 crc kubenswrapper[4809]: I1206 06:07:48.517419 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-querier-5895d59bb8-q5rsl" event={"ID":"c4ce3880-869a-4fef-b76c-e669273700e1","Type":"ContainerStarted","Data":"4c810180be98d0b6d56fcfe73ea631967c9fad26d39d09621a5a51c1d615727b"} Dec 06 06:07:48 crc kubenswrapper[4809]: I1206 06:07:48.517779 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-querier-5895d59bb8-q5rsl" Dec 06 06:07:48 crc kubenswrapper[4809]: I1206 06:07:48.519592 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-65498c4f8f-48rjq" event={"ID":"72a3ed95-3e3d-4faf-88e4-ad1731902910","Type":"ContainerStarted","Data":"3a7ceb2770fd3729c89dfeef77f89ebe9ce62ecab2080e6dfe9abb71ed462de9"} Dec 06 06:07:48 crc kubenswrapper[4809]: I1206 06:07:48.520966 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9mw6j" event={"ID":"142a1d98-f469-4d1c-8541-d4ceb151f74e","Type":"ContainerStarted","Data":"1dafbbf2842ccb8965faba1d3b6d6827d4c2554e88eafe1d919d6bcbdf705112"} Dec 06 06:07:48 crc kubenswrapper[4809]: I1206 06:07:48.521212 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9mw6j" Dec 06 06:07:48 crc kubenswrapper[4809]: I1206 06:07:48.523082 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-compactor-0" event={"ID":"2365b183-7913-4ce1-94bc-b30705d653b2","Type":"ContainerStarted","Data":"7a8e22c1b8e4d4dc2b8557d4a7ad962ed2d3183469fc0d0b5f3e72051e10bae7"} Dec 06 06:07:48 crc kubenswrapper[4809]: I1206 06:07:48.523824 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-compactor-0" Dec 06 06:07:48 crc kubenswrapper[4809]: I1206 06:07:48.525124 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq" event={"ID":"ef8f3e7e-57ea-42b8-a777-1778e5ed975b","Type":"ContainerStarted","Data":"6fd4376ce21e3816d2609666c2e2f8c976689cbd174bc5cced25676447515bec"} Dec 06 06:07:48 crc kubenswrapper[4809]: I1206 06:07:48.526151 4809 generic.go:334] "Generic (PLEG): container finished" podID="2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00" containerID="3766b2cb5f760f6460c334541c09b657c21dfe1d8c5958c91ee89455ef9506f2" exitCode=0 Dec 06 06:07:48 crc kubenswrapper[4809]: I1206 06:07:48.526277 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d5twx" event={"ID":"2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00","Type":"ContainerDied","Data":"3766b2cb5f760f6460c334541c09b657c21dfe1d8c5958c91ee89455ef9506f2"} Dec 06 06:07:48 crc kubenswrapper[4809]: I1206 06:07:48.526354 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d5twx" event={"ID":"2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00","Type":"ContainerStarted","Data":"b2a30b0e5761d1d338327013ac1a8e69f3e45942137d7517fbc8ae06be79b765"} Dec 06 06:07:48 crc kubenswrapper[4809]: I1206 06:07:48.530614 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-distributor-76cc67bf56-r5p25" event={"ID":"2ae7975b-9830-47bf-af4f-6f7fe5e47894","Type":"ContainerStarted","Data":"87f34767ec628a2d432269cc5f17cc28e6adab17385c1ecd5ec165db3a567e86"} Dec 06 06:07:48 crc kubenswrapper[4809]: I1206 06:07:48.530754 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-distributor-76cc67bf56-r5p25" Dec 06 06:07:48 crc kubenswrapper[4809]: I1206 06:07:48.533995 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-ingester-0" event={"ID":"0f3f468b-668a-4b15-af5b-c519c4ef03d7","Type":"ContainerStarted","Data":"ceb2ea318bc1c558415834e63690bf5117d21bcd64d337815f14027abee506e6"} Dec 06 06:07:48 crc kubenswrapper[4809]: I1206 06:07:48.534087 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-ingester-0" Dec 06 06:07:48 crc kubenswrapper[4809]: I1206 06:07:48.541088 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-index-gateway-0" event={"ID":"535f6f4f-93e7-4483-9c80-598c21bd1e6e","Type":"ContainerStarted","Data":"4aea08df2a4f52eb024f36c4d8598cdd983886372e3fd8817f9fa0d366dfdec7"} Dec 06 06:07:48 crc kubenswrapper[4809]: I1206 06:07:48.541248 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-index-gateway-0" Dec 06 06:07:48 crc kubenswrapper[4809]: I1206 06:07:48.541500 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-querier-5895d59bb8-q5rsl" podStartSLOduration=2.009870142 podStartE2EDuration="7.541484377s" podCreationTimestamp="2025-12-06 06:07:41 +0000 UTC" firstStartedPulling="2025-12-06 06:07:42.308670599 +0000 UTC m=+987.197653531" lastFinishedPulling="2025-12-06 06:07:47.840284824 +0000 UTC m=+992.729267766" observedRunningTime="2025-12-06 06:07:48.533733137 +0000 UTC m=+993.422716089" watchObservedRunningTime="2025-12-06 06:07:48.541484377 +0000 UTC m=+993.430467319" Dec 06 06:07:48 crc kubenswrapper[4809]: I1206 06:07:48.556571 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9mw6j" podStartSLOduration=1.7980826300000001 podStartE2EDuration="7.556550493s" podCreationTimestamp="2025-12-06 06:07:41 +0000 UTC" firstStartedPulling="2025-12-06 06:07:42.179743466 +0000 UTC m=+987.068726398" lastFinishedPulling="2025-12-06 06:07:47.938211319 +0000 UTC m=+992.827194261" observedRunningTime="2025-12-06 06:07:48.555730221 +0000 UTC m=+993.444713153" watchObservedRunningTime="2025-12-06 06:07:48.556550493 +0000 UTC m=+993.445533435" Dec 06 06:07:48 crc kubenswrapper[4809]: I1206 06:07:48.594000 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-distributor-76cc67bf56-r5p25" podStartSLOduration=2.845282375 podStartE2EDuration="8.593975364s" podCreationTimestamp="2025-12-06 06:07:40 +0000 UTC" firstStartedPulling="2025-12-06 06:07:42.045395747 +0000 UTC m=+986.934378689" lastFinishedPulling="2025-12-06 06:07:47.794088736 +0000 UTC m=+992.683071678" observedRunningTime="2025-12-06 06:07:48.591896009 +0000 UTC m=+993.480878951" watchObservedRunningTime="2025-12-06 06:07:48.593975364 +0000 UTC m=+993.482958306" Dec 06 06:07:48 crc kubenswrapper[4809]: I1206 06:07:48.614810 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-compactor-0" podStartSLOduration=2.797427697 podStartE2EDuration="7.614715585s" podCreationTimestamp="2025-12-06 06:07:41 +0000 UTC" firstStartedPulling="2025-12-06 06:07:42.920340233 +0000 UTC m=+987.809323175" lastFinishedPulling="2025-12-06 06:07:47.737628121 +0000 UTC m=+992.626611063" observedRunningTime="2025-12-06 06:07:48.608581069 +0000 UTC m=+993.497564021" watchObservedRunningTime="2025-12-06 06:07:48.614715585 +0000 UTC m=+993.503698547" Dec 06 06:07:48 crc kubenswrapper[4809]: I1206 06:07:48.627620 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-index-gateway-0" podStartSLOduration=3.063877076 podStartE2EDuration="7.627605513s" podCreationTimestamp="2025-12-06 06:07:41 +0000 UTC" firstStartedPulling="2025-12-06 06:07:43.270105192 +0000 UTC m=+988.159088134" lastFinishedPulling="2025-12-06 06:07:47.833833629 +0000 UTC m=+992.722816571" observedRunningTime="2025-12-06 06:07:48.625219779 +0000 UTC m=+993.514202721" watchObservedRunningTime="2025-12-06 06:07:48.627605513 +0000 UTC m=+993.516588455" Dec 06 06:07:48 crc kubenswrapper[4809]: I1206 06:07:48.655765 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-ingester-0" podStartSLOduration=2.7144480250000003 podStartE2EDuration="7.655743913s" podCreationTimestamp="2025-12-06 06:07:41 +0000 UTC" firstStartedPulling="2025-12-06 06:07:42.863520318 +0000 UTC m=+987.752503260" lastFinishedPulling="2025-12-06 06:07:47.804816206 +0000 UTC m=+992.693799148" observedRunningTime="2025-12-06 06:07:48.651665122 +0000 UTC m=+993.540648054" watchObservedRunningTime="2025-12-06 06:07:48.655743913 +0000 UTC m=+993.544726855" Dec 06 06:07:49 crc kubenswrapper[4809]: I1206 06:07:49.552335 4809 generic.go:334] "Generic (PLEG): container finished" podID="2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00" containerID="3eafd7a93b4c2781932a6a5633692e5a998e2267bd814a6fab9555c1cfe27a50" exitCode=0 Dec 06 06:07:49 crc kubenswrapper[4809]: I1206 06:07:49.552419 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d5twx" event={"ID":"2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00","Type":"ContainerDied","Data":"3eafd7a93b4c2781932a6a5633692e5a998e2267bd814a6fab9555c1cfe27a50"} Dec 06 06:07:51 crc kubenswrapper[4809]: I1206 06:07:51.567517 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq" event={"ID":"ef8f3e7e-57ea-42b8-a777-1778e5ed975b","Type":"ContainerStarted","Data":"88ff25f7b1f0235cc37862b2b985f0a0b3a392170f96bf9d980790a8b1d834bc"} Dec 06 06:07:51 crc kubenswrapper[4809]: I1206 06:07:51.569320 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq" Dec 06 06:07:51 crc kubenswrapper[4809]: I1206 06:07:51.569423 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq" Dec 06 06:07:51 crc kubenswrapper[4809]: I1206 06:07:51.572552 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d5twx" event={"ID":"2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00","Type":"ContainerStarted","Data":"f14f6cc8a1629492e948858156ba6278de7f6860843677daa792447867714126"} Dec 06 06:07:51 crc kubenswrapper[4809]: I1206 06:07:51.574976 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-65498c4f8f-48rjq" event={"ID":"72a3ed95-3e3d-4faf-88e4-ad1731902910","Type":"ContainerStarted","Data":"6784436a0f92cd2c3c7daeac8ec23b3e3e8b74dd7b175859c2afbbca25291b13"} Dec 06 06:07:51 crc kubenswrapper[4809]: I1206 06:07:51.575220 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-65498c4f8f-48rjq" Dec 06 06:07:51 crc kubenswrapper[4809]: I1206 06:07:51.581544 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq" Dec 06 06:07:51 crc kubenswrapper[4809]: I1206 06:07:51.582151 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq" Dec 06 06:07:51 crc kubenswrapper[4809]: I1206 06:07:51.587268 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-65498c4f8f-48rjq" Dec 06 06:07:51 crc kubenswrapper[4809]: I1206 06:07:51.594259 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq" podStartSLOduration=2.786377349 podStartE2EDuration="10.594240045s" podCreationTimestamp="2025-12-06 06:07:41 +0000 UTC" firstStartedPulling="2025-12-06 06:07:42.91984701 +0000 UTC m=+987.808829952" lastFinishedPulling="2025-12-06 06:07:50.727709706 +0000 UTC m=+995.616692648" observedRunningTime="2025-12-06 06:07:51.592825357 +0000 UTC m=+996.481808319" watchObservedRunningTime="2025-12-06 06:07:51.594240045 +0000 UTC m=+996.483222987" Dec 06 06:07:51 crc kubenswrapper[4809]: I1206 06:07:51.620348 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-gateway-65498c4f8f-48rjq" podStartSLOduration=2.5118296940000002 podStartE2EDuration="10.620330011s" podCreationTimestamp="2025-12-06 06:07:41 +0000 UTC" firstStartedPulling="2025-12-06 06:07:42.618604033 +0000 UTC m=+987.507586975" lastFinishedPulling="2025-12-06 06:07:50.72710435 +0000 UTC m=+995.616087292" observedRunningTime="2025-12-06 06:07:51.617099523 +0000 UTC m=+996.506082465" watchObservedRunningTime="2025-12-06 06:07:51.620330011 +0000 UTC m=+996.509312963" Dec 06 06:07:51 crc kubenswrapper[4809]: I1206 06:07:51.683886 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-d5twx" podStartSLOduration=4.484563353 podStartE2EDuration="6.683864767s" podCreationTimestamp="2025-12-06 06:07:45 +0000 UTC" firstStartedPulling="2025-12-06 06:07:48.528345931 +0000 UTC m=+993.417328873" lastFinishedPulling="2025-12-06 06:07:50.727647345 +0000 UTC m=+995.616630287" observedRunningTime="2025-12-06 06:07:51.678620065 +0000 UTC m=+996.567602997" watchObservedRunningTime="2025-12-06 06:07:51.683864767 +0000 UTC m=+996.572847709" Dec 06 06:07:52 crc kubenswrapper[4809]: I1206 06:07:52.334682 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-65498c4f8f-48rjq" Dec 06 06:07:52 crc kubenswrapper[4809]: I1206 06:07:52.345312 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-65498c4f8f-48rjq" Dec 06 06:07:54 crc kubenswrapper[4809]: I1206 06:07:54.012186 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nmqr8"] Dec 06 06:07:54 crc kubenswrapper[4809]: I1206 06:07:54.013847 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nmqr8" Dec 06 06:07:54 crc kubenswrapper[4809]: I1206 06:07:54.029444 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nmqr8"] Dec 06 06:07:54 crc kubenswrapper[4809]: I1206 06:07:54.144620 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b0474b0-184e-4390-af58-914ebc95f6b3-utilities\") pod \"community-operators-nmqr8\" (UID: \"2b0474b0-184e-4390-af58-914ebc95f6b3\") " pod="openshift-marketplace/community-operators-nmqr8" Dec 06 06:07:54 crc kubenswrapper[4809]: I1206 06:07:54.144668 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkvsp\" (UniqueName: \"kubernetes.io/projected/2b0474b0-184e-4390-af58-914ebc95f6b3-kube-api-access-hkvsp\") pod \"community-operators-nmqr8\" (UID: \"2b0474b0-184e-4390-af58-914ebc95f6b3\") " pod="openshift-marketplace/community-operators-nmqr8" Dec 06 06:07:54 crc kubenswrapper[4809]: I1206 06:07:54.144695 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b0474b0-184e-4390-af58-914ebc95f6b3-catalog-content\") pod \"community-operators-nmqr8\" (UID: \"2b0474b0-184e-4390-af58-914ebc95f6b3\") " pod="openshift-marketplace/community-operators-nmqr8" Dec 06 06:07:54 crc kubenswrapper[4809]: I1206 06:07:54.246083 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b0474b0-184e-4390-af58-914ebc95f6b3-utilities\") pod \"community-operators-nmqr8\" (UID: \"2b0474b0-184e-4390-af58-914ebc95f6b3\") " pod="openshift-marketplace/community-operators-nmqr8" Dec 06 06:07:54 crc kubenswrapper[4809]: I1206 06:07:54.246132 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkvsp\" (UniqueName: \"kubernetes.io/projected/2b0474b0-184e-4390-af58-914ebc95f6b3-kube-api-access-hkvsp\") pod \"community-operators-nmqr8\" (UID: \"2b0474b0-184e-4390-af58-914ebc95f6b3\") " pod="openshift-marketplace/community-operators-nmqr8" Dec 06 06:07:54 crc kubenswrapper[4809]: I1206 06:07:54.246155 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b0474b0-184e-4390-af58-914ebc95f6b3-catalog-content\") pod \"community-operators-nmqr8\" (UID: \"2b0474b0-184e-4390-af58-914ebc95f6b3\") " pod="openshift-marketplace/community-operators-nmqr8" Dec 06 06:07:54 crc kubenswrapper[4809]: I1206 06:07:54.246759 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b0474b0-184e-4390-af58-914ebc95f6b3-utilities\") pod \"community-operators-nmqr8\" (UID: \"2b0474b0-184e-4390-af58-914ebc95f6b3\") " pod="openshift-marketplace/community-operators-nmqr8" Dec 06 06:07:54 crc kubenswrapper[4809]: I1206 06:07:54.246858 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b0474b0-184e-4390-af58-914ebc95f6b3-catalog-content\") pod \"community-operators-nmqr8\" (UID: \"2b0474b0-184e-4390-af58-914ebc95f6b3\") " pod="openshift-marketplace/community-operators-nmqr8" Dec 06 06:07:54 crc kubenswrapper[4809]: I1206 06:07:54.281975 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkvsp\" (UniqueName: \"kubernetes.io/projected/2b0474b0-184e-4390-af58-914ebc95f6b3-kube-api-access-hkvsp\") pod \"community-operators-nmqr8\" (UID: \"2b0474b0-184e-4390-af58-914ebc95f6b3\") " pod="openshift-marketplace/community-operators-nmqr8" Dec 06 06:07:54 crc kubenswrapper[4809]: I1206 06:07:54.342078 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nmqr8" Dec 06 06:07:54 crc kubenswrapper[4809]: I1206 06:07:54.687512 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nmqr8"] Dec 06 06:07:55 crc kubenswrapper[4809]: I1206 06:07:55.605506 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nmqr8" event={"ID":"2b0474b0-184e-4390-af58-914ebc95f6b3","Type":"ContainerStarted","Data":"d692815f356917e7f093560be6c77fdf277f6553f42812c2aedeb043af839b46"} Dec 06 06:07:56 crc kubenswrapper[4809]: I1206 06:07:56.150800 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-d5twx" Dec 06 06:07:56 crc kubenswrapper[4809]: I1206 06:07:56.151164 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-d5twx" Dec 06 06:07:56 crc kubenswrapper[4809]: I1206 06:07:56.203699 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-d5twx" Dec 06 06:07:56 crc kubenswrapper[4809]: I1206 06:07:56.615763 4809 generic.go:334] "Generic (PLEG): container finished" podID="2b0474b0-184e-4390-af58-914ebc95f6b3" containerID="dcb5c85a733629864389d7ce301c62be30d85cde44e892a8a8235587d608e2d7" exitCode=0 Dec 06 06:07:56 crc kubenswrapper[4809]: I1206 06:07:56.615804 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nmqr8" event={"ID":"2b0474b0-184e-4390-af58-914ebc95f6b3","Type":"ContainerDied","Data":"dcb5c85a733629864389d7ce301c62be30d85cde44e892a8a8235587d608e2d7"} Dec 06 06:07:56 crc kubenswrapper[4809]: I1206 06:07:56.676640 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-d5twx" Dec 06 06:07:57 crc kubenswrapper[4809]: I1206 06:07:57.606905 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8s92z"] Dec 06 06:07:57 crc kubenswrapper[4809]: I1206 06:07:57.608973 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8s92z" Dec 06 06:07:57 crc kubenswrapper[4809]: I1206 06:07:57.621627 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8s92z"] Dec 06 06:07:57 crc kubenswrapper[4809]: I1206 06:07:57.708238 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ace93089-c55e-41e9-b41b-fd0eed2b96da-utilities\") pod \"certified-operators-8s92z\" (UID: \"ace93089-c55e-41e9-b41b-fd0eed2b96da\") " pod="openshift-marketplace/certified-operators-8s92z" Dec 06 06:07:57 crc kubenswrapper[4809]: I1206 06:07:57.708595 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ace93089-c55e-41e9-b41b-fd0eed2b96da-catalog-content\") pod \"certified-operators-8s92z\" (UID: \"ace93089-c55e-41e9-b41b-fd0eed2b96da\") " pod="openshift-marketplace/certified-operators-8s92z" Dec 06 06:07:57 crc kubenswrapper[4809]: I1206 06:07:57.708719 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvs75\" (UniqueName: \"kubernetes.io/projected/ace93089-c55e-41e9-b41b-fd0eed2b96da-kube-api-access-cvs75\") pod \"certified-operators-8s92z\" (UID: \"ace93089-c55e-41e9-b41b-fd0eed2b96da\") " pod="openshift-marketplace/certified-operators-8s92z" Dec 06 06:07:57 crc kubenswrapper[4809]: I1206 06:07:57.810468 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ace93089-c55e-41e9-b41b-fd0eed2b96da-utilities\") pod \"certified-operators-8s92z\" (UID: \"ace93089-c55e-41e9-b41b-fd0eed2b96da\") " pod="openshift-marketplace/certified-operators-8s92z" Dec 06 06:07:57 crc kubenswrapper[4809]: I1206 06:07:57.810870 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ace93089-c55e-41e9-b41b-fd0eed2b96da-catalog-content\") pod \"certified-operators-8s92z\" (UID: \"ace93089-c55e-41e9-b41b-fd0eed2b96da\") " pod="openshift-marketplace/certified-operators-8s92z" Dec 06 06:07:57 crc kubenswrapper[4809]: I1206 06:07:57.811073 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvs75\" (UniqueName: \"kubernetes.io/projected/ace93089-c55e-41e9-b41b-fd0eed2b96da-kube-api-access-cvs75\") pod \"certified-operators-8s92z\" (UID: \"ace93089-c55e-41e9-b41b-fd0eed2b96da\") " pod="openshift-marketplace/certified-operators-8s92z" Dec 06 06:07:57 crc kubenswrapper[4809]: I1206 06:07:57.811340 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ace93089-c55e-41e9-b41b-fd0eed2b96da-utilities\") pod \"certified-operators-8s92z\" (UID: \"ace93089-c55e-41e9-b41b-fd0eed2b96da\") " pod="openshift-marketplace/certified-operators-8s92z" Dec 06 06:07:57 crc kubenswrapper[4809]: I1206 06:07:57.811583 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ace93089-c55e-41e9-b41b-fd0eed2b96da-catalog-content\") pod \"certified-operators-8s92z\" (UID: \"ace93089-c55e-41e9-b41b-fd0eed2b96da\") " pod="openshift-marketplace/certified-operators-8s92z" Dec 06 06:07:57 crc kubenswrapper[4809]: I1206 06:07:57.839725 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvs75\" (UniqueName: \"kubernetes.io/projected/ace93089-c55e-41e9-b41b-fd0eed2b96da-kube-api-access-cvs75\") pod \"certified-operators-8s92z\" (UID: \"ace93089-c55e-41e9-b41b-fd0eed2b96da\") " pod="openshift-marketplace/certified-operators-8s92z" Dec 06 06:07:57 crc kubenswrapper[4809]: I1206 06:07:57.928766 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8s92z" Dec 06 06:07:58 crc kubenswrapper[4809]: I1206 06:07:58.479862 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8s92z"] Dec 06 06:07:58 crc kubenswrapper[4809]: I1206 06:07:58.637768 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8s92z" event={"ID":"ace93089-c55e-41e9-b41b-fd0eed2b96da","Type":"ContainerStarted","Data":"580df8492155587ab84b23df0eec9fbdee88f32898e38ee47e26ef9591495258"} Dec 06 06:07:59 crc kubenswrapper[4809]: I1206 06:07:59.646722 4809 generic.go:334] "Generic (PLEG): container finished" podID="2b0474b0-184e-4390-af58-914ebc95f6b3" containerID="72fffd6c6ea6edb81df3dbdabd88eb2b437251e7e640740c67b344215bf7f769" exitCode=0 Dec 06 06:07:59 crc kubenswrapper[4809]: I1206 06:07:59.646832 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nmqr8" event={"ID":"2b0474b0-184e-4390-af58-914ebc95f6b3","Type":"ContainerDied","Data":"72fffd6c6ea6edb81df3dbdabd88eb2b437251e7e640740c67b344215bf7f769"} Dec 06 06:07:59 crc kubenswrapper[4809]: I1206 06:07:59.650074 4809 generic.go:334] "Generic (PLEG): container finished" podID="ace93089-c55e-41e9-b41b-fd0eed2b96da" containerID="ff87f21a315c677f2aab8975fc7d3693a8ebe9646a4e6a164f5e04c960f53f88" exitCode=0 Dec 06 06:07:59 crc kubenswrapper[4809]: I1206 06:07:59.650124 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8s92z" event={"ID":"ace93089-c55e-41e9-b41b-fd0eed2b96da","Type":"ContainerDied","Data":"ff87f21a315c677f2aab8975fc7d3693a8ebe9646a4e6a164f5e04c960f53f88"} Dec 06 06:08:00 crc kubenswrapper[4809]: I1206 06:08:00.662362 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8s92z" event={"ID":"ace93089-c55e-41e9-b41b-fd0eed2b96da","Type":"ContainerStarted","Data":"40260dcd63e5237091815e17c1ab381f950b4a83a22cb248843767a902ea1b7d"} Dec 06 06:08:00 crc kubenswrapper[4809]: I1206 06:08:00.665909 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nmqr8" event={"ID":"2b0474b0-184e-4390-af58-914ebc95f6b3","Type":"ContainerStarted","Data":"a09a50de40db3a915e0f3f82fb532dfc10ee7a324e94fddc261e9996f3849ea0"} Dec 06 06:08:00 crc kubenswrapper[4809]: I1206 06:08:00.716298 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nmqr8" podStartSLOduration=5.257740966 podStartE2EDuration="7.716261792s" podCreationTimestamp="2025-12-06 06:07:53 +0000 UTC" firstStartedPulling="2025-12-06 06:07:57.625658851 +0000 UTC m=+1002.514641793" lastFinishedPulling="2025-12-06 06:08:00.084179677 +0000 UTC m=+1004.973162619" observedRunningTime="2025-12-06 06:08:00.707863136 +0000 UTC m=+1005.596846098" watchObservedRunningTime="2025-12-06 06:08:00.716261792 +0000 UTC m=+1005.605244734" Dec 06 06:08:01 crc kubenswrapper[4809]: I1206 06:08:01.203789 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-d5twx"] Dec 06 06:08:01 crc kubenswrapper[4809]: I1206 06:08:01.204717 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-d5twx" podUID="2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00" containerName="registry-server" containerID="cri-o://f14f6cc8a1629492e948858156ba6278de7f6860843677daa792447867714126" gracePeriod=2 Dec 06 06:08:01 crc kubenswrapper[4809]: I1206 06:08:01.677024 4809 generic.go:334] "Generic (PLEG): container finished" podID="ace93089-c55e-41e9-b41b-fd0eed2b96da" containerID="40260dcd63e5237091815e17c1ab381f950b4a83a22cb248843767a902ea1b7d" exitCode=0 Dec 06 06:08:01 crc kubenswrapper[4809]: I1206 06:08:01.677079 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8s92z" event={"ID":"ace93089-c55e-41e9-b41b-fd0eed2b96da","Type":"ContainerDied","Data":"40260dcd63e5237091815e17c1ab381f950b4a83a22cb248843767a902ea1b7d"} Dec 06 06:08:01 crc kubenswrapper[4809]: I1206 06:08:01.681007 4809 generic.go:334] "Generic (PLEG): container finished" podID="2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00" containerID="f14f6cc8a1629492e948858156ba6278de7f6860843677daa792447867714126" exitCode=0 Dec 06 06:08:01 crc kubenswrapper[4809]: I1206 06:08:01.681069 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d5twx" event={"ID":"2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00","Type":"ContainerDied","Data":"f14f6cc8a1629492e948858156ba6278de7f6860843677daa792447867714126"} Dec 06 06:08:01 crc kubenswrapper[4809]: I1206 06:08:01.681130 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d5twx" event={"ID":"2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00","Type":"ContainerDied","Data":"b2a30b0e5761d1d338327013ac1a8e69f3e45942137d7517fbc8ae06be79b765"} Dec 06 06:08:01 crc kubenswrapper[4809]: I1206 06:08:01.681150 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b2a30b0e5761d1d338327013ac1a8e69f3e45942137d7517fbc8ae06be79b765" Dec 06 06:08:01 crc kubenswrapper[4809]: I1206 06:08:01.715058 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d5twx" Dec 06 06:08:01 crc kubenswrapper[4809]: I1206 06:08:01.786390 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00-utilities\") pod \"2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00\" (UID: \"2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00\") " Dec 06 06:08:01 crc kubenswrapper[4809]: I1206 06:08:01.786577 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7bv7\" (UniqueName: \"kubernetes.io/projected/2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00-kube-api-access-g7bv7\") pod \"2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00\" (UID: \"2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00\") " Dec 06 06:08:01 crc kubenswrapper[4809]: I1206 06:08:01.786632 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00-catalog-content\") pod \"2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00\" (UID: \"2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00\") " Dec 06 06:08:01 crc kubenswrapper[4809]: I1206 06:08:01.787420 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00-utilities" (OuterVolumeSpecName: "utilities") pod "2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00" (UID: "2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:08:01 crc kubenswrapper[4809]: I1206 06:08:01.792414 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00-kube-api-access-g7bv7" (OuterVolumeSpecName: "kube-api-access-g7bv7") pod "2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00" (UID: "2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00"). InnerVolumeSpecName "kube-api-access-g7bv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:08:01 crc kubenswrapper[4809]: I1206 06:08:01.813906 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00" (UID: "2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:08:01 crc kubenswrapper[4809]: I1206 06:08:01.888457 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:08:01 crc kubenswrapper[4809]: I1206 06:08:01.888520 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:08:01 crc kubenswrapper[4809]: I1206 06:08:01.888538 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g7bv7\" (UniqueName: \"kubernetes.io/projected/2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00-kube-api-access-g7bv7\") on node \"crc\" DevicePath \"\"" Dec 06 06:08:02 crc kubenswrapper[4809]: I1206 06:08:02.573396 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-compactor-0" Dec 06 06:08:02 crc kubenswrapper[4809]: I1206 06:08:02.689817 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d5twx" Dec 06 06:08:02 crc kubenswrapper[4809]: I1206 06:08:02.689857 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8s92z" event={"ID":"ace93089-c55e-41e9-b41b-fd0eed2b96da","Type":"ContainerStarted","Data":"cb94a39aeee5b9ea5438dbe247ecdc31736f1015f3452d78067d3ab03dac3b8d"} Dec 06 06:08:02 crc kubenswrapper[4809]: I1206 06:08:02.712386 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8s92z" podStartSLOduration=3.316085613 podStartE2EDuration="5.712366487s" podCreationTimestamp="2025-12-06 06:07:57 +0000 UTC" firstStartedPulling="2025-12-06 06:07:59.652755873 +0000 UTC m=+1004.541738815" lastFinishedPulling="2025-12-06 06:08:02.049036747 +0000 UTC m=+1006.938019689" observedRunningTime="2025-12-06 06:08:02.711248816 +0000 UTC m=+1007.600231758" watchObservedRunningTime="2025-12-06 06:08:02.712366487 +0000 UTC m=+1007.601349429" Dec 06 06:08:02 crc kubenswrapper[4809]: I1206 06:08:02.770711 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-d5twx"] Dec 06 06:08:02 crc kubenswrapper[4809]: I1206 06:08:02.783202 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-d5twx"] Dec 06 06:08:03 crc kubenswrapper[4809]: I1206 06:08:03.400086 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00" path="/var/lib/kubelet/pods/2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00/volumes" Dec 06 06:08:04 crc kubenswrapper[4809]: I1206 06:08:04.342649 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nmqr8" Dec 06 06:08:04 crc kubenswrapper[4809]: I1206 06:08:04.342713 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nmqr8" Dec 06 06:08:04 crc kubenswrapper[4809]: I1206 06:08:04.417357 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nmqr8" Dec 06 06:08:04 crc kubenswrapper[4809]: I1206 06:08:04.496853 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:08:04 crc kubenswrapper[4809]: I1206 06:08:04.496941 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:08:07 crc kubenswrapper[4809]: I1206 06:08:07.929507 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8s92z" Dec 06 06:08:07 crc kubenswrapper[4809]: I1206 06:08:07.930166 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8s92z" Dec 06 06:08:07 crc kubenswrapper[4809]: I1206 06:08:07.976723 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8s92z" Dec 06 06:08:08 crc kubenswrapper[4809]: I1206 06:08:08.793454 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8s92z" Dec 06 06:08:11 crc kubenswrapper[4809]: I1206 06:08:11.198962 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8s92z"] Dec 06 06:08:11 crc kubenswrapper[4809]: I1206 06:08:11.199515 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8s92z" podUID="ace93089-c55e-41e9-b41b-fd0eed2b96da" containerName="registry-server" containerID="cri-o://cb94a39aeee5b9ea5438dbe247ecdc31736f1015f3452d78067d3ab03dac3b8d" gracePeriod=2 Dec 06 06:08:11 crc kubenswrapper[4809]: I1206 06:08:11.304657 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-distributor-76cc67bf56-r5p25" Dec 06 06:08:11 crc kubenswrapper[4809]: I1206 06:08:11.705473 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8s92z" Dec 06 06:08:11 crc kubenswrapper[4809]: I1206 06:08:11.745806 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-querier-5895d59bb8-q5rsl" Dec 06 06:08:11 crc kubenswrapper[4809]: I1206 06:08:11.766258 4809 generic.go:334] "Generic (PLEG): container finished" podID="ace93089-c55e-41e9-b41b-fd0eed2b96da" containerID="cb94a39aeee5b9ea5438dbe247ecdc31736f1015f3452d78067d3ab03dac3b8d" exitCode=0 Dec 06 06:08:11 crc kubenswrapper[4809]: I1206 06:08:11.766318 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8s92z" Dec 06 06:08:11 crc kubenswrapper[4809]: I1206 06:08:11.766315 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8s92z" event={"ID":"ace93089-c55e-41e9-b41b-fd0eed2b96da","Type":"ContainerDied","Data":"cb94a39aeee5b9ea5438dbe247ecdc31736f1015f3452d78067d3ab03dac3b8d"} Dec 06 06:08:11 crc kubenswrapper[4809]: I1206 06:08:11.766353 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8s92z" event={"ID":"ace93089-c55e-41e9-b41b-fd0eed2b96da","Type":"ContainerDied","Data":"580df8492155587ab84b23df0eec9fbdee88f32898e38ee47e26ef9591495258"} Dec 06 06:08:11 crc kubenswrapper[4809]: I1206 06:08:11.766374 4809 scope.go:117] "RemoveContainer" containerID="cb94a39aeee5b9ea5438dbe247ecdc31736f1015f3452d78067d3ab03dac3b8d" Dec 06 06:08:11 crc kubenswrapper[4809]: I1206 06:08:11.771118 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ace93089-c55e-41e9-b41b-fd0eed2b96da-utilities\") pod \"ace93089-c55e-41e9-b41b-fd0eed2b96da\" (UID: \"ace93089-c55e-41e9-b41b-fd0eed2b96da\") " Dec 06 06:08:11 crc kubenswrapper[4809]: I1206 06:08:11.771264 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cvs75\" (UniqueName: \"kubernetes.io/projected/ace93089-c55e-41e9-b41b-fd0eed2b96da-kube-api-access-cvs75\") pod \"ace93089-c55e-41e9-b41b-fd0eed2b96da\" (UID: \"ace93089-c55e-41e9-b41b-fd0eed2b96da\") " Dec 06 06:08:11 crc kubenswrapper[4809]: I1206 06:08:11.771312 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ace93089-c55e-41e9-b41b-fd0eed2b96da-catalog-content\") pod \"ace93089-c55e-41e9-b41b-fd0eed2b96da\" (UID: \"ace93089-c55e-41e9-b41b-fd0eed2b96da\") " Dec 06 06:08:11 crc kubenswrapper[4809]: I1206 06:08:11.772641 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ace93089-c55e-41e9-b41b-fd0eed2b96da-utilities" (OuterVolumeSpecName: "utilities") pod "ace93089-c55e-41e9-b41b-fd0eed2b96da" (UID: "ace93089-c55e-41e9-b41b-fd0eed2b96da"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:08:11 crc kubenswrapper[4809]: I1206 06:08:11.782371 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ace93089-c55e-41e9-b41b-fd0eed2b96da-kube-api-access-cvs75" (OuterVolumeSpecName: "kube-api-access-cvs75") pod "ace93089-c55e-41e9-b41b-fd0eed2b96da" (UID: "ace93089-c55e-41e9-b41b-fd0eed2b96da"). InnerVolumeSpecName "kube-api-access-cvs75". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:08:11 crc kubenswrapper[4809]: I1206 06:08:11.809146 4809 scope.go:117] "RemoveContainer" containerID="40260dcd63e5237091815e17c1ab381f950b4a83a22cb248843767a902ea1b7d" Dec 06 06:08:11 crc kubenswrapper[4809]: I1206 06:08:11.854061 4809 scope.go:117] "RemoveContainer" containerID="ff87f21a315c677f2aab8975fc7d3693a8ebe9646a4e6a164f5e04c960f53f88" Dec 06 06:08:11 crc kubenswrapper[4809]: I1206 06:08:11.873426 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ace93089-c55e-41e9-b41b-fd0eed2b96da-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:08:11 crc kubenswrapper[4809]: I1206 06:08:11.873481 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cvs75\" (UniqueName: \"kubernetes.io/projected/ace93089-c55e-41e9-b41b-fd0eed2b96da-kube-api-access-cvs75\") on node \"crc\" DevicePath \"\"" Dec 06 06:08:11 crc kubenswrapper[4809]: I1206 06:08:11.875492 4809 scope.go:117] "RemoveContainer" containerID="cb94a39aeee5b9ea5438dbe247ecdc31736f1015f3452d78067d3ab03dac3b8d" Dec 06 06:08:11 crc kubenswrapper[4809]: E1206 06:08:11.877880 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb94a39aeee5b9ea5438dbe247ecdc31736f1015f3452d78067d3ab03dac3b8d\": container with ID starting with cb94a39aeee5b9ea5438dbe247ecdc31736f1015f3452d78067d3ab03dac3b8d not found: ID does not exist" containerID="cb94a39aeee5b9ea5438dbe247ecdc31736f1015f3452d78067d3ab03dac3b8d" Dec 06 06:08:11 crc kubenswrapper[4809]: I1206 06:08:11.877946 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb94a39aeee5b9ea5438dbe247ecdc31736f1015f3452d78067d3ab03dac3b8d"} err="failed to get container status \"cb94a39aeee5b9ea5438dbe247ecdc31736f1015f3452d78067d3ab03dac3b8d\": rpc error: code = NotFound desc = could not find container \"cb94a39aeee5b9ea5438dbe247ecdc31736f1015f3452d78067d3ab03dac3b8d\": container with ID starting with cb94a39aeee5b9ea5438dbe247ecdc31736f1015f3452d78067d3ab03dac3b8d not found: ID does not exist" Dec 06 06:08:11 crc kubenswrapper[4809]: I1206 06:08:11.877982 4809 scope.go:117] "RemoveContainer" containerID="40260dcd63e5237091815e17c1ab381f950b4a83a22cb248843767a902ea1b7d" Dec 06 06:08:11 crc kubenswrapper[4809]: E1206 06:08:11.879394 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40260dcd63e5237091815e17c1ab381f950b4a83a22cb248843767a902ea1b7d\": container with ID starting with 40260dcd63e5237091815e17c1ab381f950b4a83a22cb248843767a902ea1b7d not found: ID does not exist" containerID="40260dcd63e5237091815e17c1ab381f950b4a83a22cb248843767a902ea1b7d" Dec 06 06:08:11 crc kubenswrapper[4809]: I1206 06:08:11.879445 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40260dcd63e5237091815e17c1ab381f950b4a83a22cb248843767a902ea1b7d"} err="failed to get container status \"40260dcd63e5237091815e17c1ab381f950b4a83a22cb248843767a902ea1b7d\": rpc error: code = NotFound desc = could not find container \"40260dcd63e5237091815e17c1ab381f950b4a83a22cb248843767a902ea1b7d\": container with ID starting with 40260dcd63e5237091815e17c1ab381f950b4a83a22cb248843767a902ea1b7d not found: ID does not exist" Dec 06 06:08:11 crc kubenswrapper[4809]: I1206 06:08:11.879483 4809 scope.go:117] "RemoveContainer" containerID="ff87f21a315c677f2aab8975fc7d3693a8ebe9646a4e6a164f5e04c960f53f88" Dec 06 06:08:11 crc kubenswrapper[4809]: E1206 06:08:11.880256 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff87f21a315c677f2aab8975fc7d3693a8ebe9646a4e6a164f5e04c960f53f88\": container with ID starting with ff87f21a315c677f2aab8975fc7d3693a8ebe9646a4e6a164f5e04c960f53f88 not found: ID does not exist" containerID="ff87f21a315c677f2aab8975fc7d3693a8ebe9646a4e6a164f5e04c960f53f88" Dec 06 06:08:11 crc kubenswrapper[4809]: I1206 06:08:11.880512 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff87f21a315c677f2aab8975fc7d3693a8ebe9646a4e6a164f5e04c960f53f88"} err="failed to get container status \"ff87f21a315c677f2aab8975fc7d3693a8ebe9646a4e6a164f5e04c960f53f88\": rpc error: code = NotFound desc = could not find container \"ff87f21a315c677f2aab8975fc7d3693a8ebe9646a4e6a164f5e04c960f53f88\": container with ID starting with ff87f21a315c677f2aab8975fc7d3693a8ebe9646a4e6a164f5e04c960f53f88 not found: ID does not exist" Dec 06 06:08:11 crc kubenswrapper[4809]: I1206 06:08:11.882024 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9mw6j" Dec 06 06:08:12 crc kubenswrapper[4809]: I1206 06:08:12.057255 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ace93089-c55e-41e9-b41b-fd0eed2b96da-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ace93089-c55e-41e9-b41b-fd0eed2b96da" (UID: "ace93089-c55e-41e9-b41b-fd0eed2b96da"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:08:12 crc kubenswrapper[4809]: I1206 06:08:12.077527 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ace93089-c55e-41e9-b41b-fd0eed2b96da-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:08:12 crc kubenswrapper[4809]: I1206 06:08:12.115250 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8s92z"] Dec 06 06:08:12 crc kubenswrapper[4809]: I1206 06:08:12.122663 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8s92z"] Dec 06 06:08:12 crc kubenswrapper[4809]: I1206 06:08:12.720296 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-index-gateway-0" Dec 06 06:08:12 crc kubenswrapper[4809]: I1206 06:08:12.881716 4809 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: this instance owns no tokens Dec 06 06:08:12 crc kubenswrapper[4809]: I1206 06:08:12.881775 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="0f3f468b-668a-4b15-af5b-c519c4ef03d7" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 06 06:08:13 crc kubenswrapper[4809]: I1206 06:08:13.398969 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ace93089-c55e-41e9-b41b-fd0eed2b96da" path="/var/lib/kubelet/pods/ace93089-c55e-41e9-b41b-fd0eed2b96da/volumes" Dec 06 06:08:14 crc kubenswrapper[4809]: I1206 06:08:14.416777 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nmqr8" Dec 06 06:08:17 crc kubenswrapper[4809]: I1206 06:08:17.800867 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nmqr8"] Dec 06 06:08:17 crc kubenswrapper[4809]: I1206 06:08:17.802214 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nmqr8" podUID="2b0474b0-184e-4390-af58-914ebc95f6b3" containerName="registry-server" containerID="cri-o://a09a50de40db3a915e0f3f82fb532dfc10ee7a324e94fddc261e9996f3849ea0" gracePeriod=2 Dec 06 06:08:18 crc kubenswrapper[4809]: I1206 06:08:18.739135 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nmqr8" Dec 06 06:08:18 crc kubenswrapper[4809]: I1206 06:08:18.805725 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hkvsp\" (UniqueName: \"kubernetes.io/projected/2b0474b0-184e-4390-af58-914ebc95f6b3-kube-api-access-hkvsp\") pod \"2b0474b0-184e-4390-af58-914ebc95f6b3\" (UID: \"2b0474b0-184e-4390-af58-914ebc95f6b3\") " Dec 06 06:08:18 crc kubenswrapper[4809]: I1206 06:08:18.805831 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b0474b0-184e-4390-af58-914ebc95f6b3-utilities\") pod \"2b0474b0-184e-4390-af58-914ebc95f6b3\" (UID: \"2b0474b0-184e-4390-af58-914ebc95f6b3\") " Dec 06 06:08:18 crc kubenswrapper[4809]: I1206 06:08:18.805873 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b0474b0-184e-4390-af58-914ebc95f6b3-catalog-content\") pod \"2b0474b0-184e-4390-af58-914ebc95f6b3\" (UID: \"2b0474b0-184e-4390-af58-914ebc95f6b3\") " Dec 06 06:08:18 crc kubenswrapper[4809]: I1206 06:08:18.807424 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b0474b0-184e-4390-af58-914ebc95f6b3-utilities" (OuterVolumeSpecName: "utilities") pod "2b0474b0-184e-4390-af58-914ebc95f6b3" (UID: "2b0474b0-184e-4390-af58-914ebc95f6b3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:08:18 crc kubenswrapper[4809]: I1206 06:08:18.821776 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b0474b0-184e-4390-af58-914ebc95f6b3-kube-api-access-hkvsp" (OuterVolumeSpecName: "kube-api-access-hkvsp") pod "2b0474b0-184e-4390-af58-914ebc95f6b3" (UID: "2b0474b0-184e-4390-af58-914ebc95f6b3"). InnerVolumeSpecName "kube-api-access-hkvsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:08:18 crc kubenswrapper[4809]: I1206 06:08:18.853123 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b0474b0-184e-4390-af58-914ebc95f6b3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2b0474b0-184e-4390-af58-914ebc95f6b3" (UID: "2b0474b0-184e-4390-af58-914ebc95f6b3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:08:18 crc kubenswrapper[4809]: I1206 06:08:18.907362 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b0474b0-184e-4390-af58-914ebc95f6b3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:08:18 crc kubenswrapper[4809]: I1206 06:08:18.907405 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hkvsp\" (UniqueName: \"kubernetes.io/projected/2b0474b0-184e-4390-af58-914ebc95f6b3-kube-api-access-hkvsp\") on node \"crc\" DevicePath \"\"" Dec 06 06:08:18 crc kubenswrapper[4809]: I1206 06:08:18.907418 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b0474b0-184e-4390-af58-914ebc95f6b3-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:08:18 crc kubenswrapper[4809]: I1206 06:08:18.927210 4809 generic.go:334] "Generic (PLEG): container finished" podID="2b0474b0-184e-4390-af58-914ebc95f6b3" containerID="a09a50de40db3a915e0f3f82fb532dfc10ee7a324e94fddc261e9996f3849ea0" exitCode=0 Dec 06 06:08:18 crc kubenswrapper[4809]: I1206 06:08:18.927260 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nmqr8" event={"ID":"2b0474b0-184e-4390-af58-914ebc95f6b3","Type":"ContainerDied","Data":"a09a50de40db3a915e0f3f82fb532dfc10ee7a324e94fddc261e9996f3849ea0"} Dec 06 06:08:18 crc kubenswrapper[4809]: I1206 06:08:18.927268 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nmqr8" Dec 06 06:08:18 crc kubenswrapper[4809]: I1206 06:08:18.927287 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nmqr8" event={"ID":"2b0474b0-184e-4390-af58-914ebc95f6b3","Type":"ContainerDied","Data":"d692815f356917e7f093560be6c77fdf277f6553f42812c2aedeb043af839b46"} Dec 06 06:08:18 crc kubenswrapper[4809]: I1206 06:08:18.927305 4809 scope.go:117] "RemoveContainer" containerID="a09a50de40db3a915e0f3f82fb532dfc10ee7a324e94fddc261e9996f3849ea0" Dec 06 06:08:18 crc kubenswrapper[4809]: I1206 06:08:18.946466 4809 scope.go:117] "RemoveContainer" containerID="72fffd6c6ea6edb81df3dbdabd88eb2b437251e7e640740c67b344215bf7f769" Dec 06 06:08:18 crc kubenswrapper[4809]: I1206 06:08:18.963213 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nmqr8"] Dec 06 06:08:18 crc kubenswrapper[4809]: I1206 06:08:18.976914 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nmqr8"] Dec 06 06:08:18 crc kubenswrapper[4809]: I1206 06:08:18.982182 4809 scope.go:117] "RemoveContainer" containerID="dcb5c85a733629864389d7ce301c62be30d85cde44e892a8a8235587d608e2d7" Dec 06 06:08:19 crc kubenswrapper[4809]: I1206 06:08:19.026977 4809 scope.go:117] "RemoveContainer" containerID="a09a50de40db3a915e0f3f82fb532dfc10ee7a324e94fddc261e9996f3849ea0" Dec 06 06:08:19 crc kubenswrapper[4809]: E1206 06:08:19.027351 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a09a50de40db3a915e0f3f82fb532dfc10ee7a324e94fddc261e9996f3849ea0\": container with ID starting with a09a50de40db3a915e0f3f82fb532dfc10ee7a324e94fddc261e9996f3849ea0 not found: ID does not exist" containerID="a09a50de40db3a915e0f3f82fb532dfc10ee7a324e94fddc261e9996f3849ea0" Dec 06 06:08:19 crc kubenswrapper[4809]: I1206 06:08:19.027402 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a09a50de40db3a915e0f3f82fb532dfc10ee7a324e94fddc261e9996f3849ea0"} err="failed to get container status \"a09a50de40db3a915e0f3f82fb532dfc10ee7a324e94fddc261e9996f3849ea0\": rpc error: code = NotFound desc = could not find container \"a09a50de40db3a915e0f3f82fb532dfc10ee7a324e94fddc261e9996f3849ea0\": container with ID starting with a09a50de40db3a915e0f3f82fb532dfc10ee7a324e94fddc261e9996f3849ea0 not found: ID does not exist" Dec 06 06:08:19 crc kubenswrapper[4809]: I1206 06:08:19.027427 4809 scope.go:117] "RemoveContainer" containerID="72fffd6c6ea6edb81df3dbdabd88eb2b437251e7e640740c67b344215bf7f769" Dec 06 06:08:19 crc kubenswrapper[4809]: E1206 06:08:19.027713 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72fffd6c6ea6edb81df3dbdabd88eb2b437251e7e640740c67b344215bf7f769\": container with ID starting with 72fffd6c6ea6edb81df3dbdabd88eb2b437251e7e640740c67b344215bf7f769 not found: ID does not exist" containerID="72fffd6c6ea6edb81df3dbdabd88eb2b437251e7e640740c67b344215bf7f769" Dec 06 06:08:19 crc kubenswrapper[4809]: I1206 06:08:19.027746 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72fffd6c6ea6edb81df3dbdabd88eb2b437251e7e640740c67b344215bf7f769"} err="failed to get container status \"72fffd6c6ea6edb81df3dbdabd88eb2b437251e7e640740c67b344215bf7f769\": rpc error: code = NotFound desc = could not find container \"72fffd6c6ea6edb81df3dbdabd88eb2b437251e7e640740c67b344215bf7f769\": container with ID starting with 72fffd6c6ea6edb81df3dbdabd88eb2b437251e7e640740c67b344215bf7f769 not found: ID does not exist" Dec 06 06:08:19 crc kubenswrapper[4809]: I1206 06:08:19.027767 4809 scope.go:117] "RemoveContainer" containerID="dcb5c85a733629864389d7ce301c62be30d85cde44e892a8a8235587d608e2d7" Dec 06 06:08:19 crc kubenswrapper[4809]: E1206 06:08:19.028189 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dcb5c85a733629864389d7ce301c62be30d85cde44e892a8a8235587d608e2d7\": container with ID starting with dcb5c85a733629864389d7ce301c62be30d85cde44e892a8a8235587d608e2d7 not found: ID does not exist" containerID="dcb5c85a733629864389d7ce301c62be30d85cde44e892a8a8235587d608e2d7" Dec 06 06:08:19 crc kubenswrapper[4809]: I1206 06:08:19.028219 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcb5c85a733629864389d7ce301c62be30d85cde44e892a8a8235587d608e2d7"} err="failed to get container status \"dcb5c85a733629864389d7ce301c62be30d85cde44e892a8a8235587d608e2d7\": rpc error: code = NotFound desc = could not find container \"dcb5c85a733629864389d7ce301c62be30d85cde44e892a8a8235587d608e2d7\": container with ID starting with dcb5c85a733629864389d7ce301c62be30d85cde44e892a8a8235587d608e2d7 not found: ID does not exist" Dec 06 06:08:19 crc kubenswrapper[4809]: I1206 06:08:19.397796 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b0474b0-184e-4390-af58-914ebc95f6b3" path="/var/lib/kubelet/pods/2b0474b0-184e-4390-af58-914ebc95f6b3/volumes" Dec 06 06:08:22 crc kubenswrapper[4809]: I1206 06:08:22.465172 4809 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: waiting for 15s after being ready Dec 06 06:08:22 crc kubenswrapper[4809]: I1206 06:08:22.465568 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="0f3f468b-668a-4b15-af5b-c519c4ef03d7" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 06 06:08:32 crc kubenswrapper[4809]: I1206 06:08:32.467382 4809 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: waiting for 15s after being ready Dec 06 06:08:32 crc kubenswrapper[4809]: I1206 06:08:32.468376 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="0f3f468b-668a-4b15-af5b-c519c4ef03d7" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 06 06:08:34 crc kubenswrapper[4809]: I1206 06:08:34.496366 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:08:34 crc kubenswrapper[4809]: I1206 06:08:34.496718 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:08:42 crc kubenswrapper[4809]: I1206 06:08:42.466342 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-ingester-0" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.811267 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/collector-cnlrl"] Dec 06 06:09:01 crc kubenswrapper[4809]: E1206 06:09:01.812176 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b0474b0-184e-4390-af58-914ebc95f6b3" containerName="extract-content" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.812196 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b0474b0-184e-4390-af58-914ebc95f6b3" containerName="extract-content" Dec 06 06:09:01 crc kubenswrapper[4809]: E1206 06:09:01.812210 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b0474b0-184e-4390-af58-914ebc95f6b3" containerName="registry-server" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.812217 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b0474b0-184e-4390-af58-914ebc95f6b3" containerName="registry-server" Dec 06 06:09:01 crc kubenswrapper[4809]: E1206 06:09:01.812229 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b0474b0-184e-4390-af58-914ebc95f6b3" containerName="extract-utilities" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.812237 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b0474b0-184e-4390-af58-914ebc95f6b3" containerName="extract-utilities" Dec 06 06:09:01 crc kubenswrapper[4809]: E1206 06:09:01.812249 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ace93089-c55e-41e9-b41b-fd0eed2b96da" containerName="registry-server" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.812256 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ace93089-c55e-41e9-b41b-fd0eed2b96da" containerName="registry-server" Dec 06 06:09:01 crc kubenswrapper[4809]: E1206 06:09:01.812273 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00" containerName="registry-server" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.812280 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00" containerName="registry-server" Dec 06 06:09:01 crc kubenswrapper[4809]: E1206 06:09:01.812302 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00" containerName="extract-content" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.812311 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00" containerName="extract-content" Dec 06 06:09:01 crc kubenswrapper[4809]: E1206 06:09:01.812329 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00" containerName="extract-utilities" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.812337 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00" containerName="extract-utilities" Dec 06 06:09:01 crc kubenswrapper[4809]: E1206 06:09:01.812346 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ace93089-c55e-41e9-b41b-fd0eed2b96da" containerName="extract-content" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.812353 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ace93089-c55e-41e9-b41b-fd0eed2b96da" containerName="extract-content" Dec 06 06:09:01 crc kubenswrapper[4809]: E1206 06:09:01.812363 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ace93089-c55e-41e9-b41b-fd0eed2b96da" containerName="extract-utilities" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.812370 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ace93089-c55e-41e9-b41b-fd0eed2b96da" containerName="extract-utilities" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.812524 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ace93089-c55e-41e9-b41b-fd0eed2b96da" containerName="registry-server" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.812540 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bd3cf7e-f882-4472-9ddd-2bb66cd9ee00" containerName="registry-server" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.812554 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b0474b0-184e-4390-af58-914ebc95f6b3" containerName="registry-server" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.813270 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-cnlrl" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.819608 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-token" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.819657 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-dockercfg-4nb2f" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.820121 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-config" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.820244 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-metrics" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.820414 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-syslog-receiver" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.826724 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/7d2b2ca1-f855-4921-ac74-71bf5e70345b-collector-syslog-receiver\") pod \"collector-cnlrl\" (UID: \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\") " pod="openshift-logging/collector-cnlrl" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.826798 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/7d2b2ca1-f855-4921-ac74-71bf5e70345b-metrics\") pod \"collector-cnlrl\" (UID: \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\") " pod="openshift-logging/collector-cnlrl" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.826841 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/7d2b2ca1-f855-4921-ac74-71bf5e70345b-collector-token\") pod \"collector-cnlrl\" (UID: \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\") " pod="openshift-logging/collector-cnlrl" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.826925 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7d2b2ca1-f855-4921-ac74-71bf5e70345b-trusted-ca\") pod \"collector-cnlrl\" (UID: \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\") " pod="openshift-logging/collector-cnlrl" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.826993 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/7d2b2ca1-f855-4921-ac74-71bf5e70345b-config-openshift-service-cacrt\") pod \"collector-cnlrl\" (UID: \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\") " pod="openshift-logging/collector-cnlrl" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.827024 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/7d2b2ca1-f855-4921-ac74-71bf5e70345b-datadir\") pod \"collector-cnlrl\" (UID: \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\") " pod="openshift-logging/collector-cnlrl" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.827106 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rtmw\" (UniqueName: \"kubernetes.io/projected/7d2b2ca1-f855-4921-ac74-71bf5e70345b-kube-api-access-8rtmw\") pod \"collector-cnlrl\" (UID: \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\") " pod="openshift-logging/collector-cnlrl" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.827135 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/7d2b2ca1-f855-4921-ac74-71bf5e70345b-entrypoint\") pod \"collector-cnlrl\" (UID: \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\") " pod="openshift-logging/collector-cnlrl" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.827170 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/7d2b2ca1-f855-4921-ac74-71bf5e70345b-tmp\") pod \"collector-cnlrl\" (UID: \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\") " pod="openshift-logging/collector-cnlrl" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.827265 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/7d2b2ca1-f855-4921-ac74-71bf5e70345b-sa-token\") pod \"collector-cnlrl\" (UID: \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\") " pod="openshift-logging/collector-cnlrl" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.827329 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d2b2ca1-f855-4921-ac74-71bf5e70345b-config\") pod \"collector-cnlrl\" (UID: \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\") " pod="openshift-logging/collector-cnlrl" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.831196 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-trustbundle" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.846738 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-cnlrl"] Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.896009 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-logging/collector-cnlrl"] Dec 06 06:09:01 crc kubenswrapper[4809]: E1206 06:09:01.896670 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[collector-syslog-receiver collector-token config config-openshift-service-cacrt datadir entrypoint kube-api-access-8rtmw metrics sa-token tmp trusted-ca], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openshift-logging/collector-cnlrl" podUID="7d2b2ca1-f855-4921-ac74-71bf5e70345b" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.927994 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d2b2ca1-f855-4921-ac74-71bf5e70345b-config\") pod \"collector-cnlrl\" (UID: \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\") " pod="openshift-logging/collector-cnlrl" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.928051 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/7d2b2ca1-f855-4921-ac74-71bf5e70345b-collector-syslog-receiver\") pod \"collector-cnlrl\" (UID: \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\") " pod="openshift-logging/collector-cnlrl" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.928080 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/7d2b2ca1-f855-4921-ac74-71bf5e70345b-metrics\") pod \"collector-cnlrl\" (UID: \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\") " pod="openshift-logging/collector-cnlrl" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.928104 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/7d2b2ca1-f855-4921-ac74-71bf5e70345b-collector-token\") pod \"collector-cnlrl\" (UID: \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\") " pod="openshift-logging/collector-cnlrl" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.928149 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7d2b2ca1-f855-4921-ac74-71bf5e70345b-trusted-ca\") pod \"collector-cnlrl\" (UID: \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\") " pod="openshift-logging/collector-cnlrl" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.928174 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/7d2b2ca1-f855-4921-ac74-71bf5e70345b-config-openshift-service-cacrt\") pod \"collector-cnlrl\" (UID: \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\") " pod="openshift-logging/collector-cnlrl" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.928205 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/7d2b2ca1-f855-4921-ac74-71bf5e70345b-datadir\") pod \"collector-cnlrl\" (UID: \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\") " pod="openshift-logging/collector-cnlrl" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.928252 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rtmw\" (UniqueName: \"kubernetes.io/projected/7d2b2ca1-f855-4921-ac74-71bf5e70345b-kube-api-access-8rtmw\") pod \"collector-cnlrl\" (UID: \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\") " pod="openshift-logging/collector-cnlrl" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.928273 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/7d2b2ca1-f855-4921-ac74-71bf5e70345b-entrypoint\") pod \"collector-cnlrl\" (UID: \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\") " pod="openshift-logging/collector-cnlrl" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.928298 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/7d2b2ca1-f855-4921-ac74-71bf5e70345b-tmp\") pod \"collector-cnlrl\" (UID: \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\") " pod="openshift-logging/collector-cnlrl" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.928350 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/7d2b2ca1-f855-4921-ac74-71bf5e70345b-sa-token\") pod \"collector-cnlrl\" (UID: \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\") " pod="openshift-logging/collector-cnlrl" Dec 06 06:09:01 crc kubenswrapper[4809]: E1206 06:09:01.928654 4809 secret.go:188] Couldn't get secret openshift-logging/collector-metrics: secret "collector-metrics" not found Dec 06 06:09:01 crc kubenswrapper[4809]: E1206 06:09:01.928696 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7d2b2ca1-f855-4921-ac74-71bf5e70345b-metrics podName:7d2b2ca1-f855-4921-ac74-71bf5e70345b nodeName:}" failed. No retries permitted until 2025-12-06 06:09:02.428680542 +0000 UTC m=+1067.317663474 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics" (UniqueName: "kubernetes.io/secret/7d2b2ca1-f855-4921-ac74-71bf5e70345b-metrics") pod "collector-cnlrl" (UID: "7d2b2ca1-f855-4921-ac74-71bf5e70345b") : secret "collector-metrics" not found Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.929043 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d2b2ca1-f855-4921-ac74-71bf5e70345b-config\") pod \"collector-cnlrl\" (UID: \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\") " pod="openshift-logging/collector-cnlrl" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.930589 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/7d2b2ca1-f855-4921-ac74-71bf5e70345b-entrypoint\") pod \"collector-cnlrl\" (UID: \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\") " pod="openshift-logging/collector-cnlrl" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.930654 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/7d2b2ca1-f855-4921-ac74-71bf5e70345b-datadir\") pod \"collector-cnlrl\" (UID: \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\") " pod="openshift-logging/collector-cnlrl" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.930716 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/7d2b2ca1-f855-4921-ac74-71bf5e70345b-config-openshift-service-cacrt\") pod \"collector-cnlrl\" (UID: \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\") " pod="openshift-logging/collector-cnlrl" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.931092 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7d2b2ca1-f855-4921-ac74-71bf5e70345b-trusted-ca\") pod \"collector-cnlrl\" (UID: \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\") " pod="openshift-logging/collector-cnlrl" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.936902 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/7d2b2ca1-f855-4921-ac74-71bf5e70345b-collector-syslog-receiver\") pod \"collector-cnlrl\" (UID: \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\") " pod="openshift-logging/collector-cnlrl" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.937265 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/7d2b2ca1-f855-4921-ac74-71bf5e70345b-collector-token\") pod \"collector-cnlrl\" (UID: \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\") " pod="openshift-logging/collector-cnlrl" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.937310 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/7d2b2ca1-f855-4921-ac74-71bf5e70345b-tmp\") pod \"collector-cnlrl\" (UID: \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\") " pod="openshift-logging/collector-cnlrl" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.947139 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/7d2b2ca1-f855-4921-ac74-71bf5e70345b-sa-token\") pod \"collector-cnlrl\" (UID: \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\") " pod="openshift-logging/collector-cnlrl" Dec 06 06:09:01 crc kubenswrapper[4809]: I1206 06:09:01.950176 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rtmw\" (UniqueName: \"kubernetes.io/projected/7d2b2ca1-f855-4921-ac74-71bf5e70345b-kube-api-access-8rtmw\") pod \"collector-cnlrl\" (UID: \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\") " pod="openshift-logging/collector-cnlrl" Dec 06 06:09:02 crc kubenswrapper[4809]: I1206 06:09:02.235434 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-cnlrl" Dec 06 06:09:02 crc kubenswrapper[4809]: I1206 06:09:02.246323 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-cnlrl" Dec 06 06:09:02 crc kubenswrapper[4809]: I1206 06:09:02.433668 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/7d2b2ca1-f855-4921-ac74-71bf5e70345b-entrypoint\") pod \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\" (UID: \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\") " Dec 06 06:09:02 crc kubenswrapper[4809]: I1206 06:09:02.433789 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/7d2b2ca1-f855-4921-ac74-71bf5e70345b-datadir\") pod \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\" (UID: \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\") " Dec 06 06:09:02 crc kubenswrapper[4809]: I1206 06:09:02.433841 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d2b2ca1-f855-4921-ac74-71bf5e70345b-config\") pod \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\" (UID: \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\") " Dec 06 06:09:02 crc kubenswrapper[4809]: I1206 06:09:02.433920 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7d2b2ca1-f855-4921-ac74-71bf5e70345b-datadir" (OuterVolumeSpecName: "datadir") pod "7d2b2ca1-f855-4921-ac74-71bf5e70345b" (UID: "7d2b2ca1-f855-4921-ac74-71bf5e70345b"). InnerVolumeSpecName "datadir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:09:02 crc kubenswrapper[4809]: I1206 06:09:02.433969 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/7d2b2ca1-f855-4921-ac74-71bf5e70345b-collector-token\") pod \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\" (UID: \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\") " Dec 06 06:09:02 crc kubenswrapper[4809]: I1206 06:09:02.434016 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/7d2b2ca1-f855-4921-ac74-71bf5e70345b-collector-syslog-receiver\") pod \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\" (UID: \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\") " Dec 06 06:09:02 crc kubenswrapper[4809]: I1206 06:09:02.434047 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/7d2b2ca1-f855-4921-ac74-71bf5e70345b-sa-token\") pod \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\" (UID: \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\") " Dec 06 06:09:02 crc kubenswrapper[4809]: I1206 06:09:02.434125 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7d2b2ca1-f855-4921-ac74-71bf5e70345b-trusted-ca\") pod \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\" (UID: \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\") " Dec 06 06:09:02 crc kubenswrapper[4809]: I1206 06:09:02.434168 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rtmw\" (UniqueName: \"kubernetes.io/projected/7d2b2ca1-f855-4921-ac74-71bf5e70345b-kube-api-access-8rtmw\") pod \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\" (UID: \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\") " Dec 06 06:09:02 crc kubenswrapper[4809]: I1206 06:09:02.434206 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/7d2b2ca1-f855-4921-ac74-71bf5e70345b-tmp\") pod \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\" (UID: \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\") " Dec 06 06:09:02 crc kubenswrapper[4809]: I1206 06:09:02.434209 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d2b2ca1-f855-4921-ac74-71bf5e70345b-entrypoint" (OuterVolumeSpecName: "entrypoint") pod "7d2b2ca1-f855-4921-ac74-71bf5e70345b" (UID: "7d2b2ca1-f855-4921-ac74-71bf5e70345b"). InnerVolumeSpecName "entrypoint". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:09:02 crc kubenswrapper[4809]: I1206 06:09:02.434604 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d2b2ca1-f855-4921-ac74-71bf5e70345b-config" (OuterVolumeSpecName: "config") pod "7d2b2ca1-f855-4921-ac74-71bf5e70345b" (UID: "7d2b2ca1-f855-4921-ac74-71bf5e70345b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:09:02 crc kubenswrapper[4809]: I1206 06:09:02.434866 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d2b2ca1-f855-4921-ac74-71bf5e70345b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "7d2b2ca1-f855-4921-ac74-71bf5e70345b" (UID: "7d2b2ca1-f855-4921-ac74-71bf5e70345b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:09:02 crc kubenswrapper[4809]: I1206 06:09:02.434878 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/7d2b2ca1-f855-4921-ac74-71bf5e70345b-config-openshift-service-cacrt\") pod \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\" (UID: \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\") " Dec 06 06:09:02 crc kubenswrapper[4809]: I1206 06:09:02.435185 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d2b2ca1-f855-4921-ac74-71bf5e70345b-config-openshift-service-cacrt" (OuterVolumeSpecName: "config-openshift-service-cacrt") pod "7d2b2ca1-f855-4921-ac74-71bf5e70345b" (UID: "7d2b2ca1-f855-4921-ac74-71bf5e70345b"). InnerVolumeSpecName "config-openshift-service-cacrt". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:09:02 crc kubenswrapper[4809]: I1206 06:09:02.435493 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/7d2b2ca1-f855-4921-ac74-71bf5e70345b-metrics\") pod \"collector-cnlrl\" (UID: \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\") " pod="openshift-logging/collector-cnlrl" Dec 06 06:09:02 crc kubenswrapper[4809]: I1206 06:09:02.435771 4809 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7d2b2ca1-f855-4921-ac74-71bf5e70345b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 06:09:02 crc kubenswrapper[4809]: I1206 06:09:02.435791 4809 reconciler_common.go:293] "Volume detached for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/7d2b2ca1-f855-4921-ac74-71bf5e70345b-config-openshift-service-cacrt\") on node \"crc\" DevicePath \"\"" Dec 06 06:09:02 crc kubenswrapper[4809]: I1206 06:09:02.435801 4809 reconciler_common.go:293] "Volume detached for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/7d2b2ca1-f855-4921-ac74-71bf5e70345b-entrypoint\") on node \"crc\" DevicePath \"\"" Dec 06 06:09:02 crc kubenswrapper[4809]: I1206 06:09:02.435810 4809 reconciler_common.go:293] "Volume detached for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/7d2b2ca1-f855-4921-ac74-71bf5e70345b-datadir\") on node \"crc\" DevicePath \"\"" Dec 06 06:09:02 crc kubenswrapper[4809]: I1206 06:09:02.435819 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d2b2ca1-f855-4921-ac74-71bf5e70345b-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:09:02 crc kubenswrapper[4809]: I1206 06:09:02.437696 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d2b2ca1-f855-4921-ac74-71bf5e70345b-kube-api-access-8rtmw" (OuterVolumeSpecName: "kube-api-access-8rtmw") pod "7d2b2ca1-f855-4921-ac74-71bf5e70345b" (UID: "7d2b2ca1-f855-4921-ac74-71bf5e70345b"). InnerVolumeSpecName "kube-api-access-8rtmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:09:02 crc kubenswrapper[4809]: I1206 06:09:02.437956 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d2b2ca1-f855-4921-ac74-71bf5e70345b-collector-syslog-receiver" (OuterVolumeSpecName: "collector-syslog-receiver") pod "7d2b2ca1-f855-4921-ac74-71bf5e70345b" (UID: "7d2b2ca1-f855-4921-ac74-71bf5e70345b"). InnerVolumeSpecName "collector-syslog-receiver". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:09:02 crc kubenswrapper[4809]: I1206 06:09:02.438235 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d2b2ca1-f855-4921-ac74-71bf5e70345b-sa-token" (OuterVolumeSpecName: "sa-token") pod "7d2b2ca1-f855-4921-ac74-71bf5e70345b" (UID: "7d2b2ca1-f855-4921-ac74-71bf5e70345b"). InnerVolumeSpecName "sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:09:02 crc kubenswrapper[4809]: I1206 06:09:02.438391 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d2b2ca1-f855-4921-ac74-71bf5e70345b-tmp" (OuterVolumeSpecName: "tmp") pod "7d2b2ca1-f855-4921-ac74-71bf5e70345b" (UID: "7d2b2ca1-f855-4921-ac74-71bf5e70345b"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:09:02 crc kubenswrapper[4809]: I1206 06:09:02.438398 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d2b2ca1-f855-4921-ac74-71bf5e70345b-collector-token" (OuterVolumeSpecName: "collector-token") pod "7d2b2ca1-f855-4921-ac74-71bf5e70345b" (UID: "7d2b2ca1-f855-4921-ac74-71bf5e70345b"). InnerVolumeSpecName "collector-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:09:02 crc kubenswrapper[4809]: I1206 06:09:02.442528 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/7d2b2ca1-f855-4921-ac74-71bf5e70345b-metrics\") pod \"collector-cnlrl\" (UID: \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\") " pod="openshift-logging/collector-cnlrl" Dec 06 06:09:02 crc kubenswrapper[4809]: I1206 06:09:02.536687 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/7d2b2ca1-f855-4921-ac74-71bf5e70345b-metrics\") pod \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\" (UID: \"7d2b2ca1-f855-4921-ac74-71bf5e70345b\") " Dec 06 06:09:02 crc kubenswrapper[4809]: I1206 06:09:02.537200 4809 reconciler_common.go:293] "Volume detached for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/7d2b2ca1-f855-4921-ac74-71bf5e70345b-collector-token\") on node \"crc\" DevicePath \"\"" Dec 06 06:09:02 crc kubenswrapper[4809]: I1206 06:09:02.537228 4809 reconciler_common.go:293] "Volume detached for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/7d2b2ca1-f855-4921-ac74-71bf5e70345b-collector-syslog-receiver\") on node \"crc\" DevicePath \"\"" Dec 06 06:09:02 crc kubenswrapper[4809]: I1206 06:09:02.537243 4809 reconciler_common.go:293] "Volume detached for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/7d2b2ca1-f855-4921-ac74-71bf5e70345b-sa-token\") on node \"crc\" DevicePath \"\"" Dec 06 06:09:02 crc kubenswrapper[4809]: I1206 06:09:02.537257 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rtmw\" (UniqueName: \"kubernetes.io/projected/7d2b2ca1-f855-4921-ac74-71bf5e70345b-kube-api-access-8rtmw\") on node \"crc\" DevicePath \"\"" Dec 06 06:09:02 crc kubenswrapper[4809]: I1206 06:09:02.537268 4809 reconciler_common.go:293] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/7d2b2ca1-f855-4921-ac74-71bf5e70345b-tmp\") on node \"crc\" DevicePath \"\"" Dec 06 06:09:02 crc kubenswrapper[4809]: I1206 06:09:02.539209 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d2b2ca1-f855-4921-ac74-71bf5e70345b-metrics" (OuterVolumeSpecName: "metrics") pod "7d2b2ca1-f855-4921-ac74-71bf5e70345b" (UID: "7d2b2ca1-f855-4921-ac74-71bf5e70345b"). InnerVolumeSpecName "metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:09:02 crc kubenswrapper[4809]: I1206 06:09:02.638374 4809 reconciler_common.go:293] "Volume detached for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/7d2b2ca1-f855-4921-ac74-71bf5e70345b-metrics\") on node \"crc\" DevicePath \"\"" Dec 06 06:09:03 crc kubenswrapper[4809]: I1206 06:09:03.243225 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-cnlrl" Dec 06 06:09:03 crc kubenswrapper[4809]: I1206 06:09:03.314784 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-logging/collector-cnlrl"] Dec 06 06:09:03 crc kubenswrapper[4809]: I1206 06:09:03.321847 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-logging/collector-cnlrl"] Dec 06 06:09:03 crc kubenswrapper[4809]: I1206 06:09:03.328532 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/collector-hmjlt"] Dec 06 06:09:03 crc kubenswrapper[4809]: I1206 06:09:03.329852 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-hmjlt" Dec 06 06:09:03 crc kubenswrapper[4809]: I1206 06:09:03.332353 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-metrics" Dec 06 06:09:03 crc kubenswrapper[4809]: I1206 06:09:03.332777 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-dockercfg-4nb2f" Dec 06 06:09:03 crc kubenswrapper[4809]: I1206 06:09:03.334672 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-syslog-receiver" Dec 06 06:09:03 crc kubenswrapper[4809]: I1206 06:09:03.334907 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-config" Dec 06 06:09:03 crc kubenswrapper[4809]: I1206 06:09:03.335134 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-token" Dec 06 06:09:03 crc kubenswrapper[4809]: I1206 06:09:03.340475 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-trustbundle" Dec 06 06:09:03 crc kubenswrapper[4809]: I1206 06:09:03.379694 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-hmjlt"] Dec 06 06:09:03 crc kubenswrapper[4809]: I1206 06:09:03.397871 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d2b2ca1-f855-4921-ac74-71bf5e70345b" path="/var/lib/kubelet/pods/7d2b2ca1-f855-4921-ac74-71bf5e70345b/volumes" Dec 06 06:09:03 crc kubenswrapper[4809]: I1206 06:09:03.450633 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/1fe931cb-8318-4749-8be6-286c69da0860-entrypoint\") pod \"collector-hmjlt\" (UID: \"1fe931cb-8318-4749-8be6-286c69da0860\") " pod="openshift-logging/collector-hmjlt" Dec 06 06:09:03 crc kubenswrapper[4809]: I1206 06:09:03.450823 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/1fe931cb-8318-4749-8be6-286c69da0860-collector-syslog-receiver\") pod \"collector-hmjlt\" (UID: \"1fe931cb-8318-4749-8be6-286c69da0860\") " pod="openshift-logging/collector-hmjlt" Dec 06 06:09:03 crc kubenswrapper[4809]: I1206 06:09:03.450895 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/1fe931cb-8318-4749-8be6-286c69da0860-collector-token\") pod \"collector-hmjlt\" (UID: \"1fe931cb-8318-4749-8be6-286c69da0860\") " pod="openshift-logging/collector-hmjlt" Dec 06 06:09:03 crc kubenswrapper[4809]: I1206 06:09:03.450919 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/1fe931cb-8318-4749-8be6-286c69da0860-sa-token\") pod \"collector-hmjlt\" (UID: \"1fe931cb-8318-4749-8be6-286c69da0860\") " pod="openshift-logging/collector-hmjlt" Dec 06 06:09:03 crc kubenswrapper[4809]: I1206 06:09:03.451004 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1fe931cb-8318-4749-8be6-286c69da0860-trusted-ca\") pod \"collector-hmjlt\" (UID: \"1fe931cb-8318-4749-8be6-286c69da0860\") " pod="openshift-logging/collector-hmjlt" Dec 06 06:09:03 crc kubenswrapper[4809]: I1206 06:09:03.451040 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/1fe931cb-8318-4749-8be6-286c69da0860-metrics\") pod \"collector-hmjlt\" (UID: \"1fe931cb-8318-4749-8be6-286c69da0860\") " pod="openshift-logging/collector-hmjlt" Dec 06 06:09:03 crc kubenswrapper[4809]: I1206 06:09:03.451068 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/1fe931cb-8318-4749-8be6-286c69da0860-config-openshift-service-cacrt\") pod \"collector-hmjlt\" (UID: \"1fe931cb-8318-4749-8be6-286c69da0860\") " pod="openshift-logging/collector-hmjlt" Dec 06 06:09:03 crc kubenswrapper[4809]: I1206 06:09:03.451107 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/1fe931cb-8318-4749-8be6-286c69da0860-tmp\") pod \"collector-hmjlt\" (UID: \"1fe931cb-8318-4749-8be6-286c69da0860\") " pod="openshift-logging/collector-hmjlt" Dec 06 06:09:03 crc kubenswrapper[4809]: I1206 06:09:03.451129 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmcbw\" (UniqueName: \"kubernetes.io/projected/1fe931cb-8318-4749-8be6-286c69da0860-kube-api-access-pmcbw\") pod \"collector-hmjlt\" (UID: \"1fe931cb-8318-4749-8be6-286c69da0860\") " pod="openshift-logging/collector-hmjlt" Dec 06 06:09:03 crc kubenswrapper[4809]: I1206 06:09:03.451161 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fe931cb-8318-4749-8be6-286c69da0860-config\") pod \"collector-hmjlt\" (UID: \"1fe931cb-8318-4749-8be6-286c69da0860\") " pod="openshift-logging/collector-hmjlt" Dec 06 06:09:03 crc kubenswrapper[4809]: I1206 06:09:03.451213 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/1fe931cb-8318-4749-8be6-286c69da0860-datadir\") pod \"collector-hmjlt\" (UID: \"1fe931cb-8318-4749-8be6-286c69da0860\") " pod="openshift-logging/collector-hmjlt" Dec 06 06:09:03 crc kubenswrapper[4809]: I1206 06:09:03.553189 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1fe931cb-8318-4749-8be6-286c69da0860-trusted-ca\") pod \"collector-hmjlt\" (UID: \"1fe931cb-8318-4749-8be6-286c69da0860\") " pod="openshift-logging/collector-hmjlt" Dec 06 06:09:03 crc kubenswrapper[4809]: I1206 06:09:03.553283 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/1fe931cb-8318-4749-8be6-286c69da0860-metrics\") pod \"collector-hmjlt\" (UID: \"1fe931cb-8318-4749-8be6-286c69da0860\") " pod="openshift-logging/collector-hmjlt" Dec 06 06:09:03 crc kubenswrapper[4809]: I1206 06:09:03.553357 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/1fe931cb-8318-4749-8be6-286c69da0860-config-openshift-service-cacrt\") pod \"collector-hmjlt\" (UID: \"1fe931cb-8318-4749-8be6-286c69da0860\") " pod="openshift-logging/collector-hmjlt" Dec 06 06:09:03 crc kubenswrapper[4809]: I1206 06:09:03.553393 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/1fe931cb-8318-4749-8be6-286c69da0860-tmp\") pod \"collector-hmjlt\" (UID: \"1fe931cb-8318-4749-8be6-286c69da0860\") " pod="openshift-logging/collector-hmjlt" Dec 06 06:09:03 crc kubenswrapper[4809]: I1206 06:09:03.553424 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmcbw\" (UniqueName: \"kubernetes.io/projected/1fe931cb-8318-4749-8be6-286c69da0860-kube-api-access-pmcbw\") pod \"collector-hmjlt\" (UID: \"1fe931cb-8318-4749-8be6-286c69da0860\") " pod="openshift-logging/collector-hmjlt" Dec 06 06:09:03 crc kubenswrapper[4809]: I1206 06:09:03.553470 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fe931cb-8318-4749-8be6-286c69da0860-config\") pod \"collector-hmjlt\" (UID: \"1fe931cb-8318-4749-8be6-286c69da0860\") " pod="openshift-logging/collector-hmjlt" Dec 06 06:09:03 crc kubenswrapper[4809]: I1206 06:09:03.553509 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/1fe931cb-8318-4749-8be6-286c69da0860-datadir\") pod \"collector-hmjlt\" (UID: \"1fe931cb-8318-4749-8be6-286c69da0860\") " pod="openshift-logging/collector-hmjlt" Dec 06 06:09:03 crc kubenswrapper[4809]: I1206 06:09:03.553583 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/1fe931cb-8318-4749-8be6-286c69da0860-entrypoint\") pod \"collector-hmjlt\" (UID: \"1fe931cb-8318-4749-8be6-286c69da0860\") " pod="openshift-logging/collector-hmjlt" Dec 06 06:09:03 crc kubenswrapper[4809]: I1206 06:09:03.553619 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/1fe931cb-8318-4749-8be6-286c69da0860-collector-syslog-receiver\") pod \"collector-hmjlt\" (UID: \"1fe931cb-8318-4749-8be6-286c69da0860\") " pod="openshift-logging/collector-hmjlt" Dec 06 06:09:03 crc kubenswrapper[4809]: I1206 06:09:03.553691 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/1fe931cb-8318-4749-8be6-286c69da0860-collector-token\") pod \"collector-hmjlt\" (UID: \"1fe931cb-8318-4749-8be6-286c69da0860\") " pod="openshift-logging/collector-hmjlt" Dec 06 06:09:03 crc kubenswrapper[4809]: I1206 06:09:03.553725 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/1fe931cb-8318-4749-8be6-286c69da0860-sa-token\") pod \"collector-hmjlt\" (UID: \"1fe931cb-8318-4749-8be6-286c69da0860\") " pod="openshift-logging/collector-hmjlt" Dec 06 06:09:03 crc kubenswrapper[4809]: I1206 06:09:03.554267 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1fe931cb-8318-4749-8be6-286c69da0860-trusted-ca\") pod \"collector-hmjlt\" (UID: \"1fe931cb-8318-4749-8be6-286c69da0860\") " pod="openshift-logging/collector-hmjlt" Dec 06 06:09:03 crc kubenswrapper[4809]: I1206 06:09:03.554366 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/1fe931cb-8318-4749-8be6-286c69da0860-datadir\") pod \"collector-hmjlt\" (UID: \"1fe931cb-8318-4749-8be6-286c69da0860\") " pod="openshift-logging/collector-hmjlt" Dec 06 06:09:03 crc kubenswrapper[4809]: I1206 06:09:03.554790 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/1fe931cb-8318-4749-8be6-286c69da0860-entrypoint\") pod \"collector-hmjlt\" (UID: \"1fe931cb-8318-4749-8be6-286c69da0860\") " pod="openshift-logging/collector-hmjlt" Dec 06 06:09:03 crc kubenswrapper[4809]: I1206 06:09:03.554800 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/1fe931cb-8318-4749-8be6-286c69da0860-config-openshift-service-cacrt\") pod \"collector-hmjlt\" (UID: \"1fe931cb-8318-4749-8be6-286c69da0860\") " pod="openshift-logging/collector-hmjlt" Dec 06 06:09:03 crc kubenswrapper[4809]: I1206 06:09:03.555787 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fe931cb-8318-4749-8be6-286c69da0860-config\") pod \"collector-hmjlt\" (UID: \"1fe931cb-8318-4749-8be6-286c69da0860\") " pod="openshift-logging/collector-hmjlt" Dec 06 06:09:03 crc kubenswrapper[4809]: I1206 06:09:03.557490 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/1fe931cb-8318-4749-8be6-286c69da0860-collector-token\") pod \"collector-hmjlt\" (UID: \"1fe931cb-8318-4749-8be6-286c69da0860\") " pod="openshift-logging/collector-hmjlt" Dec 06 06:09:03 crc kubenswrapper[4809]: I1206 06:09:03.557499 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/1fe931cb-8318-4749-8be6-286c69da0860-metrics\") pod \"collector-hmjlt\" (UID: \"1fe931cb-8318-4749-8be6-286c69da0860\") " pod="openshift-logging/collector-hmjlt" Dec 06 06:09:03 crc kubenswrapper[4809]: I1206 06:09:03.562523 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/1fe931cb-8318-4749-8be6-286c69da0860-collector-syslog-receiver\") pod \"collector-hmjlt\" (UID: \"1fe931cb-8318-4749-8be6-286c69da0860\") " pod="openshift-logging/collector-hmjlt" Dec 06 06:09:03 crc kubenswrapper[4809]: I1206 06:09:03.569353 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/1fe931cb-8318-4749-8be6-286c69da0860-tmp\") pod \"collector-hmjlt\" (UID: \"1fe931cb-8318-4749-8be6-286c69da0860\") " pod="openshift-logging/collector-hmjlt" Dec 06 06:09:03 crc kubenswrapper[4809]: I1206 06:09:03.576187 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmcbw\" (UniqueName: \"kubernetes.io/projected/1fe931cb-8318-4749-8be6-286c69da0860-kube-api-access-pmcbw\") pod \"collector-hmjlt\" (UID: \"1fe931cb-8318-4749-8be6-286c69da0860\") " pod="openshift-logging/collector-hmjlt" Dec 06 06:09:03 crc kubenswrapper[4809]: I1206 06:09:03.576907 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/1fe931cb-8318-4749-8be6-286c69da0860-sa-token\") pod \"collector-hmjlt\" (UID: \"1fe931cb-8318-4749-8be6-286c69da0860\") " pod="openshift-logging/collector-hmjlt" Dec 06 06:09:03 crc kubenswrapper[4809]: I1206 06:09:03.661420 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-hmjlt" Dec 06 06:09:04 crc kubenswrapper[4809]: I1206 06:09:04.085757 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-hmjlt"] Dec 06 06:09:04 crc kubenswrapper[4809]: W1206 06:09:04.093196 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1fe931cb_8318_4749_8be6_286c69da0860.slice/crio-1cb252247b26b42d92e6f47c11be5f3af581e704b9f7452cf51214d7a53f1cad WatchSource:0}: Error finding container 1cb252247b26b42d92e6f47c11be5f3af581e704b9f7452cf51214d7a53f1cad: Status 404 returned error can't find the container with id 1cb252247b26b42d92e6f47c11be5f3af581e704b9f7452cf51214d7a53f1cad Dec 06 06:09:04 crc kubenswrapper[4809]: I1206 06:09:04.254395 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/collector-hmjlt" event={"ID":"1fe931cb-8318-4749-8be6-286c69da0860","Type":"ContainerStarted","Data":"1cb252247b26b42d92e6f47c11be5f3af581e704b9f7452cf51214d7a53f1cad"} Dec 06 06:09:04 crc kubenswrapper[4809]: I1206 06:09:04.496699 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:09:04 crc kubenswrapper[4809]: I1206 06:09:04.497110 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:09:04 crc kubenswrapper[4809]: I1206 06:09:04.497164 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-npms2" Dec 06 06:09:04 crc kubenswrapper[4809]: I1206 06:09:04.498107 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bb59482414fdfbb9268b139562a808d2a5e04bbf56acc96dd715da16e5bd5913"} pod="openshift-machine-config-operator/machine-config-daemon-npms2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 06:09:04 crc kubenswrapper[4809]: I1206 06:09:04.498202 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" containerID="cri-o://bb59482414fdfbb9268b139562a808d2a5e04bbf56acc96dd715da16e5bd5913" gracePeriod=600 Dec 06 06:09:05 crc kubenswrapper[4809]: I1206 06:09:05.266256 4809 generic.go:334] "Generic (PLEG): container finished" podID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerID="bb59482414fdfbb9268b139562a808d2a5e04bbf56acc96dd715da16e5bd5913" exitCode=0 Dec 06 06:09:05 crc kubenswrapper[4809]: I1206 06:09:05.266343 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" event={"ID":"cbf4cf62-024e-4703-a8b8-9aecda9cd26a","Type":"ContainerDied","Data":"bb59482414fdfbb9268b139562a808d2a5e04bbf56acc96dd715da16e5bd5913"} Dec 06 06:09:05 crc kubenswrapper[4809]: I1206 06:09:05.266681 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" event={"ID":"cbf4cf62-024e-4703-a8b8-9aecda9cd26a","Type":"ContainerStarted","Data":"9b0b6808cda0fc2270940d0650cf854006725ea29b339e76c18a025e4f361d32"} Dec 06 06:09:05 crc kubenswrapper[4809]: I1206 06:09:05.266708 4809 scope.go:117] "RemoveContainer" containerID="49ef8bff82d74abdc9d792d5696488d1189b4c9562bc3eb566b2639019bb757d" Dec 06 06:09:12 crc kubenswrapper[4809]: I1206 06:09:12.325410 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/collector-hmjlt" event={"ID":"1fe931cb-8318-4749-8be6-286c69da0860","Type":"ContainerStarted","Data":"75d29f2842e4eeeb81ca21ba96a0857617433d1c364a3c37d6703502ba42f89a"} Dec 06 06:09:12 crc kubenswrapper[4809]: I1206 06:09:12.351722 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/collector-hmjlt" podStartSLOduration=1.709047843 podStartE2EDuration="9.351695334s" podCreationTimestamp="2025-12-06 06:09:03 +0000 UTC" firstStartedPulling="2025-12-06 06:09:04.107658786 +0000 UTC m=+1068.996641728" lastFinishedPulling="2025-12-06 06:09:11.750306277 +0000 UTC m=+1076.639289219" observedRunningTime="2025-12-06 06:09:12.345922808 +0000 UTC m=+1077.234905750" watchObservedRunningTime="2025-12-06 06:09:12.351695334 +0000 UTC m=+1077.240678276" Dec 06 06:09:41 crc kubenswrapper[4809]: I1206 06:09:41.848154 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwt88"] Dec 06 06:09:41 crc kubenswrapper[4809]: I1206 06:09:41.851188 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwt88" Dec 06 06:09:41 crc kubenswrapper[4809]: I1206 06:09:41.857654 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwt88"] Dec 06 06:09:41 crc kubenswrapper[4809]: I1206 06:09:41.858557 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 06 06:09:41 crc kubenswrapper[4809]: I1206 06:09:41.991735 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzfc5\" (UniqueName: \"kubernetes.io/projected/d8953b5f-9e44-407f-aab5-6a434e11a7ef-kube-api-access-rzfc5\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwt88\" (UID: \"d8953b5f-9e44-407f-aab5-6a434e11a7ef\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwt88" Dec 06 06:09:41 crc kubenswrapper[4809]: I1206 06:09:41.991828 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d8953b5f-9e44-407f-aab5-6a434e11a7ef-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwt88\" (UID: \"d8953b5f-9e44-407f-aab5-6a434e11a7ef\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwt88" Dec 06 06:09:41 crc kubenswrapper[4809]: I1206 06:09:41.991866 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d8953b5f-9e44-407f-aab5-6a434e11a7ef-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwt88\" (UID: \"d8953b5f-9e44-407f-aab5-6a434e11a7ef\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwt88" Dec 06 06:09:42 crc kubenswrapper[4809]: I1206 06:09:42.094021 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzfc5\" (UniqueName: \"kubernetes.io/projected/d8953b5f-9e44-407f-aab5-6a434e11a7ef-kube-api-access-rzfc5\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwt88\" (UID: \"d8953b5f-9e44-407f-aab5-6a434e11a7ef\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwt88" Dec 06 06:09:42 crc kubenswrapper[4809]: I1206 06:09:42.094112 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d8953b5f-9e44-407f-aab5-6a434e11a7ef-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwt88\" (UID: \"d8953b5f-9e44-407f-aab5-6a434e11a7ef\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwt88" Dec 06 06:09:42 crc kubenswrapper[4809]: I1206 06:09:42.094140 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d8953b5f-9e44-407f-aab5-6a434e11a7ef-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwt88\" (UID: \"d8953b5f-9e44-407f-aab5-6a434e11a7ef\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwt88" Dec 06 06:09:42 crc kubenswrapper[4809]: I1206 06:09:42.094607 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d8953b5f-9e44-407f-aab5-6a434e11a7ef-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwt88\" (UID: \"d8953b5f-9e44-407f-aab5-6a434e11a7ef\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwt88" Dec 06 06:09:42 crc kubenswrapper[4809]: I1206 06:09:42.094744 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d8953b5f-9e44-407f-aab5-6a434e11a7ef-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwt88\" (UID: \"d8953b5f-9e44-407f-aab5-6a434e11a7ef\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwt88" Dec 06 06:09:42 crc kubenswrapper[4809]: I1206 06:09:42.115976 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzfc5\" (UniqueName: \"kubernetes.io/projected/d8953b5f-9e44-407f-aab5-6a434e11a7ef-kube-api-access-rzfc5\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwt88\" (UID: \"d8953b5f-9e44-407f-aab5-6a434e11a7ef\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwt88" Dec 06 06:09:42 crc kubenswrapper[4809]: I1206 06:09:42.173909 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwt88" Dec 06 06:09:42 crc kubenswrapper[4809]: I1206 06:09:42.575449 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwt88"] Dec 06 06:09:42 crc kubenswrapper[4809]: I1206 06:09:42.644855 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwt88" event={"ID":"d8953b5f-9e44-407f-aab5-6a434e11a7ef","Type":"ContainerStarted","Data":"38a36fe3782c37d3f7d85de4c6abb4a3c7ec7cc16e742a4f1faf26d51c8b52ae"} Dec 06 06:09:43 crc kubenswrapper[4809]: I1206 06:09:43.652963 4809 generic.go:334] "Generic (PLEG): container finished" podID="d8953b5f-9e44-407f-aab5-6a434e11a7ef" containerID="80a2119c8dc0b42c677f39edc3f8412624db90186495fcb4346be9fb7b8f4f53" exitCode=0 Dec 06 06:09:43 crc kubenswrapper[4809]: I1206 06:09:43.653018 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwt88" event={"ID":"d8953b5f-9e44-407f-aab5-6a434e11a7ef","Type":"ContainerDied","Data":"80a2119c8dc0b42c677f39edc3f8412624db90186495fcb4346be9fb7b8f4f53"} Dec 06 06:09:46 crc kubenswrapper[4809]: I1206 06:09:46.679803 4809 generic.go:334] "Generic (PLEG): container finished" podID="d8953b5f-9e44-407f-aab5-6a434e11a7ef" containerID="44e7b1d9a0ffd22e830226d0231c490013711717889e909960fae39d95481429" exitCode=0 Dec 06 06:09:46 crc kubenswrapper[4809]: I1206 06:09:46.680466 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwt88" event={"ID":"d8953b5f-9e44-407f-aab5-6a434e11a7ef","Type":"ContainerDied","Data":"44e7b1d9a0ffd22e830226d0231c490013711717889e909960fae39d95481429"} Dec 06 06:09:47 crc kubenswrapper[4809]: I1206 06:09:47.690508 4809 generic.go:334] "Generic (PLEG): container finished" podID="d8953b5f-9e44-407f-aab5-6a434e11a7ef" containerID="3c9b99003faaaa71a48281c18e1cd5e6cf02297acd0b2c83ba73b3dd5b78562c" exitCode=0 Dec 06 06:09:47 crc kubenswrapper[4809]: I1206 06:09:47.690609 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwt88" event={"ID":"d8953b5f-9e44-407f-aab5-6a434e11a7ef","Type":"ContainerDied","Data":"3c9b99003faaaa71a48281c18e1cd5e6cf02297acd0b2c83ba73b3dd5b78562c"} Dec 06 06:09:49 crc kubenswrapper[4809]: I1206 06:09:49.108677 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwt88" Dec 06 06:09:49 crc kubenswrapper[4809]: I1206 06:09:49.228331 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d8953b5f-9e44-407f-aab5-6a434e11a7ef-util\") pod \"d8953b5f-9e44-407f-aab5-6a434e11a7ef\" (UID: \"d8953b5f-9e44-407f-aab5-6a434e11a7ef\") " Dec 06 06:09:49 crc kubenswrapper[4809]: I1206 06:09:49.228413 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d8953b5f-9e44-407f-aab5-6a434e11a7ef-bundle\") pod \"d8953b5f-9e44-407f-aab5-6a434e11a7ef\" (UID: \"d8953b5f-9e44-407f-aab5-6a434e11a7ef\") " Dec 06 06:09:49 crc kubenswrapper[4809]: I1206 06:09:49.228647 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rzfc5\" (UniqueName: \"kubernetes.io/projected/d8953b5f-9e44-407f-aab5-6a434e11a7ef-kube-api-access-rzfc5\") pod \"d8953b5f-9e44-407f-aab5-6a434e11a7ef\" (UID: \"d8953b5f-9e44-407f-aab5-6a434e11a7ef\") " Dec 06 06:09:49 crc kubenswrapper[4809]: I1206 06:09:49.229502 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8953b5f-9e44-407f-aab5-6a434e11a7ef-bundle" (OuterVolumeSpecName: "bundle") pod "d8953b5f-9e44-407f-aab5-6a434e11a7ef" (UID: "d8953b5f-9e44-407f-aab5-6a434e11a7ef"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:09:49 crc kubenswrapper[4809]: I1206 06:09:49.241014 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8953b5f-9e44-407f-aab5-6a434e11a7ef-util" (OuterVolumeSpecName: "util") pod "d8953b5f-9e44-407f-aab5-6a434e11a7ef" (UID: "d8953b5f-9e44-407f-aab5-6a434e11a7ef"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:09:49 crc kubenswrapper[4809]: I1206 06:09:49.249782 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8953b5f-9e44-407f-aab5-6a434e11a7ef-kube-api-access-rzfc5" (OuterVolumeSpecName: "kube-api-access-rzfc5") pod "d8953b5f-9e44-407f-aab5-6a434e11a7ef" (UID: "d8953b5f-9e44-407f-aab5-6a434e11a7ef"). InnerVolumeSpecName "kube-api-access-rzfc5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:09:49 crc kubenswrapper[4809]: I1206 06:09:49.331637 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rzfc5\" (UniqueName: \"kubernetes.io/projected/d8953b5f-9e44-407f-aab5-6a434e11a7ef-kube-api-access-rzfc5\") on node \"crc\" DevicePath \"\"" Dec 06 06:09:49 crc kubenswrapper[4809]: I1206 06:09:49.332187 4809 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d8953b5f-9e44-407f-aab5-6a434e11a7ef-util\") on node \"crc\" DevicePath \"\"" Dec 06 06:09:49 crc kubenswrapper[4809]: I1206 06:09:49.332206 4809 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d8953b5f-9e44-407f-aab5-6a434e11a7ef-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:09:49 crc kubenswrapper[4809]: I1206 06:09:49.718273 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwt88" event={"ID":"d8953b5f-9e44-407f-aab5-6a434e11a7ef","Type":"ContainerDied","Data":"38a36fe3782c37d3f7d85de4c6abb4a3c7ec7cc16e742a4f1faf26d51c8b52ae"} Dec 06 06:09:49 crc kubenswrapper[4809]: I1206 06:09:49.718338 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwt88" Dec 06 06:09:49 crc kubenswrapper[4809]: I1206 06:09:49.718348 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="38a36fe3782c37d3f7d85de4c6abb4a3c7ec7cc16e742a4f1faf26d51c8b52ae" Dec 06 06:09:54 crc kubenswrapper[4809]: I1206 06:09:54.107174 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-t82p5"] Dec 06 06:09:54 crc kubenswrapper[4809]: E1206 06:09:54.110174 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8953b5f-9e44-407f-aab5-6a434e11a7ef" containerName="util" Dec 06 06:09:54 crc kubenswrapper[4809]: I1206 06:09:54.110333 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8953b5f-9e44-407f-aab5-6a434e11a7ef" containerName="util" Dec 06 06:09:54 crc kubenswrapper[4809]: E1206 06:09:54.110470 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8953b5f-9e44-407f-aab5-6a434e11a7ef" containerName="pull" Dec 06 06:09:54 crc kubenswrapper[4809]: I1206 06:09:54.110556 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8953b5f-9e44-407f-aab5-6a434e11a7ef" containerName="pull" Dec 06 06:09:54 crc kubenswrapper[4809]: E1206 06:09:54.110628 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8953b5f-9e44-407f-aab5-6a434e11a7ef" containerName="extract" Dec 06 06:09:54 crc kubenswrapper[4809]: I1206 06:09:54.110688 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8953b5f-9e44-407f-aab5-6a434e11a7ef" containerName="extract" Dec 06 06:09:54 crc kubenswrapper[4809]: I1206 06:09:54.110883 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8953b5f-9e44-407f-aab5-6a434e11a7ef" containerName="extract" Dec 06 06:09:54 crc kubenswrapper[4809]: I1206 06:09:54.111816 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-t82p5" Dec 06 06:09:54 crc kubenswrapper[4809]: I1206 06:09:54.114600 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-5nk9c" Dec 06 06:09:54 crc kubenswrapper[4809]: I1206 06:09:54.117770 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 06 06:09:54 crc kubenswrapper[4809]: I1206 06:09:54.120291 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 06 06:09:54 crc kubenswrapper[4809]: I1206 06:09:54.133606 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-t82p5"] Dec 06 06:09:54 crc kubenswrapper[4809]: I1206 06:09:54.221192 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wgsw\" (UniqueName: \"kubernetes.io/projected/b86e935d-b651-435e-8218-f39878ec04b9-kube-api-access-2wgsw\") pod \"nmstate-operator-5b5b58f5c8-t82p5\" (UID: \"b86e935d-b651-435e-8218-f39878ec04b9\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-t82p5" Dec 06 06:09:54 crc kubenswrapper[4809]: I1206 06:09:54.324051 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wgsw\" (UniqueName: \"kubernetes.io/projected/b86e935d-b651-435e-8218-f39878ec04b9-kube-api-access-2wgsw\") pod \"nmstate-operator-5b5b58f5c8-t82p5\" (UID: \"b86e935d-b651-435e-8218-f39878ec04b9\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-t82p5" Dec 06 06:09:54 crc kubenswrapper[4809]: I1206 06:09:54.351236 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wgsw\" (UniqueName: \"kubernetes.io/projected/b86e935d-b651-435e-8218-f39878ec04b9-kube-api-access-2wgsw\") pod \"nmstate-operator-5b5b58f5c8-t82p5\" (UID: \"b86e935d-b651-435e-8218-f39878ec04b9\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-t82p5" Dec 06 06:09:54 crc kubenswrapper[4809]: I1206 06:09:54.438550 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-t82p5" Dec 06 06:09:55 crc kubenswrapper[4809]: I1206 06:09:55.113490 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-t82p5"] Dec 06 06:09:55 crc kubenswrapper[4809]: I1206 06:09:55.839442 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-t82p5" event={"ID":"b86e935d-b651-435e-8218-f39878ec04b9","Type":"ContainerStarted","Data":"9d2420e934279cc83f4eb52558898637a5043a0960c764bdae620ecf698ae855"} Dec 06 06:09:57 crc kubenswrapper[4809]: I1206 06:09:57.868975 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-t82p5" event={"ID":"b86e935d-b651-435e-8218-f39878ec04b9","Type":"ContainerStarted","Data":"1b338e5d3d400e727db3cba521e5ebe087f2bcaed62704e9342c7a1ce38ff116"} Dec 06 06:09:57 crc kubenswrapper[4809]: I1206 06:09:57.892327 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-t82p5" podStartSLOduration=1.958574401 podStartE2EDuration="3.892304665s" podCreationTimestamp="2025-12-06 06:09:54 +0000 UTC" firstStartedPulling="2025-12-06 06:09:55.121288712 +0000 UTC m=+1120.010271654" lastFinishedPulling="2025-12-06 06:09:57.055018976 +0000 UTC m=+1121.944001918" observedRunningTime="2025-12-06 06:09:57.892045888 +0000 UTC m=+1122.781028830" watchObservedRunningTime="2025-12-06 06:09:57.892304665 +0000 UTC m=+1122.781287607" Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.248463 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-pvgrd"] Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.251515 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-pvgrd" Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.262654 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-5rhsc" Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.277727 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-pvgrd"] Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.320106 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmg78\" (UniqueName: \"kubernetes.io/projected/d39362b1-ae5e-4852-be43-a335be68151b-kube-api-access-pmg78\") pod \"nmstate-metrics-7f946cbc9-pvgrd\" (UID: \"d39362b1-ae5e-4852-be43-a335be68151b\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-pvgrd" Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.344838 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pw6wx"] Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.346520 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pw6wx" Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.350216 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.364012 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-jbb6s"] Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.366716 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-jbb6s" Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.413617 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pw6wx"] Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.422349 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/50b406f3-bc91-492f-89ea-8eeae9ae132e-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-pw6wx\" (UID: \"50b406f3-bc91-492f-89ea-8eeae9ae132e\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pw6wx" Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.422434 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h72xx\" (UniqueName: \"kubernetes.io/projected/57d09bf3-cccc-4745-834d-57eaad3f96a1-kube-api-access-h72xx\") pod \"nmstate-handler-jbb6s\" (UID: \"57d09bf3-cccc-4745-834d-57eaad3f96a1\") " pod="openshift-nmstate/nmstate-handler-jbb6s" Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.422481 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmg78\" (UniqueName: \"kubernetes.io/projected/d39362b1-ae5e-4852-be43-a335be68151b-kube-api-access-pmg78\") pod \"nmstate-metrics-7f946cbc9-pvgrd\" (UID: \"d39362b1-ae5e-4852-be43-a335be68151b\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-pvgrd" Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.422587 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/57d09bf3-cccc-4745-834d-57eaad3f96a1-dbus-socket\") pod \"nmstate-handler-jbb6s\" (UID: \"57d09bf3-cccc-4745-834d-57eaad3f96a1\") " pod="openshift-nmstate/nmstate-handler-jbb6s" Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.422632 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/57d09bf3-cccc-4745-834d-57eaad3f96a1-nmstate-lock\") pod \"nmstate-handler-jbb6s\" (UID: \"57d09bf3-cccc-4745-834d-57eaad3f96a1\") " pod="openshift-nmstate/nmstate-handler-jbb6s" Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.422693 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/57d09bf3-cccc-4745-834d-57eaad3f96a1-ovs-socket\") pod \"nmstate-handler-jbb6s\" (UID: \"57d09bf3-cccc-4745-834d-57eaad3f96a1\") " pod="openshift-nmstate/nmstate-handler-jbb6s" Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.422724 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95vkm\" (UniqueName: \"kubernetes.io/projected/50b406f3-bc91-492f-89ea-8eeae9ae132e-kube-api-access-95vkm\") pod \"nmstate-webhook-5f6d4c5ccb-pw6wx\" (UID: \"50b406f3-bc91-492f-89ea-8eeae9ae132e\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pw6wx" Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.470439 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmg78\" (UniqueName: \"kubernetes.io/projected/d39362b1-ae5e-4852-be43-a335be68151b-kube-api-access-pmg78\") pod \"nmstate-metrics-7f946cbc9-pvgrd\" (UID: \"d39362b1-ae5e-4852-be43-a335be68151b\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-pvgrd" Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.526069 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/50b406f3-bc91-492f-89ea-8eeae9ae132e-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-pw6wx\" (UID: \"50b406f3-bc91-492f-89ea-8eeae9ae132e\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pw6wx" Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.526142 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h72xx\" (UniqueName: \"kubernetes.io/projected/57d09bf3-cccc-4745-834d-57eaad3f96a1-kube-api-access-h72xx\") pod \"nmstate-handler-jbb6s\" (UID: \"57d09bf3-cccc-4745-834d-57eaad3f96a1\") " pod="openshift-nmstate/nmstate-handler-jbb6s" Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.526185 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/57d09bf3-cccc-4745-834d-57eaad3f96a1-dbus-socket\") pod \"nmstate-handler-jbb6s\" (UID: \"57d09bf3-cccc-4745-834d-57eaad3f96a1\") " pod="openshift-nmstate/nmstate-handler-jbb6s" Dec 06 06:10:03 crc kubenswrapper[4809]: E1206 06:10:03.526356 4809 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Dec 06 06:10:03 crc kubenswrapper[4809]: E1206 06:10:03.526431 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/50b406f3-bc91-492f-89ea-8eeae9ae132e-tls-key-pair podName:50b406f3-bc91-492f-89ea-8eeae9ae132e nodeName:}" failed. No retries permitted until 2025-12-06 06:10:04.026397037 +0000 UTC m=+1128.915379979 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/50b406f3-bc91-492f-89ea-8eeae9ae132e-tls-key-pair") pod "nmstate-webhook-5f6d4c5ccb-pw6wx" (UID: "50b406f3-bc91-492f-89ea-8eeae9ae132e") : secret "openshift-nmstate-webhook" not found Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.527395 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/57d09bf3-cccc-4745-834d-57eaad3f96a1-dbus-socket\") pod \"nmstate-handler-jbb6s\" (UID: \"57d09bf3-cccc-4745-834d-57eaad3f96a1\") " pod="openshift-nmstate/nmstate-handler-jbb6s" Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.527492 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/57d09bf3-cccc-4745-834d-57eaad3f96a1-nmstate-lock\") pod \"nmstate-handler-jbb6s\" (UID: \"57d09bf3-cccc-4745-834d-57eaad3f96a1\") " pod="openshift-nmstate/nmstate-handler-jbb6s" Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.527548 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/57d09bf3-cccc-4745-834d-57eaad3f96a1-ovs-socket\") pod \"nmstate-handler-jbb6s\" (UID: \"57d09bf3-cccc-4745-834d-57eaad3f96a1\") " pod="openshift-nmstate/nmstate-handler-jbb6s" Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.527586 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95vkm\" (UniqueName: \"kubernetes.io/projected/50b406f3-bc91-492f-89ea-8eeae9ae132e-kube-api-access-95vkm\") pod \"nmstate-webhook-5f6d4c5ccb-pw6wx\" (UID: \"50b406f3-bc91-492f-89ea-8eeae9ae132e\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pw6wx" Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.527877 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/57d09bf3-cccc-4745-834d-57eaad3f96a1-nmstate-lock\") pod \"nmstate-handler-jbb6s\" (UID: \"57d09bf3-cccc-4745-834d-57eaad3f96a1\") " pod="openshift-nmstate/nmstate-handler-jbb6s" Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.527915 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/57d09bf3-cccc-4745-834d-57eaad3f96a1-ovs-socket\") pod \"nmstate-handler-jbb6s\" (UID: \"57d09bf3-cccc-4745-834d-57eaad3f96a1\") " pod="openshift-nmstate/nmstate-handler-jbb6s" Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.548027 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7dpj4"] Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.549511 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7dpj4" Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.559952 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-9tszw" Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.561219 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.561528 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.563054 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h72xx\" (UniqueName: \"kubernetes.io/projected/57d09bf3-cccc-4745-834d-57eaad3f96a1-kube-api-access-h72xx\") pod \"nmstate-handler-jbb6s\" (UID: \"57d09bf3-cccc-4745-834d-57eaad3f96a1\") " pod="openshift-nmstate/nmstate-handler-jbb6s" Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.573099 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7dpj4"] Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.575583 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-pvgrd" Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.595950 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95vkm\" (UniqueName: \"kubernetes.io/projected/50b406f3-bc91-492f-89ea-8eeae9ae132e-kube-api-access-95vkm\") pod \"nmstate-webhook-5f6d4c5ccb-pw6wx\" (UID: \"50b406f3-bc91-492f-89ea-8eeae9ae132e\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pw6wx" Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.632776 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/87deabac-9840-43a3-945b-cb4d054e34ba-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-7dpj4\" (UID: \"87deabac-9840-43a3-945b-cb4d054e34ba\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7dpj4" Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.633348 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xzns\" (UniqueName: \"kubernetes.io/projected/87deabac-9840-43a3-945b-cb4d054e34ba-kube-api-access-8xzns\") pod \"nmstate-console-plugin-7fbb5f6569-7dpj4\" (UID: \"87deabac-9840-43a3-945b-cb4d054e34ba\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7dpj4" Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.633386 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/87deabac-9840-43a3-945b-cb4d054e34ba-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-7dpj4\" (UID: \"87deabac-9840-43a3-945b-cb4d054e34ba\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7dpj4" Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.694403 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-jbb6s" Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.748468 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xzns\" (UniqueName: \"kubernetes.io/projected/87deabac-9840-43a3-945b-cb4d054e34ba-kube-api-access-8xzns\") pod \"nmstate-console-plugin-7fbb5f6569-7dpj4\" (UID: \"87deabac-9840-43a3-945b-cb4d054e34ba\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7dpj4" Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.748524 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/87deabac-9840-43a3-945b-cb4d054e34ba-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-7dpj4\" (UID: \"87deabac-9840-43a3-945b-cb4d054e34ba\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7dpj4" Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.748599 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/87deabac-9840-43a3-945b-cb4d054e34ba-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-7dpj4\" (UID: \"87deabac-9840-43a3-945b-cb4d054e34ba\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7dpj4" Dec 06 06:10:03 crc kubenswrapper[4809]: E1206 06:10:03.748874 4809 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Dec 06 06:10:03 crc kubenswrapper[4809]: E1206 06:10:03.748950 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/87deabac-9840-43a3-945b-cb4d054e34ba-plugin-serving-cert podName:87deabac-9840-43a3-945b-cb4d054e34ba nodeName:}" failed. No retries permitted until 2025-12-06 06:10:04.248918532 +0000 UTC m=+1129.137901474 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/87deabac-9840-43a3-945b-cb4d054e34ba-plugin-serving-cert") pod "nmstate-console-plugin-7fbb5f6569-7dpj4" (UID: "87deabac-9840-43a3-945b-cb4d054e34ba") : secret "plugin-serving-cert" not found Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.749525 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/87deabac-9840-43a3-945b-cb4d054e34ba-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-7dpj4\" (UID: \"87deabac-9840-43a3-945b-cb4d054e34ba\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7dpj4" Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.772633 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.808270 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-6499fbfdc8-929cv"] Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.809774 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6499fbfdc8-929cv" Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.838139 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6499fbfdc8-929cv"] Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.841091 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xzns\" (UniqueName: \"kubernetes.io/projected/87deabac-9840-43a3-945b-cb4d054e34ba-kube-api-access-8xzns\") pod \"nmstate-console-plugin-7fbb5f6569-7dpj4\" (UID: \"87deabac-9840-43a3-945b-cb4d054e34ba\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7dpj4" Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.938281 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-jbb6s" event={"ID":"57d09bf3-cccc-4745-834d-57eaad3f96a1","Type":"ContainerStarted","Data":"8399530dc9ec33c817dbff61a60a7178d4051ebce2e8a8605de93edc9dd0a29d"} Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.952429 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mh2mm\" (UniqueName: \"kubernetes.io/projected/ba14ef06-3f68-45a3-8fde-36980cd2d194-kube-api-access-mh2mm\") pod \"console-6499fbfdc8-929cv\" (UID: \"ba14ef06-3f68-45a3-8fde-36980cd2d194\") " pod="openshift-console/console-6499fbfdc8-929cv" Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.952535 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ba14ef06-3f68-45a3-8fde-36980cd2d194-service-ca\") pod \"console-6499fbfdc8-929cv\" (UID: \"ba14ef06-3f68-45a3-8fde-36980cd2d194\") " pod="openshift-console/console-6499fbfdc8-929cv" Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.952565 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ba14ef06-3f68-45a3-8fde-36980cd2d194-console-config\") pod \"console-6499fbfdc8-929cv\" (UID: \"ba14ef06-3f68-45a3-8fde-36980cd2d194\") " pod="openshift-console/console-6499fbfdc8-929cv" Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.952622 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ba14ef06-3f68-45a3-8fde-36980cd2d194-trusted-ca-bundle\") pod \"console-6499fbfdc8-929cv\" (UID: \"ba14ef06-3f68-45a3-8fde-36980cd2d194\") " pod="openshift-console/console-6499fbfdc8-929cv" Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.952695 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ba14ef06-3f68-45a3-8fde-36980cd2d194-console-serving-cert\") pod \"console-6499fbfdc8-929cv\" (UID: \"ba14ef06-3f68-45a3-8fde-36980cd2d194\") " pod="openshift-console/console-6499fbfdc8-929cv" Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.952720 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ba14ef06-3f68-45a3-8fde-36980cd2d194-console-oauth-config\") pod \"console-6499fbfdc8-929cv\" (UID: \"ba14ef06-3f68-45a3-8fde-36980cd2d194\") " pod="openshift-console/console-6499fbfdc8-929cv" Dec 06 06:10:03 crc kubenswrapper[4809]: I1206 06:10:03.952757 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ba14ef06-3f68-45a3-8fde-36980cd2d194-oauth-serving-cert\") pod \"console-6499fbfdc8-929cv\" (UID: \"ba14ef06-3f68-45a3-8fde-36980cd2d194\") " pod="openshift-console/console-6499fbfdc8-929cv" Dec 06 06:10:04 crc kubenswrapper[4809]: I1206 06:10:04.054829 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ba14ef06-3f68-45a3-8fde-36980cd2d194-console-serving-cert\") pod \"console-6499fbfdc8-929cv\" (UID: \"ba14ef06-3f68-45a3-8fde-36980cd2d194\") " pod="openshift-console/console-6499fbfdc8-929cv" Dec 06 06:10:04 crc kubenswrapper[4809]: I1206 06:10:04.054883 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ba14ef06-3f68-45a3-8fde-36980cd2d194-console-oauth-config\") pod \"console-6499fbfdc8-929cv\" (UID: \"ba14ef06-3f68-45a3-8fde-36980cd2d194\") " pod="openshift-console/console-6499fbfdc8-929cv" Dec 06 06:10:04 crc kubenswrapper[4809]: I1206 06:10:04.054912 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ba14ef06-3f68-45a3-8fde-36980cd2d194-oauth-serving-cert\") pod \"console-6499fbfdc8-929cv\" (UID: \"ba14ef06-3f68-45a3-8fde-36980cd2d194\") " pod="openshift-console/console-6499fbfdc8-929cv" Dec 06 06:10:04 crc kubenswrapper[4809]: I1206 06:10:04.055024 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mh2mm\" (UniqueName: \"kubernetes.io/projected/ba14ef06-3f68-45a3-8fde-36980cd2d194-kube-api-access-mh2mm\") pod \"console-6499fbfdc8-929cv\" (UID: \"ba14ef06-3f68-45a3-8fde-36980cd2d194\") " pod="openshift-console/console-6499fbfdc8-929cv" Dec 06 06:10:04 crc kubenswrapper[4809]: I1206 06:10:04.055060 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ba14ef06-3f68-45a3-8fde-36980cd2d194-service-ca\") pod \"console-6499fbfdc8-929cv\" (UID: \"ba14ef06-3f68-45a3-8fde-36980cd2d194\") " pod="openshift-console/console-6499fbfdc8-929cv" Dec 06 06:10:04 crc kubenswrapper[4809]: I1206 06:10:04.055088 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ba14ef06-3f68-45a3-8fde-36980cd2d194-console-config\") pod \"console-6499fbfdc8-929cv\" (UID: \"ba14ef06-3f68-45a3-8fde-36980cd2d194\") " pod="openshift-console/console-6499fbfdc8-929cv" Dec 06 06:10:04 crc kubenswrapper[4809]: I1206 06:10:04.055136 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ba14ef06-3f68-45a3-8fde-36980cd2d194-trusted-ca-bundle\") pod \"console-6499fbfdc8-929cv\" (UID: \"ba14ef06-3f68-45a3-8fde-36980cd2d194\") " pod="openshift-console/console-6499fbfdc8-929cv" Dec 06 06:10:04 crc kubenswrapper[4809]: I1206 06:10:04.055167 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/50b406f3-bc91-492f-89ea-8eeae9ae132e-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-pw6wx\" (UID: \"50b406f3-bc91-492f-89ea-8eeae9ae132e\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pw6wx" Dec 06 06:10:04 crc kubenswrapper[4809]: I1206 06:10:04.056732 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ba14ef06-3f68-45a3-8fde-36980cd2d194-oauth-serving-cert\") pod \"console-6499fbfdc8-929cv\" (UID: \"ba14ef06-3f68-45a3-8fde-36980cd2d194\") " pod="openshift-console/console-6499fbfdc8-929cv" Dec 06 06:10:04 crc kubenswrapper[4809]: I1206 06:10:04.057154 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ba14ef06-3f68-45a3-8fde-36980cd2d194-service-ca\") pod \"console-6499fbfdc8-929cv\" (UID: \"ba14ef06-3f68-45a3-8fde-36980cd2d194\") " pod="openshift-console/console-6499fbfdc8-929cv" Dec 06 06:10:04 crc kubenswrapper[4809]: I1206 06:10:04.057158 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ba14ef06-3f68-45a3-8fde-36980cd2d194-console-config\") pod \"console-6499fbfdc8-929cv\" (UID: \"ba14ef06-3f68-45a3-8fde-36980cd2d194\") " pod="openshift-console/console-6499fbfdc8-929cv" Dec 06 06:10:04 crc kubenswrapper[4809]: I1206 06:10:04.057383 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ba14ef06-3f68-45a3-8fde-36980cd2d194-trusted-ca-bundle\") pod \"console-6499fbfdc8-929cv\" (UID: \"ba14ef06-3f68-45a3-8fde-36980cd2d194\") " pod="openshift-console/console-6499fbfdc8-929cv" Dec 06 06:10:04 crc kubenswrapper[4809]: I1206 06:10:04.059814 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/50b406f3-bc91-492f-89ea-8eeae9ae132e-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-pw6wx\" (UID: \"50b406f3-bc91-492f-89ea-8eeae9ae132e\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pw6wx" Dec 06 06:10:04 crc kubenswrapper[4809]: I1206 06:10:04.068187 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ba14ef06-3f68-45a3-8fde-36980cd2d194-console-serving-cert\") pod \"console-6499fbfdc8-929cv\" (UID: \"ba14ef06-3f68-45a3-8fde-36980cd2d194\") " pod="openshift-console/console-6499fbfdc8-929cv" Dec 06 06:10:04 crc kubenswrapper[4809]: I1206 06:10:04.071189 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ba14ef06-3f68-45a3-8fde-36980cd2d194-console-oauth-config\") pod \"console-6499fbfdc8-929cv\" (UID: \"ba14ef06-3f68-45a3-8fde-36980cd2d194\") " pod="openshift-console/console-6499fbfdc8-929cv" Dec 06 06:10:04 crc kubenswrapper[4809]: I1206 06:10:04.079617 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mh2mm\" (UniqueName: \"kubernetes.io/projected/ba14ef06-3f68-45a3-8fde-36980cd2d194-kube-api-access-mh2mm\") pod \"console-6499fbfdc8-929cv\" (UID: \"ba14ef06-3f68-45a3-8fde-36980cd2d194\") " pod="openshift-console/console-6499fbfdc8-929cv" Dec 06 06:10:04 crc kubenswrapper[4809]: I1206 06:10:04.145140 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-pvgrd"] Dec 06 06:10:04 crc kubenswrapper[4809]: I1206 06:10:04.181126 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6499fbfdc8-929cv" Dec 06 06:10:04 crc kubenswrapper[4809]: I1206 06:10:04.259594 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/87deabac-9840-43a3-945b-cb4d054e34ba-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-7dpj4\" (UID: \"87deabac-9840-43a3-945b-cb4d054e34ba\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7dpj4" Dec 06 06:10:04 crc kubenswrapper[4809]: I1206 06:10:04.266058 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/87deabac-9840-43a3-945b-cb4d054e34ba-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-7dpj4\" (UID: \"87deabac-9840-43a3-945b-cb4d054e34ba\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7dpj4" Dec 06 06:10:04 crc kubenswrapper[4809]: I1206 06:10:04.268069 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7dpj4" Dec 06 06:10:04 crc kubenswrapper[4809]: I1206 06:10:04.269167 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pw6wx" Dec 06 06:10:04 crc kubenswrapper[4809]: I1206 06:10:04.678363 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6499fbfdc8-929cv"] Dec 06 06:10:04 crc kubenswrapper[4809]: I1206 06:10:04.834959 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pw6wx"] Dec 06 06:10:04 crc kubenswrapper[4809]: I1206 06:10:04.924422 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7dpj4"] Dec 06 06:10:04 crc kubenswrapper[4809]: W1206 06:10:04.935253 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod87deabac_9840_43a3_945b_cb4d054e34ba.slice/crio-4183b1633840fce9d33a47174e0e13afbb6cacc869870bec1892107fb2761c95 WatchSource:0}: Error finding container 4183b1633840fce9d33a47174e0e13afbb6cacc869870bec1892107fb2761c95: Status 404 returned error can't find the container with id 4183b1633840fce9d33a47174e0e13afbb6cacc869870bec1892107fb2761c95 Dec 06 06:10:04 crc kubenswrapper[4809]: I1206 06:10:04.956044 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6499fbfdc8-929cv" event={"ID":"ba14ef06-3f68-45a3-8fde-36980cd2d194","Type":"ContainerStarted","Data":"783e760a1bf33d4a66aaf5565d4bab5a8309f7adb668a85a0e93f84348ca28a4"} Dec 06 06:10:04 crc kubenswrapper[4809]: I1206 06:10:04.956168 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6499fbfdc8-929cv" event={"ID":"ba14ef06-3f68-45a3-8fde-36980cd2d194","Type":"ContainerStarted","Data":"9daa8bc4fa7f8980924840bbf78cd6023ac2490b195c4022042729a2a1dcc06d"} Dec 06 06:10:04 crc kubenswrapper[4809]: I1206 06:10:04.960375 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pw6wx" event={"ID":"50b406f3-bc91-492f-89ea-8eeae9ae132e","Type":"ContainerStarted","Data":"f769bf2b09bda78e6b8cee5176a48efef18b4aa9aa8522afab39400246de6a04"} Dec 06 06:10:04 crc kubenswrapper[4809]: I1206 06:10:04.962705 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-pvgrd" event={"ID":"d39362b1-ae5e-4852-be43-a335be68151b","Type":"ContainerStarted","Data":"431e53ca0bfebd7f3fb9e078aaa5af417d3dae9a9ea8fa149e3e0513396e177d"} Dec 06 06:10:04 crc kubenswrapper[4809]: I1206 06:10:04.964479 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7dpj4" event={"ID":"87deabac-9840-43a3-945b-cb4d054e34ba","Type":"ContainerStarted","Data":"4183b1633840fce9d33a47174e0e13afbb6cacc869870bec1892107fb2761c95"} Dec 06 06:10:04 crc kubenswrapper[4809]: I1206 06:10:04.989912 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-6499fbfdc8-929cv" podStartSLOduration=1.989875208 podStartE2EDuration="1.989875208s" podCreationTimestamp="2025-12-06 06:10:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:10:04.977805222 +0000 UTC m=+1129.866788164" watchObservedRunningTime="2025-12-06 06:10:04.989875208 +0000 UTC m=+1129.878858260" Dec 06 06:10:08 crc kubenswrapper[4809]: I1206 06:10:08.001729 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pw6wx" event={"ID":"50b406f3-bc91-492f-89ea-8eeae9ae132e","Type":"ContainerStarted","Data":"f0d8609da3b75f283c0dba67decac914ab2e69e782d8a9a9de73d12802057779"} Dec 06 06:10:08 crc kubenswrapper[4809]: I1206 06:10:08.002654 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pw6wx" Dec 06 06:10:08 crc kubenswrapper[4809]: I1206 06:10:08.005534 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-pvgrd" event={"ID":"d39362b1-ae5e-4852-be43-a335be68151b","Type":"ContainerStarted","Data":"ba43c70e68c87a0e128126e3fb4be2367b5ef884b1ae9f4c11f3c5a4aef6204a"} Dec 06 06:10:08 crc kubenswrapper[4809]: I1206 06:10:08.007500 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-jbb6s" event={"ID":"57d09bf3-cccc-4745-834d-57eaad3f96a1","Type":"ContainerStarted","Data":"4eadf5bf465bab46cc5f6d94b584d975fcacb987e0a064ee158488ba6f9a464b"} Dec 06 06:10:08 crc kubenswrapper[4809]: I1206 06:10:08.008364 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-jbb6s" Dec 06 06:10:08 crc kubenswrapper[4809]: I1206 06:10:08.041286 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pw6wx" podStartSLOduration=2.872685546 podStartE2EDuration="5.041257669s" podCreationTimestamp="2025-12-06 06:10:03 +0000 UTC" firstStartedPulling="2025-12-06 06:10:04.860060624 +0000 UTC m=+1129.749043566" lastFinishedPulling="2025-12-06 06:10:07.028632747 +0000 UTC m=+1131.917615689" observedRunningTime="2025-12-06 06:10:08.029671576 +0000 UTC m=+1132.918654548" watchObservedRunningTime="2025-12-06 06:10:08.041257669 +0000 UTC m=+1132.930240611" Dec 06 06:10:08 crc kubenswrapper[4809]: I1206 06:10:08.075183 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-jbb6s" podStartSLOduration=1.847522166 podStartE2EDuration="5.075152824s" podCreationTimestamp="2025-12-06 06:10:03 +0000 UTC" firstStartedPulling="2025-12-06 06:10:03.772373156 +0000 UTC m=+1128.661356098" lastFinishedPulling="2025-12-06 06:10:07.000003794 +0000 UTC m=+1131.888986756" observedRunningTime="2025-12-06 06:10:08.070042916 +0000 UTC m=+1132.959025858" watchObservedRunningTime="2025-12-06 06:10:08.075152824 +0000 UTC m=+1132.964135766" Dec 06 06:10:09 crc kubenswrapper[4809]: I1206 06:10:09.018877 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7dpj4" event={"ID":"87deabac-9840-43a3-945b-cb4d054e34ba","Type":"ContainerStarted","Data":"e8bfd3b3e14d87886d7fe67d9a7cc16e5ddbe3714e44b13ce921d30a6aa8cc49"} Dec 06 06:10:13 crc kubenswrapper[4809]: I1206 06:10:13.728206 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-jbb6s" Dec 06 06:10:13 crc kubenswrapper[4809]: I1206 06:10:13.758956 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7dpj4" podStartSLOduration=7.612107579 podStartE2EDuration="10.758902346s" podCreationTimestamp="2025-12-06 06:10:03 +0000 UTC" firstStartedPulling="2025-12-06 06:10:04.938612834 +0000 UTC m=+1129.827595776" lastFinishedPulling="2025-12-06 06:10:08.085407601 +0000 UTC m=+1132.974390543" observedRunningTime="2025-12-06 06:10:09.044594701 +0000 UTC m=+1133.933577653" watchObservedRunningTime="2025-12-06 06:10:13.758902346 +0000 UTC m=+1138.647885288" Dec 06 06:10:14 crc kubenswrapper[4809]: I1206 06:10:14.072155 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-pvgrd" event={"ID":"d39362b1-ae5e-4852-be43-a335be68151b","Type":"ContainerStarted","Data":"8386ca05cf81b316a6c9e28fb52072cbbf74408b46953a501fa4242a39131989"} Dec 06 06:10:14 crc kubenswrapper[4809]: I1206 06:10:14.120745 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-pvgrd" podStartSLOduration=2.374570511 podStartE2EDuration="11.120706312s" podCreationTimestamp="2025-12-06 06:10:03 +0000 UTC" firstStartedPulling="2025-12-06 06:10:04.157247944 +0000 UTC m=+1129.046230886" lastFinishedPulling="2025-12-06 06:10:12.903383745 +0000 UTC m=+1137.792366687" observedRunningTime="2025-12-06 06:10:14.093414445 +0000 UTC m=+1138.982397387" watchObservedRunningTime="2025-12-06 06:10:14.120706312 +0000 UTC m=+1139.009689264" Dec 06 06:10:14 crc kubenswrapper[4809]: I1206 06:10:14.181610 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-6499fbfdc8-929cv" Dec 06 06:10:14 crc kubenswrapper[4809]: I1206 06:10:14.181758 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-6499fbfdc8-929cv" Dec 06 06:10:14 crc kubenswrapper[4809]: I1206 06:10:14.188177 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-6499fbfdc8-929cv" Dec 06 06:10:15 crc kubenswrapper[4809]: I1206 06:10:15.085589 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-6499fbfdc8-929cv" Dec 06 06:10:15 crc kubenswrapper[4809]: I1206 06:10:15.189094 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-664586d6fb-zgzv8"] Dec 06 06:10:24 crc kubenswrapper[4809]: I1206 06:10:24.275534 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pw6wx" Dec 06 06:10:40 crc kubenswrapper[4809]: I1206 06:10:40.233117 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-664586d6fb-zgzv8" podUID="7b54a496-5ecf-48c6-a9ba-d30828286c79" containerName="console" containerID="cri-o://b89bc9800243faf10718de24d8ac3f3d8d6489019f431edef21c84ce711eb4d4" gracePeriod=15 Dec 06 06:10:40 crc kubenswrapper[4809]: I1206 06:10:40.447868 4809 patch_prober.go:28] interesting pod/console-664586d6fb-zgzv8 container/console namespace/openshift-console: Readiness probe status=failure output="Get \"https://10.217.0.80:8443/health\": dial tcp 10.217.0.80:8443: connect: connection refused" start-of-body= Dec 06 06:10:40 crc kubenswrapper[4809]: I1206 06:10:40.448686 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/console-664586d6fb-zgzv8" podUID="7b54a496-5ecf-48c6-a9ba-d30828286c79" containerName="console" probeResult="failure" output="Get \"https://10.217.0.80:8443/health\": dial tcp 10.217.0.80:8443: connect: connection refused" Dec 06 06:10:40 crc kubenswrapper[4809]: I1206 06:10:40.697848 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-664586d6fb-zgzv8_7b54a496-5ecf-48c6-a9ba-d30828286c79/console/0.log" Dec 06 06:10:40 crc kubenswrapper[4809]: I1206 06:10:40.698346 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-664586d6fb-zgzv8" Dec 06 06:10:40 crc kubenswrapper[4809]: I1206 06:10:40.774624 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7b54a496-5ecf-48c6-a9ba-d30828286c79-oauth-serving-cert\") pod \"7b54a496-5ecf-48c6-a9ba-d30828286c79\" (UID: \"7b54a496-5ecf-48c6-a9ba-d30828286c79\") " Dec 06 06:10:40 crc kubenswrapper[4809]: I1206 06:10:40.774701 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7b54a496-5ecf-48c6-a9ba-d30828286c79-service-ca\") pod \"7b54a496-5ecf-48c6-a9ba-d30828286c79\" (UID: \"7b54a496-5ecf-48c6-a9ba-d30828286c79\") " Dec 06 06:10:40 crc kubenswrapper[4809]: I1206 06:10:40.774819 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7b54a496-5ecf-48c6-a9ba-d30828286c79-console-oauth-config\") pod \"7b54a496-5ecf-48c6-a9ba-d30828286c79\" (UID: \"7b54a496-5ecf-48c6-a9ba-d30828286c79\") " Dec 06 06:10:40 crc kubenswrapper[4809]: I1206 06:10:40.774849 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7b54a496-5ecf-48c6-a9ba-d30828286c79-trusted-ca-bundle\") pod \"7b54a496-5ecf-48c6-a9ba-d30828286c79\" (UID: \"7b54a496-5ecf-48c6-a9ba-d30828286c79\") " Dec 06 06:10:40 crc kubenswrapper[4809]: I1206 06:10:40.774895 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7b54a496-5ecf-48c6-a9ba-d30828286c79-console-config\") pod \"7b54a496-5ecf-48c6-a9ba-d30828286c79\" (UID: \"7b54a496-5ecf-48c6-a9ba-d30828286c79\") " Dec 06 06:10:40 crc kubenswrapper[4809]: I1206 06:10:40.775085 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwd2s\" (UniqueName: \"kubernetes.io/projected/7b54a496-5ecf-48c6-a9ba-d30828286c79-kube-api-access-lwd2s\") pod \"7b54a496-5ecf-48c6-a9ba-d30828286c79\" (UID: \"7b54a496-5ecf-48c6-a9ba-d30828286c79\") " Dec 06 06:10:40 crc kubenswrapper[4809]: I1206 06:10:40.775129 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7b54a496-5ecf-48c6-a9ba-d30828286c79-console-serving-cert\") pod \"7b54a496-5ecf-48c6-a9ba-d30828286c79\" (UID: \"7b54a496-5ecf-48c6-a9ba-d30828286c79\") " Dec 06 06:10:40 crc kubenswrapper[4809]: I1206 06:10:40.776813 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b54a496-5ecf-48c6-a9ba-d30828286c79-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "7b54a496-5ecf-48c6-a9ba-d30828286c79" (UID: "7b54a496-5ecf-48c6-a9ba-d30828286c79"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:10:40 crc kubenswrapper[4809]: I1206 06:10:40.776881 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b54a496-5ecf-48c6-a9ba-d30828286c79-service-ca" (OuterVolumeSpecName: "service-ca") pod "7b54a496-5ecf-48c6-a9ba-d30828286c79" (UID: "7b54a496-5ecf-48c6-a9ba-d30828286c79"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:10:40 crc kubenswrapper[4809]: I1206 06:10:40.777444 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b54a496-5ecf-48c6-a9ba-d30828286c79-console-config" (OuterVolumeSpecName: "console-config") pod "7b54a496-5ecf-48c6-a9ba-d30828286c79" (UID: "7b54a496-5ecf-48c6-a9ba-d30828286c79"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:10:40 crc kubenswrapper[4809]: I1206 06:10:40.777809 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b54a496-5ecf-48c6-a9ba-d30828286c79-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "7b54a496-5ecf-48c6-a9ba-d30828286c79" (UID: "7b54a496-5ecf-48c6-a9ba-d30828286c79"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:10:40 crc kubenswrapper[4809]: I1206 06:10:40.783705 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b54a496-5ecf-48c6-a9ba-d30828286c79-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "7b54a496-5ecf-48c6-a9ba-d30828286c79" (UID: "7b54a496-5ecf-48c6-a9ba-d30828286c79"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:10:40 crc kubenswrapper[4809]: I1206 06:10:40.784126 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b54a496-5ecf-48c6-a9ba-d30828286c79-kube-api-access-lwd2s" (OuterVolumeSpecName: "kube-api-access-lwd2s") pod "7b54a496-5ecf-48c6-a9ba-d30828286c79" (UID: "7b54a496-5ecf-48c6-a9ba-d30828286c79"). InnerVolumeSpecName "kube-api-access-lwd2s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:10:40 crc kubenswrapper[4809]: I1206 06:10:40.785490 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b54a496-5ecf-48c6-a9ba-d30828286c79-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "7b54a496-5ecf-48c6-a9ba-d30828286c79" (UID: "7b54a496-5ecf-48c6-a9ba-d30828286c79"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:10:40 crc kubenswrapper[4809]: I1206 06:10:40.877705 4809 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7b54a496-5ecf-48c6-a9ba-d30828286c79-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 06:10:40 crc kubenswrapper[4809]: I1206 06:10:40.877791 4809 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7b54a496-5ecf-48c6-a9ba-d30828286c79-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 06:10:40 crc kubenswrapper[4809]: I1206 06:10:40.877810 4809 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7b54a496-5ecf-48c6-a9ba-d30828286c79-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:10:40 crc kubenswrapper[4809]: I1206 06:10:40.877827 4809 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7b54a496-5ecf-48c6-a9ba-d30828286c79-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:10:40 crc kubenswrapper[4809]: I1206 06:10:40.877871 4809 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7b54a496-5ecf-48c6-a9ba-d30828286c79-console-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:10:40 crc kubenswrapper[4809]: I1206 06:10:40.877889 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwd2s\" (UniqueName: \"kubernetes.io/projected/7b54a496-5ecf-48c6-a9ba-d30828286c79-kube-api-access-lwd2s\") on node \"crc\" DevicePath \"\"" Dec 06 06:10:40 crc kubenswrapper[4809]: I1206 06:10:40.877905 4809 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7b54a496-5ecf-48c6-a9ba-d30828286c79-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 06:10:41 crc kubenswrapper[4809]: I1206 06:10:41.358214 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-664586d6fb-zgzv8_7b54a496-5ecf-48c6-a9ba-d30828286c79/console/0.log" Dec 06 06:10:41 crc kubenswrapper[4809]: I1206 06:10:41.358270 4809 generic.go:334] "Generic (PLEG): container finished" podID="7b54a496-5ecf-48c6-a9ba-d30828286c79" containerID="b89bc9800243faf10718de24d8ac3f3d8d6489019f431edef21c84ce711eb4d4" exitCode=2 Dec 06 06:10:41 crc kubenswrapper[4809]: I1206 06:10:41.358303 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-664586d6fb-zgzv8" event={"ID":"7b54a496-5ecf-48c6-a9ba-d30828286c79","Type":"ContainerDied","Data":"b89bc9800243faf10718de24d8ac3f3d8d6489019f431edef21c84ce711eb4d4"} Dec 06 06:10:41 crc kubenswrapper[4809]: I1206 06:10:41.358326 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-664586d6fb-zgzv8" Dec 06 06:10:41 crc kubenswrapper[4809]: I1206 06:10:41.358343 4809 scope.go:117] "RemoveContainer" containerID="b89bc9800243faf10718de24d8ac3f3d8d6489019f431edef21c84ce711eb4d4" Dec 06 06:10:41 crc kubenswrapper[4809]: I1206 06:10:41.358333 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-664586d6fb-zgzv8" event={"ID":"7b54a496-5ecf-48c6-a9ba-d30828286c79","Type":"ContainerDied","Data":"4ef66f19a1e70d8f9939127aba4edacd0985245ec312ed68c403b0bfd23a9866"} Dec 06 06:10:41 crc kubenswrapper[4809]: I1206 06:10:41.381342 4809 scope.go:117] "RemoveContainer" containerID="b89bc9800243faf10718de24d8ac3f3d8d6489019f431edef21c84ce711eb4d4" Dec 06 06:10:41 crc kubenswrapper[4809]: E1206 06:10:41.384466 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b89bc9800243faf10718de24d8ac3f3d8d6489019f431edef21c84ce711eb4d4\": container with ID starting with b89bc9800243faf10718de24d8ac3f3d8d6489019f431edef21c84ce711eb4d4 not found: ID does not exist" containerID="b89bc9800243faf10718de24d8ac3f3d8d6489019f431edef21c84ce711eb4d4" Dec 06 06:10:41 crc kubenswrapper[4809]: I1206 06:10:41.384532 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b89bc9800243faf10718de24d8ac3f3d8d6489019f431edef21c84ce711eb4d4"} err="failed to get container status \"b89bc9800243faf10718de24d8ac3f3d8d6489019f431edef21c84ce711eb4d4\": rpc error: code = NotFound desc = could not find container \"b89bc9800243faf10718de24d8ac3f3d8d6489019f431edef21c84ce711eb4d4\": container with ID starting with b89bc9800243faf10718de24d8ac3f3d8d6489019f431edef21c84ce711eb4d4 not found: ID does not exist" Dec 06 06:10:41 crc kubenswrapper[4809]: I1206 06:10:41.401451 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-664586d6fb-zgzv8"] Dec 06 06:10:41 crc kubenswrapper[4809]: I1206 06:10:41.401490 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-664586d6fb-zgzv8"] Dec 06 06:10:43 crc kubenswrapper[4809]: I1206 06:10:43.399359 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b54a496-5ecf-48c6-a9ba-d30828286c79" path="/var/lib/kubelet/pods/7b54a496-5ecf-48c6-a9ba-d30828286c79/volumes" Dec 06 06:10:43 crc kubenswrapper[4809]: I1206 06:10:43.457095 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rhmbp"] Dec 06 06:10:43 crc kubenswrapper[4809]: E1206 06:10:43.457371 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b54a496-5ecf-48c6-a9ba-d30828286c79" containerName="console" Dec 06 06:10:43 crc kubenswrapper[4809]: I1206 06:10:43.457385 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b54a496-5ecf-48c6-a9ba-d30828286c79" containerName="console" Dec 06 06:10:43 crc kubenswrapper[4809]: I1206 06:10:43.457542 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b54a496-5ecf-48c6-a9ba-d30828286c79" containerName="console" Dec 06 06:10:43 crc kubenswrapper[4809]: I1206 06:10:43.458689 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rhmbp" Dec 06 06:10:43 crc kubenswrapper[4809]: I1206 06:10:43.461507 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 06 06:10:43 crc kubenswrapper[4809]: I1206 06:10:43.477966 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rhmbp"] Dec 06 06:10:43 crc kubenswrapper[4809]: I1206 06:10:43.629770 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ade7d3b7-533c-40a1-9964-4e5ac7b88214-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rhmbp\" (UID: \"ade7d3b7-533c-40a1-9964-4e5ac7b88214\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rhmbp" Dec 06 06:10:43 crc kubenswrapper[4809]: I1206 06:10:43.629835 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bss5d\" (UniqueName: \"kubernetes.io/projected/ade7d3b7-533c-40a1-9964-4e5ac7b88214-kube-api-access-bss5d\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rhmbp\" (UID: \"ade7d3b7-533c-40a1-9964-4e5ac7b88214\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rhmbp" Dec 06 06:10:43 crc kubenswrapper[4809]: I1206 06:10:43.629899 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ade7d3b7-533c-40a1-9964-4e5ac7b88214-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rhmbp\" (UID: \"ade7d3b7-533c-40a1-9964-4e5ac7b88214\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rhmbp" Dec 06 06:10:43 crc kubenswrapper[4809]: I1206 06:10:43.731688 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ade7d3b7-533c-40a1-9964-4e5ac7b88214-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rhmbp\" (UID: \"ade7d3b7-533c-40a1-9964-4e5ac7b88214\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rhmbp" Dec 06 06:10:43 crc kubenswrapper[4809]: I1206 06:10:43.731822 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ade7d3b7-533c-40a1-9964-4e5ac7b88214-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rhmbp\" (UID: \"ade7d3b7-533c-40a1-9964-4e5ac7b88214\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rhmbp" Dec 06 06:10:43 crc kubenswrapper[4809]: I1206 06:10:43.731872 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bss5d\" (UniqueName: \"kubernetes.io/projected/ade7d3b7-533c-40a1-9964-4e5ac7b88214-kube-api-access-bss5d\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rhmbp\" (UID: \"ade7d3b7-533c-40a1-9964-4e5ac7b88214\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rhmbp" Dec 06 06:10:43 crc kubenswrapper[4809]: I1206 06:10:43.733684 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ade7d3b7-533c-40a1-9964-4e5ac7b88214-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rhmbp\" (UID: \"ade7d3b7-533c-40a1-9964-4e5ac7b88214\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rhmbp" Dec 06 06:10:43 crc kubenswrapper[4809]: I1206 06:10:43.733904 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ade7d3b7-533c-40a1-9964-4e5ac7b88214-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rhmbp\" (UID: \"ade7d3b7-533c-40a1-9964-4e5ac7b88214\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rhmbp" Dec 06 06:10:43 crc kubenswrapper[4809]: I1206 06:10:43.761517 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bss5d\" (UniqueName: \"kubernetes.io/projected/ade7d3b7-533c-40a1-9964-4e5ac7b88214-kube-api-access-bss5d\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rhmbp\" (UID: \"ade7d3b7-533c-40a1-9964-4e5ac7b88214\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rhmbp" Dec 06 06:10:43 crc kubenswrapper[4809]: I1206 06:10:43.783127 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rhmbp" Dec 06 06:10:44 crc kubenswrapper[4809]: I1206 06:10:44.257389 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rhmbp"] Dec 06 06:10:44 crc kubenswrapper[4809]: I1206 06:10:44.391345 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rhmbp" event={"ID":"ade7d3b7-533c-40a1-9964-4e5ac7b88214","Type":"ContainerStarted","Data":"9fe6c707aa410174ae445dde8f1675fdfb782e2a5829ebe1a12fca00b613a4aa"} Dec 06 06:10:45 crc kubenswrapper[4809]: I1206 06:10:45.411769 4809 generic.go:334] "Generic (PLEG): container finished" podID="ade7d3b7-533c-40a1-9964-4e5ac7b88214" containerID="707d2d3bbe0f73f6fccf14eda150b5405491f529af6015fc5efe72fd8b4f9f44" exitCode=0 Dec 06 06:10:45 crc kubenswrapper[4809]: I1206 06:10:45.411807 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rhmbp" event={"ID":"ade7d3b7-533c-40a1-9964-4e5ac7b88214","Type":"ContainerDied","Data":"707d2d3bbe0f73f6fccf14eda150b5405491f529af6015fc5efe72fd8b4f9f44"} Dec 06 06:10:47 crc kubenswrapper[4809]: I1206 06:10:47.433051 4809 generic.go:334] "Generic (PLEG): container finished" podID="ade7d3b7-533c-40a1-9964-4e5ac7b88214" containerID="cfb9a4a57abf31bde7a6e9a7bffad904dc7fe987bc030a34507a0d5a7362b4e9" exitCode=0 Dec 06 06:10:47 crc kubenswrapper[4809]: I1206 06:10:47.433168 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rhmbp" event={"ID":"ade7d3b7-533c-40a1-9964-4e5ac7b88214","Type":"ContainerDied","Data":"cfb9a4a57abf31bde7a6e9a7bffad904dc7fe987bc030a34507a0d5a7362b4e9"} Dec 06 06:10:48 crc kubenswrapper[4809]: I1206 06:10:48.446027 4809 generic.go:334] "Generic (PLEG): container finished" podID="ade7d3b7-533c-40a1-9964-4e5ac7b88214" containerID="25c0714e0701d0a56ee163a4f6541ca01110171a412cc42f36dd46c0893d0b79" exitCode=0 Dec 06 06:10:48 crc kubenswrapper[4809]: I1206 06:10:48.446129 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rhmbp" event={"ID":"ade7d3b7-533c-40a1-9964-4e5ac7b88214","Type":"ContainerDied","Data":"25c0714e0701d0a56ee163a4f6541ca01110171a412cc42f36dd46c0893d0b79"} Dec 06 06:10:49 crc kubenswrapper[4809]: I1206 06:10:49.804151 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rhmbp" Dec 06 06:10:49 crc kubenswrapper[4809]: I1206 06:10:49.956826 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ade7d3b7-533c-40a1-9964-4e5ac7b88214-util\") pod \"ade7d3b7-533c-40a1-9964-4e5ac7b88214\" (UID: \"ade7d3b7-533c-40a1-9964-4e5ac7b88214\") " Dec 06 06:10:49 crc kubenswrapper[4809]: I1206 06:10:49.957218 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ade7d3b7-533c-40a1-9964-4e5ac7b88214-bundle\") pod \"ade7d3b7-533c-40a1-9964-4e5ac7b88214\" (UID: \"ade7d3b7-533c-40a1-9964-4e5ac7b88214\") " Dec 06 06:10:49 crc kubenswrapper[4809]: I1206 06:10:49.957398 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bss5d\" (UniqueName: \"kubernetes.io/projected/ade7d3b7-533c-40a1-9964-4e5ac7b88214-kube-api-access-bss5d\") pod \"ade7d3b7-533c-40a1-9964-4e5ac7b88214\" (UID: \"ade7d3b7-533c-40a1-9964-4e5ac7b88214\") " Dec 06 06:10:49 crc kubenswrapper[4809]: I1206 06:10:49.958459 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ade7d3b7-533c-40a1-9964-4e5ac7b88214-bundle" (OuterVolumeSpecName: "bundle") pod "ade7d3b7-533c-40a1-9964-4e5ac7b88214" (UID: "ade7d3b7-533c-40a1-9964-4e5ac7b88214"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:10:49 crc kubenswrapper[4809]: I1206 06:10:49.963544 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ade7d3b7-533c-40a1-9964-4e5ac7b88214-kube-api-access-bss5d" (OuterVolumeSpecName: "kube-api-access-bss5d") pod "ade7d3b7-533c-40a1-9964-4e5ac7b88214" (UID: "ade7d3b7-533c-40a1-9964-4e5ac7b88214"). InnerVolumeSpecName "kube-api-access-bss5d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:10:50 crc kubenswrapper[4809]: I1206 06:10:50.060133 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bss5d\" (UniqueName: \"kubernetes.io/projected/ade7d3b7-533c-40a1-9964-4e5ac7b88214-kube-api-access-bss5d\") on node \"crc\" DevicePath \"\"" Dec 06 06:10:50 crc kubenswrapper[4809]: I1206 06:10:50.060182 4809 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ade7d3b7-533c-40a1-9964-4e5ac7b88214-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:10:50 crc kubenswrapper[4809]: I1206 06:10:50.114526 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ade7d3b7-533c-40a1-9964-4e5ac7b88214-util" (OuterVolumeSpecName: "util") pod "ade7d3b7-533c-40a1-9964-4e5ac7b88214" (UID: "ade7d3b7-533c-40a1-9964-4e5ac7b88214"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:10:50 crc kubenswrapper[4809]: I1206 06:10:50.162117 4809 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ade7d3b7-533c-40a1-9964-4e5ac7b88214-util\") on node \"crc\" DevicePath \"\"" Dec 06 06:10:50 crc kubenswrapper[4809]: I1206 06:10:50.462250 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rhmbp" event={"ID":"ade7d3b7-533c-40a1-9964-4e5ac7b88214","Type":"ContainerDied","Data":"9fe6c707aa410174ae445dde8f1675fdfb782e2a5829ebe1a12fca00b613a4aa"} Dec 06 06:10:50 crc kubenswrapper[4809]: I1206 06:10:50.462306 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rhmbp" Dec 06 06:10:50 crc kubenswrapper[4809]: I1206 06:10:50.462310 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9fe6c707aa410174ae445dde8f1675fdfb782e2a5829ebe1a12fca00b613a4aa" Dec 06 06:10:57 crc kubenswrapper[4809]: I1206 06:10:57.622238 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-6f667b6558-c67wt"] Dec 06 06:10:57 crc kubenswrapper[4809]: E1206 06:10:57.623986 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ade7d3b7-533c-40a1-9964-4e5ac7b88214" containerName="extract" Dec 06 06:10:57 crc kubenswrapper[4809]: I1206 06:10:57.624100 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ade7d3b7-533c-40a1-9964-4e5ac7b88214" containerName="extract" Dec 06 06:10:57 crc kubenswrapper[4809]: E1206 06:10:57.624180 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ade7d3b7-533c-40a1-9964-4e5ac7b88214" containerName="pull" Dec 06 06:10:57 crc kubenswrapper[4809]: I1206 06:10:57.624239 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ade7d3b7-533c-40a1-9964-4e5ac7b88214" containerName="pull" Dec 06 06:10:57 crc kubenswrapper[4809]: E1206 06:10:57.624325 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ade7d3b7-533c-40a1-9964-4e5ac7b88214" containerName="util" Dec 06 06:10:57 crc kubenswrapper[4809]: I1206 06:10:57.624389 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ade7d3b7-533c-40a1-9964-4e5ac7b88214" containerName="util" Dec 06 06:10:57 crc kubenswrapper[4809]: I1206 06:10:57.624585 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ade7d3b7-533c-40a1-9964-4e5ac7b88214" containerName="extract" Dec 06 06:10:57 crc kubenswrapper[4809]: I1206 06:10:57.625259 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6f667b6558-c67wt" Dec 06 06:10:57 crc kubenswrapper[4809]: I1206 06:10:57.628101 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-kdvpp" Dec 06 06:10:57 crc kubenswrapper[4809]: I1206 06:10:57.628723 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 06 06:10:57 crc kubenswrapper[4809]: I1206 06:10:57.628959 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 06 06:10:57 crc kubenswrapper[4809]: I1206 06:10:57.629081 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 06 06:10:57 crc kubenswrapper[4809]: I1206 06:10:57.629411 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 06 06:10:57 crc kubenswrapper[4809]: I1206 06:10:57.685589 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/69484422-8317-4050-945c-9c9b127552b1-apiservice-cert\") pod \"metallb-operator-controller-manager-6f667b6558-c67wt\" (UID: \"69484422-8317-4050-945c-9c9b127552b1\") " pod="metallb-system/metallb-operator-controller-manager-6f667b6558-c67wt" Dec 06 06:10:57 crc kubenswrapper[4809]: I1206 06:10:57.685679 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9nzg\" (UniqueName: \"kubernetes.io/projected/69484422-8317-4050-945c-9c9b127552b1-kube-api-access-s9nzg\") pod \"metallb-operator-controller-manager-6f667b6558-c67wt\" (UID: \"69484422-8317-4050-945c-9c9b127552b1\") " pod="metallb-system/metallb-operator-controller-manager-6f667b6558-c67wt" Dec 06 06:10:57 crc kubenswrapper[4809]: I1206 06:10:57.686029 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/69484422-8317-4050-945c-9c9b127552b1-webhook-cert\") pod \"metallb-operator-controller-manager-6f667b6558-c67wt\" (UID: \"69484422-8317-4050-945c-9c9b127552b1\") " pod="metallb-system/metallb-operator-controller-manager-6f667b6558-c67wt" Dec 06 06:10:57 crc kubenswrapper[4809]: I1206 06:10:57.712494 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6f667b6558-c67wt"] Dec 06 06:10:57 crc kubenswrapper[4809]: I1206 06:10:57.787992 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/69484422-8317-4050-945c-9c9b127552b1-apiservice-cert\") pod \"metallb-operator-controller-manager-6f667b6558-c67wt\" (UID: \"69484422-8317-4050-945c-9c9b127552b1\") " pod="metallb-system/metallb-operator-controller-manager-6f667b6558-c67wt" Dec 06 06:10:57 crc kubenswrapper[4809]: I1206 06:10:57.788050 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9nzg\" (UniqueName: \"kubernetes.io/projected/69484422-8317-4050-945c-9c9b127552b1-kube-api-access-s9nzg\") pod \"metallb-operator-controller-manager-6f667b6558-c67wt\" (UID: \"69484422-8317-4050-945c-9c9b127552b1\") " pod="metallb-system/metallb-operator-controller-manager-6f667b6558-c67wt" Dec 06 06:10:57 crc kubenswrapper[4809]: I1206 06:10:57.788133 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/69484422-8317-4050-945c-9c9b127552b1-webhook-cert\") pod \"metallb-operator-controller-manager-6f667b6558-c67wt\" (UID: \"69484422-8317-4050-945c-9c9b127552b1\") " pod="metallb-system/metallb-operator-controller-manager-6f667b6558-c67wt" Dec 06 06:10:57 crc kubenswrapper[4809]: I1206 06:10:57.794644 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/69484422-8317-4050-945c-9c9b127552b1-apiservice-cert\") pod \"metallb-operator-controller-manager-6f667b6558-c67wt\" (UID: \"69484422-8317-4050-945c-9c9b127552b1\") " pod="metallb-system/metallb-operator-controller-manager-6f667b6558-c67wt" Dec 06 06:10:57 crc kubenswrapper[4809]: I1206 06:10:57.802924 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/69484422-8317-4050-945c-9c9b127552b1-webhook-cert\") pod \"metallb-operator-controller-manager-6f667b6558-c67wt\" (UID: \"69484422-8317-4050-945c-9c9b127552b1\") " pod="metallb-system/metallb-operator-controller-manager-6f667b6558-c67wt" Dec 06 06:10:57 crc kubenswrapper[4809]: I1206 06:10:57.809739 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9nzg\" (UniqueName: \"kubernetes.io/projected/69484422-8317-4050-945c-9c9b127552b1-kube-api-access-s9nzg\") pod \"metallb-operator-controller-manager-6f667b6558-c67wt\" (UID: \"69484422-8317-4050-945c-9c9b127552b1\") " pod="metallb-system/metallb-operator-controller-manager-6f667b6558-c67wt" Dec 06 06:10:57 crc kubenswrapper[4809]: I1206 06:10:57.947460 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6f667b6558-c67wt" Dec 06 06:10:58 crc kubenswrapper[4809]: I1206 06:10:58.209190 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-765b7d7f99-8dngt"] Dec 06 06:10:58 crc kubenswrapper[4809]: I1206 06:10:58.210695 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-765b7d7f99-8dngt" Dec 06 06:10:58 crc kubenswrapper[4809]: I1206 06:10:58.214521 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 06 06:10:58 crc kubenswrapper[4809]: I1206 06:10:58.214562 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-qxgdx" Dec 06 06:10:58 crc kubenswrapper[4809]: I1206 06:10:58.214837 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 06 06:10:58 crc kubenswrapper[4809]: I1206 06:10:58.239795 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-765b7d7f99-8dngt"] Dec 06 06:10:58 crc kubenswrapper[4809]: I1206 06:10:58.299159 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/60798ea0-75be-4a0c-a8c6-9fb431ba0e67-apiservice-cert\") pod \"metallb-operator-webhook-server-765b7d7f99-8dngt\" (UID: \"60798ea0-75be-4a0c-a8c6-9fb431ba0e67\") " pod="metallb-system/metallb-operator-webhook-server-765b7d7f99-8dngt" Dec 06 06:10:58 crc kubenswrapper[4809]: I1206 06:10:58.299281 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbhzn\" (UniqueName: \"kubernetes.io/projected/60798ea0-75be-4a0c-a8c6-9fb431ba0e67-kube-api-access-hbhzn\") pod \"metallb-operator-webhook-server-765b7d7f99-8dngt\" (UID: \"60798ea0-75be-4a0c-a8c6-9fb431ba0e67\") " pod="metallb-system/metallb-operator-webhook-server-765b7d7f99-8dngt" Dec 06 06:10:58 crc kubenswrapper[4809]: I1206 06:10:58.299304 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/60798ea0-75be-4a0c-a8c6-9fb431ba0e67-webhook-cert\") pod \"metallb-operator-webhook-server-765b7d7f99-8dngt\" (UID: \"60798ea0-75be-4a0c-a8c6-9fb431ba0e67\") " pod="metallb-system/metallb-operator-webhook-server-765b7d7f99-8dngt" Dec 06 06:10:58 crc kubenswrapper[4809]: I1206 06:10:58.400898 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbhzn\" (UniqueName: \"kubernetes.io/projected/60798ea0-75be-4a0c-a8c6-9fb431ba0e67-kube-api-access-hbhzn\") pod \"metallb-operator-webhook-server-765b7d7f99-8dngt\" (UID: \"60798ea0-75be-4a0c-a8c6-9fb431ba0e67\") " pod="metallb-system/metallb-operator-webhook-server-765b7d7f99-8dngt" Dec 06 06:10:58 crc kubenswrapper[4809]: I1206 06:10:58.400973 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/60798ea0-75be-4a0c-a8c6-9fb431ba0e67-webhook-cert\") pod \"metallb-operator-webhook-server-765b7d7f99-8dngt\" (UID: \"60798ea0-75be-4a0c-a8c6-9fb431ba0e67\") " pod="metallb-system/metallb-operator-webhook-server-765b7d7f99-8dngt" Dec 06 06:10:58 crc kubenswrapper[4809]: I1206 06:10:58.401080 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/60798ea0-75be-4a0c-a8c6-9fb431ba0e67-apiservice-cert\") pod \"metallb-operator-webhook-server-765b7d7f99-8dngt\" (UID: \"60798ea0-75be-4a0c-a8c6-9fb431ba0e67\") " pod="metallb-system/metallb-operator-webhook-server-765b7d7f99-8dngt" Dec 06 06:10:58 crc kubenswrapper[4809]: I1206 06:10:58.414893 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/60798ea0-75be-4a0c-a8c6-9fb431ba0e67-webhook-cert\") pod \"metallb-operator-webhook-server-765b7d7f99-8dngt\" (UID: \"60798ea0-75be-4a0c-a8c6-9fb431ba0e67\") " pod="metallb-system/metallb-operator-webhook-server-765b7d7f99-8dngt" Dec 06 06:10:58 crc kubenswrapper[4809]: I1206 06:10:58.418369 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/60798ea0-75be-4a0c-a8c6-9fb431ba0e67-apiservice-cert\") pod \"metallb-operator-webhook-server-765b7d7f99-8dngt\" (UID: \"60798ea0-75be-4a0c-a8c6-9fb431ba0e67\") " pod="metallb-system/metallb-operator-webhook-server-765b7d7f99-8dngt" Dec 06 06:10:58 crc kubenswrapper[4809]: I1206 06:10:58.425503 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbhzn\" (UniqueName: \"kubernetes.io/projected/60798ea0-75be-4a0c-a8c6-9fb431ba0e67-kube-api-access-hbhzn\") pod \"metallb-operator-webhook-server-765b7d7f99-8dngt\" (UID: \"60798ea0-75be-4a0c-a8c6-9fb431ba0e67\") " pod="metallb-system/metallb-operator-webhook-server-765b7d7f99-8dngt" Dec 06 06:10:58 crc kubenswrapper[4809]: I1206 06:10:58.500204 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6f667b6558-c67wt"] Dec 06 06:10:58 crc kubenswrapper[4809]: I1206 06:10:58.527693 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6f667b6558-c67wt" event={"ID":"69484422-8317-4050-945c-9c9b127552b1","Type":"ContainerStarted","Data":"ea831d62c48ba78670eb897540a0b9a896ce935e61e5439a215f75c7478d0e2d"} Dec 06 06:10:58 crc kubenswrapper[4809]: I1206 06:10:58.536744 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-765b7d7f99-8dngt" Dec 06 06:10:58 crc kubenswrapper[4809]: I1206 06:10:58.987103 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-765b7d7f99-8dngt"] Dec 06 06:10:58 crc kubenswrapper[4809]: W1206 06:10:58.987422 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod60798ea0_75be_4a0c_a8c6_9fb431ba0e67.slice/crio-18d145374d875c256a33ab009c33e8a7d4b9bf95ccbe40f598ded1a338784088 WatchSource:0}: Error finding container 18d145374d875c256a33ab009c33e8a7d4b9bf95ccbe40f598ded1a338784088: Status 404 returned error can't find the container with id 18d145374d875c256a33ab009c33e8a7d4b9bf95ccbe40f598ded1a338784088 Dec 06 06:10:59 crc kubenswrapper[4809]: I1206 06:10:59.542695 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-765b7d7f99-8dngt" event={"ID":"60798ea0-75be-4a0c-a8c6-9fb431ba0e67","Type":"ContainerStarted","Data":"18d145374d875c256a33ab009c33e8a7d4b9bf95ccbe40f598ded1a338784088"} Dec 06 06:11:02 crc kubenswrapper[4809]: I1206 06:11:02.583043 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6f667b6558-c67wt" event={"ID":"69484422-8317-4050-945c-9c9b127552b1","Type":"ContainerStarted","Data":"637e1811ab433c9dfe08139379c1881512e228702f53e7cf1fabff15cfc85d09"} Dec 06 06:11:02 crc kubenswrapper[4809]: I1206 06:11:02.584282 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-6f667b6558-c67wt" Dec 06 06:11:04 crc kubenswrapper[4809]: I1206 06:11:04.496958 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:11:04 crc kubenswrapper[4809]: I1206 06:11:04.497308 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:11:05 crc kubenswrapper[4809]: I1206 06:11:05.424098 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-6f667b6558-c67wt" podStartSLOduration=5.439470254 podStartE2EDuration="8.424077843s" podCreationTimestamp="2025-12-06 06:10:57 +0000 UTC" firstStartedPulling="2025-12-06 06:10:58.516716503 +0000 UTC m=+1183.405699435" lastFinishedPulling="2025-12-06 06:11:01.501324072 +0000 UTC m=+1186.390307024" observedRunningTime="2025-12-06 06:11:02.608325802 +0000 UTC m=+1187.497308754" watchObservedRunningTime="2025-12-06 06:11:05.424077843 +0000 UTC m=+1190.313060785" Dec 06 06:11:05 crc kubenswrapper[4809]: I1206 06:11:05.627790 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-765b7d7f99-8dngt" event={"ID":"60798ea0-75be-4a0c-a8c6-9fb431ba0e67","Type":"ContainerStarted","Data":"ad1f567da5ebb2f3573fa8e2012447f1d943b0c935615bfc016344d75011d52d"} Dec 06 06:11:05 crc kubenswrapper[4809]: I1206 06:11:05.628466 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-765b7d7f99-8dngt" Dec 06 06:11:16 crc kubenswrapper[4809]: I1206 06:11:16.155783 4809 scope.go:117] "RemoveContainer" containerID="d86f9164d10d620c990222a1d697eef8646f314e472220d3c0405add4a8eb3fe" Dec 06 06:11:16 crc kubenswrapper[4809]: I1206 06:11:16.183329 4809 scope.go:117] "RemoveContainer" containerID="e6a979852ffe0757e6bd2d20d548ff969847341def135a6c0dd423eb2ef71951" Dec 06 06:11:16 crc kubenswrapper[4809]: I1206 06:11:16.225761 4809 scope.go:117] "RemoveContainer" containerID="8ef2d96a8b34261df3655cd85565e0a2642719e0394a5536122dba6ca33a07a0" Dec 06 06:11:18 crc kubenswrapper[4809]: I1206 06:11:18.543742 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-765b7d7f99-8dngt" Dec 06 06:11:18 crc kubenswrapper[4809]: I1206 06:11:18.565772 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-765b7d7f99-8dngt" podStartSLOduration=14.445524512 podStartE2EDuration="20.565740215s" podCreationTimestamp="2025-12-06 06:10:58 +0000 UTC" firstStartedPulling="2025-12-06 06:10:58.992586118 +0000 UTC m=+1183.881569060" lastFinishedPulling="2025-12-06 06:11:05.112801821 +0000 UTC m=+1190.001784763" observedRunningTime="2025-12-06 06:11:05.656504717 +0000 UTC m=+1190.545487669" watchObservedRunningTime="2025-12-06 06:11:18.565740215 +0000 UTC m=+1203.454723157" Dec 06 06:11:34 crc kubenswrapper[4809]: I1206 06:11:34.496923 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:11:34 crc kubenswrapper[4809]: I1206 06:11:34.497488 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:11:37 crc kubenswrapper[4809]: I1206 06:11:37.950584 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-6f667b6558-c67wt" Dec 06 06:11:38 crc kubenswrapper[4809]: I1206 06:11:38.962473 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-z95cs"] Dec 06 06:11:38 crc kubenswrapper[4809]: I1206 06:11:38.971242 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-z95cs" Dec 06 06:11:38 crc kubenswrapper[4809]: I1206 06:11:38.977143 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 06 06:11:38 crc kubenswrapper[4809]: I1206 06:11:38.977280 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-5h7xk" Dec 06 06:11:38 crc kubenswrapper[4809]: I1206 06:11:38.982015 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 06 06:11:38 crc kubenswrapper[4809]: I1206 06:11:38.991303 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-ttn9g"] Dec 06 06:11:38 crc kubenswrapper[4809]: I1206 06:11:38.993415 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-ttn9g" Dec 06 06:11:38 crc kubenswrapper[4809]: I1206 06:11:38.996218 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.015267 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-ttn9g"] Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.068738 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-5t9z6"] Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.074360 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-5t9z6" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.079398 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.079605 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-8hp28" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.079730 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.079822 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.081954 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-d4fgg"] Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.085154 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-d4fgg" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.098454 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.100063 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/a3b6197b-05d4-49b2-9a67-56b675046a1d-reloader\") pod \"frr-k8s-z95cs\" (UID: \"a3b6197b-05d4-49b2-9a67-56b675046a1d\") " pod="metallb-system/frr-k8s-z95cs" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.100110 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a3b6197b-05d4-49b2-9a67-56b675046a1d-metrics-certs\") pod \"frr-k8s-z95cs\" (UID: \"a3b6197b-05d4-49b2-9a67-56b675046a1d\") " pod="metallb-system/frr-k8s-z95cs" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.100158 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlf2n\" (UniqueName: \"kubernetes.io/projected/a3b6197b-05d4-49b2-9a67-56b675046a1d-kube-api-access-rlf2n\") pod \"frr-k8s-z95cs\" (UID: \"a3b6197b-05d4-49b2-9a67-56b675046a1d\") " pod="metallb-system/frr-k8s-z95cs" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.100200 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/a3b6197b-05d4-49b2-9a67-56b675046a1d-metrics\") pod \"frr-k8s-z95cs\" (UID: \"a3b6197b-05d4-49b2-9a67-56b675046a1d\") " pod="metallb-system/frr-k8s-z95cs" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.100263 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/a3b6197b-05d4-49b2-9a67-56b675046a1d-frr-conf\") pod \"frr-k8s-z95cs\" (UID: \"a3b6197b-05d4-49b2-9a67-56b675046a1d\") " pod="metallb-system/frr-k8s-z95cs" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.100280 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/a3b6197b-05d4-49b2-9a67-56b675046a1d-frr-sockets\") pod \"frr-k8s-z95cs\" (UID: \"a3b6197b-05d4-49b2-9a67-56b675046a1d\") " pod="metallb-system/frr-k8s-z95cs" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.100310 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b483382f-c2e3-40bf-ab4b-b58c2f3f9132-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-ttn9g\" (UID: \"b483382f-c2e3-40bf-ab4b-b58c2f3f9132\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-ttn9g" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.100326 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5vzl\" (UniqueName: \"kubernetes.io/projected/b483382f-c2e3-40bf-ab4b-b58c2f3f9132-kube-api-access-q5vzl\") pod \"frr-k8s-webhook-server-7fcb986d4-ttn9g\" (UID: \"b483382f-c2e3-40bf-ab4b-b58c2f3f9132\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-ttn9g" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.100417 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/a3b6197b-05d4-49b2-9a67-56b675046a1d-frr-startup\") pod \"frr-k8s-z95cs\" (UID: \"a3b6197b-05d4-49b2-9a67-56b675046a1d\") " pod="metallb-system/frr-k8s-z95cs" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.109714 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-d4fgg"] Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.201711 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/21411d89-4a48-4556-8ad2-193f1c8be0db-metallb-excludel2\") pod \"speaker-5t9z6\" (UID: \"21411d89-4a48-4556-8ad2-193f1c8be0db\") " pod="metallb-system/speaker-5t9z6" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.201771 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/a3b6197b-05d4-49b2-9a67-56b675046a1d-frr-conf\") pod \"frr-k8s-z95cs\" (UID: \"a3b6197b-05d4-49b2-9a67-56b675046a1d\") " pod="metallb-system/frr-k8s-z95cs" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.201802 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/a3b6197b-05d4-49b2-9a67-56b675046a1d-frr-sockets\") pod \"frr-k8s-z95cs\" (UID: \"a3b6197b-05d4-49b2-9a67-56b675046a1d\") " pod="metallb-system/frr-k8s-z95cs" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.201829 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b483382f-c2e3-40bf-ab4b-b58c2f3f9132-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-ttn9g\" (UID: \"b483382f-c2e3-40bf-ab4b-b58c2f3f9132\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-ttn9g" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.201849 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5vzl\" (UniqueName: \"kubernetes.io/projected/b483382f-c2e3-40bf-ab4b-b58c2f3f9132-kube-api-access-q5vzl\") pod \"frr-k8s-webhook-server-7fcb986d4-ttn9g\" (UID: \"b483382f-c2e3-40bf-ab4b-b58c2f3f9132\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-ttn9g" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.201888 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/a3b6197b-05d4-49b2-9a67-56b675046a1d-frr-startup\") pod \"frr-k8s-z95cs\" (UID: \"a3b6197b-05d4-49b2-9a67-56b675046a1d\") " pod="metallb-system/frr-k8s-z95cs" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.201916 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twrvc\" (UniqueName: \"kubernetes.io/projected/073938c5-bc80-47d3-91e8-1d6d7da0f000-kube-api-access-twrvc\") pod \"controller-f8648f98b-d4fgg\" (UID: \"073938c5-bc80-47d3-91e8-1d6d7da0f000\") " pod="metallb-system/controller-f8648f98b-d4fgg" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.201973 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/21411d89-4a48-4556-8ad2-193f1c8be0db-metrics-certs\") pod \"speaker-5t9z6\" (UID: \"21411d89-4a48-4556-8ad2-193f1c8be0db\") " pod="metallb-system/speaker-5t9z6" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.201994 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/21411d89-4a48-4556-8ad2-193f1c8be0db-memberlist\") pod \"speaker-5t9z6\" (UID: \"21411d89-4a48-4556-8ad2-193f1c8be0db\") " pod="metallb-system/speaker-5t9z6" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.202013 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zd82n\" (UniqueName: \"kubernetes.io/projected/21411d89-4a48-4556-8ad2-193f1c8be0db-kube-api-access-zd82n\") pod \"speaker-5t9z6\" (UID: \"21411d89-4a48-4556-8ad2-193f1c8be0db\") " pod="metallb-system/speaker-5t9z6" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.202038 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/a3b6197b-05d4-49b2-9a67-56b675046a1d-reloader\") pod \"frr-k8s-z95cs\" (UID: \"a3b6197b-05d4-49b2-9a67-56b675046a1d\") " pod="metallb-system/frr-k8s-z95cs" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.202056 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a3b6197b-05d4-49b2-9a67-56b675046a1d-metrics-certs\") pod \"frr-k8s-z95cs\" (UID: \"a3b6197b-05d4-49b2-9a67-56b675046a1d\") " pod="metallb-system/frr-k8s-z95cs" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.202218 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/a3b6197b-05d4-49b2-9a67-56b675046a1d-frr-conf\") pod \"frr-k8s-z95cs\" (UID: \"a3b6197b-05d4-49b2-9a67-56b675046a1d\") " pod="metallb-system/frr-k8s-z95cs" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.202525 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/a3b6197b-05d4-49b2-9a67-56b675046a1d-frr-sockets\") pod \"frr-k8s-z95cs\" (UID: \"a3b6197b-05d4-49b2-9a67-56b675046a1d\") " pod="metallb-system/frr-k8s-z95cs" Dec 06 06:11:39 crc kubenswrapper[4809]: E1206 06:11:39.203252 4809 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Dec 06 06:11:39 crc kubenswrapper[4809]: E1206 06:11:39.203325 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a3b6197b-05d4-49b2-9a67-56b675046a1d-metrics-certs podName:a3b6197b-05d4-49b2-9a67-56b675046a1d nodeName:}" failed. No retries permitted until 2025-12-06 06:11:39.703301642 +0000 UTC m=+1224.592284584 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a3b6197b-05d4-49b2-9a67-56b675046a1d-metrics-certs") pod "frr-k8s-z95cs" (UID: "a3b6197b-05d4-49b2-9a67-56b675046a1d") : secret "frr-k8s-certs-secret" not found Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.203372 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/a3b6197b-05d4-49b2-9a67-56b675046a1d-reloader\") pod \"frr-k8s-z95cs\" (UID: \"a3b6197b-05d4-49b2-9a67-56b675046a1d\") " pod="metallb-system/frr-k8s-z95cs" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.203474 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/073938c5-bc80-47d3-91e8-1d6d7da0f000-cert\") pod \"controller-f8648f98b-d4fgg\" (UID: \"073938c5-bc80-47d3-91e8-1d6d7da0f000\") " pod="metallb-system/controller-f8648f98b-d4fgg" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.203515 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlf2n\" (UniqueName: \"kubernetes.io/projected/a3b6197b-05d4-49b2-9a67-56b675046a1d-kube-api-access-rlf2n\") pod \"frr-k8s-z95cs\" (UID: \"a3b6197b-05d4-49b2-9a67-56b675046a1d\") " pod="metallb-system/frr-k8s-z95cs" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.203549 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/073938c5-bc80-47d3-91e8-1d6d7da0f000-metrics-certs\") pod \"controller-f8648f98b-d4fgg\" (UID: \"073938c5-bc80-47d3-91e8-1d6d7da0f000\") " pod="metallb-system/controller-f8648f98b-d4fgg" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.203577 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/a3b6197b-05d4-49b2-9a67-56b675046a1d-metrics\") pod \"frr-k8s-z95cs\" (UID: \"a3b6197b-05d4-49b2-9a67-56b675046a1d\") " pod="metallb-system/frr-k8s-z95cs" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.203784 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/a3b6197b-05d4-49b2-9a67-56b675046a1d-frr-startup\") pod \"frr-k8s-z95cs\" (UID: \"a3b6197b-05d4-49b2-9a67-56b675046a1d\") " pod="metallb-system/frr-k8s-z95cs" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.203812 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/a3b6197b-05d4-49b2-9a67-56b675046a1d-metrics\") pod \"frr-k8s-z95cs\" (UID: \"a3b6197b-05d4-49b2-9a67-56b675046a1d\") " pod="metallb-system/frr-k8s-z95cs" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.217573 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b483382f-c2e3-40bf-ab4b-b58c2f3f9132-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-ttn9g\" (UID: \"b483382f-c2e3-40bf-ab4b-b58c2f3f9132\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-ttn9g" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.223677 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5vzl\" (UniqueName: \"kubernetes.io/projected/b483382f-c2e3-40bf-ab4b-b58c2f3f9132-kube-api-access-q5vzl\") pod \"frr-k8s-webhook-server-7fcb986d4-ttn9g\" (UID: \"b483382f-c2e3-40bf-ab4b-b58c2f3f9132\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-ttn9g" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.224669 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlf2n\" (UniqueName: \"kubernetes.io/projected/a3b6197b-05d4-49b2-9a67-56b675046a1d-kube-api-access-rlf2n\") pod \"frr-k8s-z95cs\" (UID: \"a3b6197b-05d4-49b2-9a67-56b675046a1d\") " pod="metallb-system/frr-k8s-z95cs" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.305297 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zd82n\" (UniqueName: \"kubernetes.io/projected/21411d89-4a48-4556-8ad2-193f1c8be0db-kube-api-access-zd82n\") pod \"speaker-5t9z6\" (UID: \"21411d89-4a48-4556-8ad2-193f1c8be0db\") " pod="metallb-system/speaker-5t9z6" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.308011 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/073938c5-bc80-47d3-91e8-1d6d7da0f000-cert\") pod \"controller-f8648f98b-d4fgg\" (UID: \"073938c5-bc80-47d3-91e8-1d6d7da0f000\") " pod="metallb-system/controller-f8648f98b-d4fgg" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.308399 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/073938c5-bc80-47d3-91e8-1d6d7da0f000-metrics-certs\") pod \"controller-f8648f98b-d4fgg\" (UID: \"073938c5-bc80-47d3-91e8-1d6d7da0f000\") " pod="metallb-system/controller-f8648f98b-d4fgg" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.308789 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/21411d89-4a48-4556-8ad2-193f1c8be0db-metallb-excludel2\") pod \"speaker-5t9z6\" (UID: \"21411d89-4a48-4556-8ad2-193f1c8be0db\") " pod="metallb-system/speaker-5t9z6" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.309259 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twrvc\" (UniqueName: \"kubernetes.io/projected/073938c5-bc80-47d3-91e8-1d6d7da0f000-kube-api-access-twrvc\") pod \"controller-f8648f98b-d4fgg\" (UID: \"073938c5-bc80-47d3-91e8-1d6d7da0f000\") " pod="metallb-system/controller-f8648f98b-d4fgg" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.309448 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/21411d89-4a48-4556-8ad2-193f1c8be0db-metrics-certs\") pod \"speaker-5t9z6\" (UID: \"21411d89-4a48-4556-8ad2-193f1c8be0db\") " pod="metallb-system/speaker-5t9z6" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.309570 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/21411d89-4a48-4556-8ad2-193f1c8be0db-memberlist\") pod \"speaker-5t9z6\" (UID: \"21411d89-4a48-4556-8ad2-193f1c8be0db\") " pod="metallb-system/speaker-5t9z6" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.309598 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/21411d89-4a48-4556-8ad2-193f1c8be0db-metallb-excludel2\") pod \"speaker-5t9z6\" (UID: \"21411d89-4a48-4556-8ad2-193f1c8be0db\") " pod="metallb-system/speaker-5t9z6" Dec 06 06:11:39 crc kubenswrapper[4809]: E1206 06:11:39.309737 4809 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 06 06:11:39 crc kubenswrapper[4809]: E1206 06:11:39.310136 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/21411d89-4a48-4556-8ad2-193f1c8be0db-memberlist podName:21411d89-4a48-4556-8ad2-193f1c8be0db nodeName:}" failed. No retries permitted until 2025-12-06 06:11:39.810106745 +0000 UTC m=+1224.699089687 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/21411d89-4a48-4556-8ad2-193f1c8be0db-memberlist") pod "speaker-5t9z6" (UID: "21411d89-4a48-4556-8ad2-193f1c8be0db") : secret "metallb-memberlist" not found Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.311140 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.312799 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/21411d89-4a48-4556-8ad2-193f1c8be0db-metrics-certs\") pod \"speaker-5t9z6\" (UID: \"21411d89-4a48-4556-8ad2-193f1c8be0db\") " pod="metallb-system/speaker-5t9z6" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.313155 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/073938c5-bc80-47d3-91e8-1d6d7da0f000-metrics-certs\") pod \"controller-f8648f98b-d4fgg\" (UID: \"073938c5-bc80-47d3-91e8-1d6d7da0f000\") " pod="metallb-system/controller-f8648f98b-d4fgg" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.321818 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-ttn9g" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.322694 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/073938c5-bc80-47d3-91e8-1d6d7da0f000-cert\") pod \"controller-f8648f98b-d4fgg\" (UID: \"073938c5-bc80-47d3-91e8-1d6d7da0f000\") " pod="metallb-system/controller-f8648f98b-d4fgg" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.326334 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zd82n\" (UniqueName: \"kubernetes.io/projected/21411d89-4a48-4556-8ad2-193f1c8be0db-kube-api-access-zd82n\") pod \"speaker-5t9z6\" (UID: \"21411d89-4a48-4556-8ad2-193f1c8be0db\") " pod="metallb-system/speaker-5t9z6" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.326967 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twrvc\" (UniqueName: \"kubernetes.io/projected/073938c5-bc80-47d3-91e8-1d6d7da0f000-kube-api-access-twrvc\") pod \"controller-f8648f98b-d4fgg\" (UID: \"073938c5-bc80-47d3-91e8-1d6d7da0f000\") " pod="metallb-system/controller-f8648f98b-d4fgg" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.426752 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-d4fgg" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.724036 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a3b6197b-05d4-49b2-9a67-56b675046a1d-metrics-certs\") pod \"frr-k8s-z95cs\" (UID: \"a3b6197b-05d4-49b2-9a67-56b675046a1d\") " pod="metallb-system/frr-k8s-z95cs" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.730767 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a3b6197b-05d4-49b2-9a67-56b675046a1d-metrics-certs\") pod \"frr-k8s-z95cs\" (UID: \"a3b6197b-05d4-49b2-9a67-56b675046a1d\") " pod="metallb-system/frr-k8s-z95cs" Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.764505 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-ttn9g"] Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.825713 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/21411d89-4a48-4556-8ad2-193f1c8be0db-memberlist\") pod \"speaker-5t9z6\" (UID: \"21411d89-4a48-4556-8ad2-193f1c8be0db\") " pod="metallb-system/speaker-5t9z6" Dec 06 06:11:39 crc kubenswrapper[4809]: E1206 06:11:39.825972 4809 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 06 06:11:39 crc kubenswrapper[4809]: E1206 06:11:39.826221 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/21411d89-4a48-4556-8ad2-193f1c8be0db-memberlist podName:21411d89-4a48-4556-8ad2-193f1c8be0db nodeName:}" failed. No retries permitted until 2025-12-06 06:11:40.826199665 +0000 UTC m=+1225.715182607 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/21411d89-4a48-4556-8ad2-193f1c8be0db-memberlist") pod "speaker-5t9z6" (UID: "21411d89-4a48-4556-8ad2-193f1c8be0db") : secret "metallb-memberlist" not found Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.875205 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-d4fgg"] Dec 06 06:11:39 crc kubenswrapper[4809]: W1206 06:11:39.877230 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod073938c5_bc80_47d3_91e8_1d6d7da0f000.slice/crio-d9496868086dc614a1eb9b44fbe91c8d09cffa56f0392eb82fd0c18eab975d02 WatchSource:0}: Error finding container d9496868086dc614a1eb9b44fbe91c8d09cffa56f0392eb82fd0c18eab975d02: Status 404 returned error can't find the container with id d9496868086dc614a1eb9b44fbe91c8d09cffa56f0392eb82fd0c18eab975d02 Dec 06 06:11:39 crc kubenswrapper[4809]: I1206 06:11:39.897677 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-z95cs" Dec 06 06:11:40 crc kubenswrapper[4809]: I1206 06:11:40.127518 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-d4fgg" event={"ID":"073938c5-bc80-47d3-91e8-1d6d7da0f000","Type":"ContainerStarted","Data":"33056281dc387b10a749a489bf6ecfe9f48b49ce62da155ba063b4c35c1a75d5"} Dec 06 06:11:40 crc kubenswrapper[4809]: I1206 06:11:40.127587 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-d4fgg" event={"ID":"073938c5-bc80-47d3-91e8-1d6d7da0f000","Type":"ContainerStarted","Data":"d9496868086dc614a1eb9b44fbe91c8d09cffa56f0392eb82fd0c18eab975d02"} Dec 06 06:11:40 crc kubenswrapper[4809]: I1206 06:11:40.129559 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-ttn9g" event={"ID":"b483382f-c2e3-40bf-ab4b-b58c2f3f9132","Type":"ContainerStarted","Data":"ce783a93cc65fedb87cdc68011d82cca911d70ca42c119fba240784fcd51b993"} Dec 06 06:11:40 crc kubenswrapper[4809]: I1206 06:11:40.130747 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-z95cs" event={"ID":"a3b6197b-05d4-49b2-9a67-56b675046a1d","Type":"ContainerStarted","Data":"90ec887d5f6750ffc425d90ed0758d7e41f85e900a1ab3ea47983173dd9c2681"} Dec 06 06:11:40 crc kubenswrapper[4809]: I1206 06:11:40.848687 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/21411d89-4a48-4556-8ad2-193f1c8be0db-memberlist\") pod \"speaker-5t9z6\" (UID: \"21411d89-4a48-4556-8ad2-193f1c8be0db\") " pod="metallb-system/speaker-5t9z6" Dec 06 06:11:40 crc kubenswrapper[4809]: I1206 06:11:40.855421 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/21411d89-4a48-4556-8ad2-193f1c8be0db-memberlist\") pod \"speaker-5t9z6\" (UID: \"21411d89-4a48-4556-8ad2-193f1c8be0db\") " pod="metallb-system/speaker-5t9z6" Dec 06 06:11:40 crc kubenswrapper[4809]: I1206 06:11:40.911442 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-5t9z6" Dec 06 06:11:41 crc kubenswrapper[4809]: I1206 06:11:41.143819 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-5t9z6" event={"ID":"21411d89-4a48-4556-8ad2-193f1c8be0db","Type":"ContainerStarted","Data":"276fd1bec0516369d52b0ab1b23ebd9d711c18ec31f6330d6ac78cb5bd27dc2b"} Dec 06 06:11:41 crc kubenswrapper[4809]: I1206 06:11:41.164279 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-d4fgg" event={"ID":"073938c5-bc80-47d3-91e8-1d6d7da0f000","Type":"ContainerStarted","Data":"bf466223c86cb6e998dfd4cbbc1fb1aa01447c5f324dd5b4a38d94b0e0636261"} Dec 06 06:11:41 crc kubenswrapper[4809]: I1206 06:11:41.165007 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-d4fgg" Dec 06 06:11:41 crc kubenswrapper[4809]: I1206 06:11:41.198358 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-d4fgg" podStartSLOduration=2.19831138 podStartE2EDuration="2.19831138s" podCreationTimestamp="2025-12-06 06:11:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:11:41.188136226 +0000 UTC m=+1226.077119168" watchObservedRunningTime="2025-12-06 06:11:41.19831138 +0000 UTC m=+1226.087294332" Dec 06 06:11:42 crc kubenswrapper[4809]: I1206 06:11:42.179025 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-5t9z6" event={"ID":"21411d89-4a48-4556-8ad2-193f1c8be0db","Type":"ContainerStarted","Data":"4833e06abe461d59123e3e07ca03f43ef414d1ca555de3b3a12f7c96b996987c"} Dec 06 06:11:42 crc kubenswrapper[4809]: I1206 06:11:42.179336 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-5t9z6" event={"ID":"21411d89-4a48-4556-8ad2-193f1c8be0db","Type":"ContainerStarted","Data":"37d537a2579de6465d1abec7d6ae058c1d93e0ff3f00d12fa614427d15762e7a"} Dec 06 06:11:43 crc kubenswrapper[4809]: I1206 06:11:43.188953 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-5t9z6" Dec 06 06:11:45 crc kubenswrapper[4809]: I1206 06:11:45.434845 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-5t9z6" podStartSLOduration=6.434815359 podStartE2EDuration="6.434815359s" podCreationTimestamp="2025-12-06 06:11:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:11:42.210296745 +0000 UTC m=+1227.099279687" watchObservedRunningTime="2025-12-06 06:11:45.434815359 +0000 UTC m=+1230.323798311" Dec 06 06:11:48 crc kubenswrapper[4809]: I1206 06:11:48.249338 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-ttn9g" event={"ID":"b483382f-c2e3-40bf-ab4b-b58c2f3f9132","Type":"ContainerStarted","Data":"e7a53ea59878c173ebecb490955772a29252510c4ac27e6564be54f178d968d9"} Dec 06 06:11:48 crc kubenswrapper[4809]: I1206 06:11:48.250200 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-ttn9g" Dec 06 06:11:48 crc kubenswrapper[4809]: I1206 06:11:48.251920 4809 generic.go:334] "Generic (PLEG): container finished" podID="a3b6197b-05d4-49b2-9a67-56b675046a1d" containerID="a4a7b422932ed2ecfdd0bc867cf31ff409534d02af3e273df5d3e86a91c4ec4b" exitCode=0 Dec 06 06:11:48 crc kubenswrapper[4809]: I1206 06:11:48.252007 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-z95cs" event={"ID":"a3b6197b-05d4-49b2-9a67-56b675046a1d","Type":"ContainerDied","Data":"a4a7b422932ed2ecfdd0bc867cf31ff409534d02af3e273df5d3e86a91c4ec4b"} Dec 06 06:11:48 crc kubenswrapper[4809]: I1206 06:11:48.274204 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-ttn9g" podStartSLOduration=2.434915475 podStartE2EDuration="10.274175828s" podCreationTimestamp="2025-12-06 06:11:38 +0000 UTC" firstStartedPulling="2025-12-06 06:11:39.771894059 +0000 UTC m=+1224.660877001" lastFinishedPulling="2025-12-06 06:11:47.611154372 +0000 UTC m=+1232.500137354" observedRunningTime="2025-12-06 06:11:48.267370974 +0000 UTC m=+1233.156353926" watchObservedRunningTime="2025-12-06 06:11:48.274175828 +0000 UTC m=+1233.163158770" Dec 06 06:11:49 crc kubenswrapper[4809]: I1206 06:11:49.294515 4809 generic.go:334] "Generic (PLEG): container finished" podID="a3b6197b-05d4-49b2-9a67-56b675046a1d" containerID="55ae55fbbe1bb1e93a46ad0107b63e2b52a23ded53c89a1143376f194736d474" exitCode=0 Dec 06 06:11:49 crc kubenswrapper[4809]: I1206 06:11:49.294615 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-z95cs" event={"ID":"a3b6197b-05d4-49b2-9a67-56b675046a1d","Type":"ContainerDied","Data":"55ae55fbbe1bb1e93a46ad0107b63e2b52a23ded53c89a1143376f194736d474"} Dec 06 06:11:50 crc kubenswrapper[4809]: I1206 06:11:50.308265 4809 generic.go:334] "Generic (PLEG): container finished" podID="a3b6197b-05d4-49b2-9a67-56b675046a1d" containerID="d54ae8481553df1fffba78d9daec8e93d2d1ac8d316c983237b524a388c1f930" exitCode=0 Dec 06 06:11:50 crc kubenswrapper[4809]: I1206 06:11:50.308883 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-z95cs" event={"ID":"a3b6197b-05d4-49b2-9a67-56b675046a1d","Type":"ContainerDied","Data":"d54ae8481553df1fffba78d9daec8e93d2d1ac8d316c983237b524a388c1f930"} Dec 06 06:11:51 crc kubenswrapper[4809]: I1206 06:11:51.322411 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-z95cs" event={"ID":"a3b6197b-05d4-49b2-9a67-56b675046a1d","Type":"ContainerStarted","Data":"a142613f5005612f27fe1cd29d5982f40edcf6d20b662a7d616493ef24f67a06"} Dec 06 06:11:51 crc kubenswrapper[4809]: I1206 06:11:51.324060 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-z95cs" event={"ID":"a3b6197b-05d4-49b2-9a67-56b675046a1d","Type":"ContainerStarted","Data":"d2f78dc928e7d574f9735b631dc72a6942893122be0c54f51f44a44b05417428"} Dec 06 06:11:52 crc kubenswrapper[4809]: I1206 06:11:52.337284 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-z95cs" event={"ID":"a3b6197b-05d4-49b2-9a67-56b675046a1d","Type":"ContainerStarted","Data":"a0efec97117769c9990f20aafd08691eec6224d15f57805a508d47b68b1a8d02"} Dec 06 06:11:52 crc kubenswrapper[4809]: I1206 06:11:52.337726 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-z95cs" event={"ID":"a3b6197b-05d4-49b2-9a67-56b675046a1d","Type":"ContainerStarted","Data":"5bc5d490b42aa27f602c85bcc086fb7abe4ada3f4f32bfc33e1cd9191fd69984"} Dec 06 06:11:52 crc kubenswrapper[4809]: I1206 06:11:52.337758 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-z95cs" Dec 06 06:11:52 crc kubenswrapper[4809]: I1206 06:11:52.337777 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-z95cs" event={"ID":"a3b6197b-05d4-49b2-9a67-56b675046a1d","Type":"ContainerStarted","Data":"faa431a6ca43d7628f8e909e0b52d28659966fb2442b770d57577597cb2e34d9"} Dec 06 06:11:52 crc kubenswrapper[4809]: I1206 06:11:52.337797 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-z95cs" event={"ID":"a3b6197b-05d4-49b2-9a67-56b675046a1d","Type":"ContainerStarted","Data":"620140909b8df59ae70f0dbab8e3f7a46073bd4d3ca3e9e65bdb3e31e66d75ed"} Dec 06 06:11:52 crc kubenswrapper[4809]: I1206 06:11:52.373515 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-z95cs" podStartSLOduration=6.788861634 podStartE2EDuration="14.373478695s" podCreationTimestamp="2025-12-06 06:11:38 +0000 UTC" firstStartedPulling="2025-12-06 06:11:40.035475993 +0000 UTC m=+1224.924458935" lastFinishedPulling="2025-12-06 06:11:47.620093054 +0000 UTC m=+1232.509075996" observedRunningTime="2025-12-06 06:11:52.367099533 +0000 UTC m=+1237.256082475" watchObservedRunningTime="2025-12-06 06:11:52.373478695 +0000 UTC m=+1237.262461817" Dec 06 06:11:54 crc kubenswrapper[4809]: I1206 06:11:54.898879 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-z95cs" Dec 06 06:11:54 crc kubenswrapper[4809]: I1206 06:11:54.936746 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-z95cs" Dec 06 06:11:59 crc kubenswrapper[4809]: I1206 06:11:59.328687 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-ttn9g" Dec 06 06:11:59 crc kubenswrapper[4809]: I1206 06:11:59.433583 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-d4fgg" Dec 06 06:12:00 crc kubenswrapper[4809]: I1206 06:12:00.917440 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-5t9z6" Dec 06 06:12:04 crc kubenswrapper[4809]: I1206 06:12:04.496378 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:12:04 crc kubenswrapper[4809]: I1206 06:12:04.497301 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:12:04 crc kubenswrapper[4809]: I1206 06:12:04.497395 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-npms2" Dec 06 06:12:04 crc kubenswrapper[4809]: I1206 06:12:04.498650 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9b0b6808cda0fc2270940d0650cf854006725ea29b339e76c18a025e4f361d32"} pod="openshift-machine-config-operator/machine-config-daemon-npms2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 06:12:04 crc kubenswrapper[4809]: I1206 06:12:04.498737 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" containerID="cri-o://9b0b6808cda0fc2270940d0650cf854006725ea29b339e76c18a025e4f361d32" gracePeriod=600 Dec 06 06:12:05 crc kubenswrapper[4809]: I1206 06:12:05.461302 4809 generic.go:334] "Generic (PLEG): container finished" podID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerID="9b0b6808cda0fc2270940d0650cf854006725ea29b339e76c18a025e4f361d32" exitCode=0 Dec 06 06:12:05 crc kubenswrapper[4809]: I1206 06:12:05.461414 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" event={"ID":"cbf4cf62-024e-4703-a8b8-9aecda9cd26a","Type":"ContainerDied","Data":"9b0b6808cda0fc2270940d0650cf854006725ea29b339e76c18a025e4f361d32"} Dec 06 06:12:05 crc kubenswrapper[4809]: I1206 06:12:05.462176 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" event={"ID":"cbf4cf62-024e-4703-a8b8-9aecda9cd26a","Type":"ContainerStarted","Data":"fac0817a1ba54b752b764ab02ba5dbd6db50ecfcf2de9f18d391e8a61207257c"} Dec 06 06:12:05 crc kubenswrapper[4809]: I1206 06:12:05.462229 4809 scope.go:117] "RemoveContainer" containerID="bb59482414fdfbb9268b139562a808d2a5e04bbf56acc96dd715da16e5bd5913" Dec 06 06:12:09 crc kubenswrapper[4809]: I1206 06:12:09.900302 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-z95cs" Dec 06 06:12:11 crc kubenswrapper[4809]: I1206 06:12:11.028106 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-hv5l4"] Dec 06 06:12:11 crc kubenswrapper[4809]: I1206 06:12:11.029741 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-hv5l4" Dec 06 06:12:11 crc kubenswrapper[4809]: I1206 06:12:11.033054 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 06 06:12:11 crc kubenswrapper[4809]: I1206 06:12:11.033206 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 06 06:12:11 crc kubenswrapper[4809]: I1206 06:12:11.033474 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-tqnm7" Dec 06 06:12:11 crc kubenswrapper[4809]: I1206 06:12:11.046656 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-hv5l4"] Dec 06 06:12:11 crc kubenswrapper[4809]: I1206 06:12:11.087959 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fn8fh\" (UniqueName: \"kubernetes.io/projected/45b1df34-ffd2-42ca-85ec-885efb8ec7e5-kube-api-access-fn8fh\") pod \"openstack-operator-index-hv5l4\" (UID: \"45b1df34-ffd2-42ca-85ec-885efb8ec7e5\") " pod="openstack-operators/openstack-operator-index-hv5l4" Dec 06 06:12:11 crc kubenswrapper[4809]: I1206 06:12:11.190228 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fn8fh\" (UniqueName: \"kubernetes.io/projected/45b1df34-ffd2-42ca-85ec-885efb8ec7e5-kube-api-access-fn8fh\") pod \"openstack-operator-index-hv5l4\" (UID: \"45b1df34-ffd2-42ca-85ec-885efb8ec7e5\") " pod="openstack-operators/openstack-operator-index-hv5l4" Dec 06 06:12:11 crc kubenswrapper[4809]: I1206 06:12:11.482994 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fn8fh\" (UniqueName: \"kubernetes.io/projected/45b1df34-ffd2-42ca-85ec-885efb8ec7e5-kube-api-access-fn8fh\") pod \"openstack-operator-index-hv5l4\" (UID: \"45b1df34-ffd2-42ca-85ec-885efb8ec7e5\") " pod="openstack-operators/openstack-operator-index-hv5l4" Dec 06 06:12:11 crc kubenswrapper[4809]: I1206 06:12:11.653789 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-hv5l4" Dec 06 06:12:12 crc kubenswrapper[4809]: I1206 06:12:12.138906 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-hv5l4"] Dec 06 06:12:12 crc kubenswrapper[4809]: I1206 06:12:12.532087 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-hv5l4" event={"ID":"45b1df34-ffd2-42ca-85ec-885efb8ec7e5","Type":"ContainerStarted","Data":"f615adb81a3298cf7177e5d00cba57ccfdac3ba1e05c5a07455005c6f39f6210"} Dec 06 06:12:17 crc kubenswrapper[4809]: I1206 06:12:17.583820 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-hv5l4" event={"ID":"45b1df34-ffd2-42ca-85ec-885efb8ec7e5","Type":"ContainerStarted","Data":"883997c35bc3a4c8dfd22fd89e02078ab5f7af9c1b30eef358d157d9b356db04"} Dec 06 06:12:17 crc kubenswrapper[4809]: I1206 06:12:17.611775 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-hv5l4" podStartSLOduration=2.229549109 podStartE2EDuration="6.611751853s" podCreationTimestamp="2025-12-06 06:12:11 +0000 UTC" firstStartedPulling="2025-12-06 06:12:12.147848516 +0000 UTC m=+1257.036831468" lastFinishedPulling="2025-12-06 06:12:16.53005127 +0000 UTC m=+1261.419034212" observedRunningTime="2025-12-06 06:12:17.602618097 +0000 UTC m=+1262.491601029" watchObservedRunningTime="2025-12-06 06:12:17.611751853 +0000 UTC m=+1262.500734795" Dec 06 06:12:21 crc kubenswrapper[4809]: I1206 06:12:21.654972 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-hv5l4" Dec 06 06:12:21 crc kubenswrapper[4809]: I1206 06:12:21.655598 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-hv5l4" Dec 06 06:12:21 crc kubenswrapper[4809]: I1206 06:12:21.691577 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-hv5l4" Dec 06 06:12:22 crc kubenswrapper[4809]: I1206 06:12:22.670473 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-hv5l4" Dec 06 06:12:24 crc kubenswrapper[4809]: I1206 06:12:24.266529 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/2a8c7bb1b949e5c76441d571adf78af7c2049b664ecf2ad024c54dd9ba67kbg"] Dec 06 06:12:24 crc kubenswrapper[4809]: I1206 06:12:24.268573 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/2a8c7bb1b949e5c76441d571adf78af7c2049b664ecf2ad024c54dd9ba67kbg" Dec 06 06:12:24 crc kubenswrapper[4809]: I1206 06:12:24.270919 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-zbkjv" Dec 06 06:12:24 crc kubenswrapper[4809]: I1206 06:12:24.278212 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/2a8c7bb1b949e5c76441d571adf78af7c2049b664ecf2ad024c54dd9ba67kbg"] Dec 06 06:12:24 crc kubenswrapper[4809]: I1206 06:12:24.381403 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/99d85dd8-4137-4876-9062-3c23ee0bb730-bundle\") pod \"2a8c7bb1b949e5c76441d571adf78af7c2049b664ecf2ad024c54dd9ba67kbg\" (UID: \"99d85dd8-4137-4876-9062-3c23ee0bb730\") " pod="openstack-operators/2a8c7bb1b949e5c76441d571adf78af7c2049b664ecf2ad024c54dd9ba67kbg" Dec 06 06:12:24 crc kubenswrapper[4809]: I1206 06:12:24.381571 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/99d85dd8-4137-4876-9062-3c23ee0bb730-util\") pod \"2a8c7bb1b949e5c76441d571adf78af7c2049b664ecf2ad024c54dd9ba67kbg\" (UID: \"99d85dd8-4137-4876-9062-3c23ee0bb730\") " pod="openstack-operators/2a8c7bb1b949e5c76441d571adf78af7c2049b664ecf2ad024c54dd9ba67kbg" Dec 06 06:12:24 crc kubenswrapper[4809]: I1206 06:12:24.381612 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n75rb\" (UniqueName: \"kubernetes.io/projected/99d85dd8-4137-4876-9062-3c23ee0bb730-kube-api-access-n75rb\") pod \"2a8c7bb1b949e5c76441d571adf78af7c2049b664ecf2ad024c54dd9ba67kbg\" (UID: \"99d85dd8-4137-4876-9062-3c23ee0bb730\") " pod="openstack-operators/2a8c7bb1b949e5c76441d571adf78af7c2049b664ecf2ad024c54dd9ba67kbg" Dec 06 06:12:24 crc kubenswrapper[4809]: I1206 06:12:24.483368 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/99d85dd8-4137-4876-9062-3c23ee0bb730-bundle\") pod \"2a8c7bb1b949e5c76441d571adf78af7c2049b664ecf2ad024c54dd9ba67kbg\" (UID: \"99d85dd8-4137-4876-9062-3c23ee0bb730\") " pod="openstack-operators/2a8c7bb1b949e5c76441d571adf78af7c2049b664ecf2ad024c54dd9ba67kbg" Dec 06 06:12:24 crc kubenswrapper[4809]: I1206 06:12:24.483492 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/99d85dd8-4137-4876-9062-3c23ee0bb730-util\") pod \"2a8c7bb1b949e5c76441d571adf78af7c2049b664ecf2ad024c54dd9ba67kbg\" (UID: \"99d85dd8-4137-4876-9062-3c23ee0bb730\") " pod="openstack-operators/2a8c7bb1b949e5c76441d571adf78af7c2049b664ecf2ad024c54dd9ba67kbg" Dec 06 06:12:24 crc kubenswrapper[4809]: I1206 06:12:24.483545 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n75rb\" (UniqueName: \"kubernetes.io/projected/99d85dd8-4137-4876-9062-3c23ee0bb730-kube-api-access-n75rb\") pod \"2a8c7bb1b949e5c76441d571adf78af7c2049b664ecf2ad024c54dd9ba67kbg\" (UID: \"99d85dd8-4137-4876-9062-3c23ee0bb730\") " pod="openstack-operators/2a8c7bb1b949e5c76441d571adf78af7c2049b664ecf2ad024c54dd9ba67kbg" Dec 06 06:12:24 crc kubenswrapper[4809]: I1206 06:12:24.484398 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/99d85dd8-4137-4876-9062-3c23ee0bb730-bundle\") pod \"2a8c7bb1b949e5c76441d571adf78af7c2049b664ecf2ad024c54dd9ba67kbg\" (UID: \"99d85dd8-4137-4876-9062-3c23ee0bb730\") " pod="openstack-operators/2a8c7bb1b949e5c76441d571adf78af7c2049b664ecf2ad024c54dd9ba67kbg" Dec 06 06:12:24 crc kubenswrapper[4809]: I1206 06:12:24.484506 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/99d85dd8-4137-4876-9062-3c23ee0bb730-util\") pod \"2a8c7bb1b949e5c76441d571adf78af7c2049b664ecf2ad024c54dd9ba67kbg\" (UID: \"99d85dd8-4137-4876-9062-3c23ee0bb730\") " pod="openstack-operators/2a8c7bb1b949e5c76441d571adf78af7c2049b664ecf2ad024c54dd9ba67kbg" Dec 06 06:12:24 crc kubenswrapper[4809]: I1206 06:12:24.517917 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n75rb\" (UniqueName: \"kubernetes.io/projected/99d85dd8-4137-4876-9062-3c23ee0bb730-kube-api-access-n75rb\") pod \"2a8c7bb1b949e5c76441d571adf78af7c2049b664ecf2ad024c54dd9ba67kbg\" (UID: \"99d85dd8-4137-4876-9062-3c23ee0bb730\") " pod="openstack-operators/2a8c7bb1b949e5c76441d571adf78af7c2049b664ecf2ad024c54dd9ba67kbg" Dec 06 06:12:24 crc kubenswrapper[4809]: I1206 06:12:24.587865 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/2a8c7bb1b949e5c76441d571adf78af7c2049b664ecf2ad024c54dd9ba67kbg" Dec 06 06:12:25 crc kubenswrapper[4809]: W1206 06:12:25.050273 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod99d85dd8_4137_4876_9062_3c23ee0bb730.slice/crio-476dca1fce4ee7991749b26309960ab17e95faf98a81ab202851ef6ec33d2adc WatchSource:0}: Error finding container 476dca1fce4ee7991749b26309960ab17e95faf98a81ab202851ef6ec33d2adc: Status 404 returned error can't find the container with id 476dca1fce4ee7991749b26309960ab17e95faf98a81ab202851ef6ec33d2adc Dec 06 06:12:25 crc kubenswrapper[4809]: I1206 06:12:25.053392 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/2a8c7bb1b949e5c76441d571adf78af7c2049b664ecf2ad024c54dd9ba67kbg"] Dec 06 06:12:25 crc kubenswrapper[4809]: I1206 06:12:25.660438 4809 generic.go:334] "Generic (PLEG): container finished" podID="99d85dd8-4137-4876-9062-3c23ee0bb730" containerID="6294be3b8a96d8ba05ddf01ebfd1468115f8d2fc3dedeb9a004a9a47399f947c" exitCode=0 Dec 06 06:12:25 crc kubenswrapper[4809]: I1206 06:12:25.660506 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/2a8c7bb1b949e5c76441d571adf78af7c2049b664ecf2ad024c54dd9ba67kbg" event={"ID":"99d85dd8-4137-4876-9062-3c23ee0bb730","Type":"ContainerDied","Data":"6294be3b8a96d8ba05ddf01ebfd1468115f8d2fc3dedeb9a004a9a47399f947c"} Dec 06 06:12:25 crc kubenswrapper[4809]: I1206 06:12:25.660787 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/2a8c7bb1b949e5c76441d571adf78af7c2049b664ecf2ad024c54dd9ba67kbg" event={"ID":"99d85dd8-4137-4876-9062-3c23ee0bb730","Type":"ContainerStarted","Data":"476dca1fce4ee7991749b26309960ab17e95faf98a81ab202851ef6ec33d2adc"} Dec 06 06:12:26 crc kubenswrapper[4809]: I1206 06:12:26.672923 4809 generic.go:334] "Generic (PLEG): container finished" podID="99d85dd8-4137-4876-9062-3c23ee0bb730" containerID="af3e960c27e095a824694cb74c78df4b0e65f703bcb9e43191c5f7cf2b8795d5" exitCode=0 Dec 06 06:12:26 crc kubenswrapper[4809]: I1206 06:12:26.673017 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/2a8c7bb1b949e5c76441d571adf78af7c2049b664ecf2ad024c54dd9ba67kbg" event={"ID":"99d85dd8-4137-4876-9062-3c23ee0bb730","Type":"ContainerDied","Data":"af3e960c27e095a824694cb74c78df4b0e65f703bcb9e43191c5f7cf2b8795d5"} Dec 06 06:12:27 crc kubenswrapper[4809]: I1206 06:12:27.685193 4809 generic.go:334] "Generic (PLEG): container finished" podID="99d85dd8-4137-4876-9062-3c23ee0bb730" containerID="e72b917948831ed516a45fc34d12815689dd3f124fb92cefccf19db85380c106" exitCode=0 Dec 06 06:12:27 crc kubenswrapper[4809]: I1206 06:12:27.685246 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/2a8c7bb1b949e5c76441d571adf78af7c2049b664ecf2ad024c54dd9ba67kbg" event={"ID":"99d85dd8-4137-4876-9062-3c23ee0bb730","Type":"ContainerDied","Data":"e72b917948831ed516a45fc34d12815689dd3f124fb92cefccf19db85380c106"} Dec 06 06:12:29 crc kubenswrapper[4809]: I1206 06:12:29.050321 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/2a8c7bb1b949e5c76441d571adf78af7c2049b664ecf2ad024c54dd9ba67kbg" Dec 06 06:12:29 crc kubenswrapper[4809]: I1206 06:12:29.084741 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n75rb\" (UniqueName: \"kubernetes.io/projected/99d85dd8-4137-4876-9062-3c23ee0bb730-kube-api-access-n75rb\") pod \"99d85dd8-4137-4876-9062-3c23ee0bb730\" (UID: \"99d85dd8-4137-4876-9062-3c23ee0bb730\") " Dec 06 06:12:29 crc kubenswrapper[4809]: I1206 06:12:29.085076 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/99d85dd8-4137-4876-9062-3c23ee0bb730-bundle\") pod \"99d85dd8-4137-4876-9062-3c23ee0bb730\" (UID: \"99d85dd8-4137-4876-9062-3c23ee0bb730\") " Dec 06 06:12:29 crc kubenswrapper[4809]: I1206 06:12:29.085106 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/99d85dd8-4137-4876-9062-3c23ee0bb730-util\") pod \"99d85dd8-4137-4876-9062-3c23ee0bb730\" (UID: \"99d85dd8-4137-4876-9062-3c23ee0bb730\") " Dec 06 06:12:29 crc kubenswrapper[4809]: I1206 06:12:29.085773 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99d85dd8-4137-4876-9062-3c23ee0bb730-bundle" (OuterVolumeSpecName: "bundle") pod "99d85dd8-4137-4876-9062-3c23ee0bb730" (UID: "99d85dd8-4137-4876-9062-3c23ee0bb730"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:12:29 crc kubenswrapper[4809]: I1206 06:12:29.092314 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99d85dd8-4137-4876-9062-3c23ee0bb730-kube-api-access-n75rb" (OuterVolumeSpecName: "kube-api-access-n75rb") pod "99d85dd8-4137-4876-9062-3c23ee0bb730" (UID: "99d85dd8-4137-4876-9062-3c23ee0bb730"). InnerVolumeSpecName "kube-api-access-n75rb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:12:29 crc kubenswrapper[4809]: I1206 06:12:29.099751 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99d85dd8-4137-4876-9062-3c23ee0bb730-util" (OuterVolumeSpecName: "util") pod "99d85dd8-4137-4876-9062-3c23ee0bb730" (UID: "99d85dd8-4137-4876-9062-3c23ee0bb730"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:12:29 crc kubenswrapper[4809]: I1206 06:12:29.187308 4809 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/99d85dd8-4137-4876-9062-3c23ee0bb730-util\") on node \"crc\" DevicePath \"\"" Dec 06 06:12:29 crc kubenswrapper[4809]: I1206 06:12:29.187354 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n75rb\" (UniqueName: \"kubernetes.io/projected/99d85dd8-4137-4876-9062-3c23ee0bb730-kube-api-access-n75rb\") on node \"crc\" DevicePath \"\"" Dec 06 06:12:29 crc kubenswrapper[4809]: I1206 06:12:29.187366 4809 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/99d85dd8-4137-4876-9062-3c23ee0bb730-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:12:29 crc kubenswrapper[4809]: I1206 06:12:29.709161 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/2a8c7bb1b949e5c76441d571adf78af7c2049b664ecf2ad024c54dd9ba67kbg" event={"ID":"99d85dd8-4137-4876-9062-3c23ee0bb730","Type":"ContainerDied","Data":"476dca1fce4ee7991749b26309960ab17e95faf98a81ab202851ef6ec33d2adc"} Dec 06 06:12:29 crc kubenswrapper[4809]: I1206 06:12:29.709211 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="476dca1fce4ee7991749b26309960ab17e95faf98a81ab202851ef6ec33d2adc" Dec 06 06:12:29 crc kubenswrapper[4809]: I1206 06:12:29.709290 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/2a8c7bb1b949e5c76441d571adf78af7c2049b664ecf2ad024c54dd9ba67kbg" Dec 06 06:12:34 crc kubenswrapper[4809]: I1206 06:12:34.328482 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6598f77bc9-xjfmg"] Dec 06 06:12:34 crc kubenswrapper[4809]: E1206 06:12:34.329167 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99d85dd8-4137-4876-9062-3c23ee0bb730" containerName="extract" Dec 06 06:12:34 crc kubenswrapper[4809]: I1206 06:12:34.329181 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="99d85dd8-4137-4876-9062-3c23ee0bb730" containerName="extract" Dec 06 06:12:34 crc kubenswrapper[4809]: E1206 06:12:34.329223 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99d85dd8-4137-4876-9062-3c23ee0bb730" containerName="util" Dec 06 06:12:34 crc kubenswrapper[4809]: I1206 06:12:34.329229 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="99d85dd8-4137-4876-9062-3c23ee0bb730" containerName="util" Dec 06 06:12:34 crc kubenswrapper[4809]: E1206 06:12:34.329251 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99d85dd8-4137-4876-9062-3c23ee0bb730" containerName="pull" Dec 06 06:12:34 crc kubenswrapper[4809]: I1206 06:12:34.329257 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="99d85dd8-4137-4876-9062-3c23ee0bb730" containerName="pull" Dec 06 06:12:34 crc kubenswrapper[4809]: I1206 06:12:34.329421 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="99d85dd8-4137-4876-9062-3c23ee0bb730" containerName="extract" Dec 06 06:12:34 crc kubenswrapper[4809]: I1206 06:12:34.330122 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6598f77bc9-xjfmg" Dec 06 06:12:34 crc kubenswrapper[4809]: I1206 06:12:34.332074 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-9ssph" Dec 06 06:12:34 crc kubenswrapper[4809]: I1206 06:12:34.365397 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6598f77bc9-xjfmg"] Dec 06 06:12:34 crc kubenswrapper[4809]: I1206 06:12:34.485446 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9cxk\" (UniqueName: \"kubernetes.io/projected/2d9f04dc-3365-4b3e-b57c-604e657c6483-kube-api-access-h9cxk\") pod \"openstack-operator-controller-operator-6598f77bc9-xjfmg\" (UID: \"2d9f04dc-3365-4b3e-b57c-604e657c6483\") " pod="openstack-operators/openstack-operator-controller-operator-6598f77bc9-xjfmg" Dec 06 06:12:34 crc kubenswrapper[4809]: I1206 06:12:34.587391 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9cxk\" (UniqueName: \"kubernetes.io/projected/2d9f04dc-3365-4b3e-b57c-604e657c6483-kube-api-access-h9cxk\") pod \"openstack-operator-controller-operator-6598f77bc9-xjfmg\" (UID: \"2d9f04dc-3365-4b3e-b57c-604e657c6483\") " pod="openstack-operators/openstack-operator-controller-operator-6598f77bc9-xjfmg" Dec 06 06:12:34 crc kubenswrapper[4809]: I1206 06:12:34.608000 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9cxk\" (UniqueName: \"kubernetes.io/projected/2d9f04dc-3365-4b3e-b57c-604e657c6483-kube-api-access-h9cxk\") pod \"openstack-operator-controller-operator-6598f77bc9-xjfmg\" (UID: \"2d9f04dc-3365-4b3e-b57c-604e657c6483\") " pod="openstack-operators/openstack-operator-controller-operator-6598f77bc9-xjfmg" Dec 06 06:12:34 crc kubenswrapper[4809]: I1206 06:12:34.648727 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6598f77bc9-xjfmg" Dec 06 06:12:35 crc kubenswrapper[4809]: I1206 06:12:35.177842 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6598f77bc9-xjfmg"] Dec 06 06:12:35 crc kubenswrapper[4809]: I1206 06:12:35.762839 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6598f77bc9-xjfmg" event={"ID":"2d9f04dc-3365-4b3e-b57c-604e657c6483","Type":"ContainerStarted","Data":"6ba3c52da622f1618acdb75363e41f07f1fa5eae12162ec522a64c6ddff3bfd5"} Dec 06 06:12:40 crc kubenswrapper[4809]: I1206 06:12:40.848828 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6598f77bc9-xjfmg" event={"ID":"2d9f04dc-3365-4b3e-b57c-604e657c6483","Type":"ContainerStarted","Data":"5aae50b8c9f015a57d5e306c520766ad33ad1ed6ef7eec8a12710d9aff47c115"} Dec 06 06:12:40 crc kubenswrapper[4809]: I1206 06:12:40.849549 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-6598f77bc9-xjfmg" Dec 06 06:12:40 crc kubenswrapper[4809]: I1206 06:12:40.880839 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-6598f77bc9-xjfmg" podStartSLOduration=2.291222776 podStartE2EDuration="6.88080807s" podCreationTimestamp="2025-12-06 06:12:34 +0000 UTC" firstStartedPulling="2025-12-06 06:12:35.185275119 +0000 UTC m=+1280.074258061" lastFinishedPulling="2025-12-06 06:12:39.774860413 +0000 UTC m=+1284.663843355" observedRunningTime="2025-12-06 06:12:40.874655545 +0000 UTC m=+1285.763638507" watchObservedRunningTime="2025-12-06 06:12:40.88080807 +0000 UTC m=+1285.769791012" Dec 06 06:12:44 crc kubenswrapper[4809]: I1206 06:12:44.653298 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-6598f77bc9-xjfmg" Dec 06 06:13:03 crc kubenswrapper[4809]: I1206 06:13:03.836702 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-l8clc"] Dec 06 06:13:03 crc kubenswrapper[4809]: I1206 06:13:03.839987 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-l8clc" Dec 06 06:13:03 crc kubenswrapper[4809]: I1206 06:13:03.843196 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-zkhr2" Dec 06 06:13:03 crc kubenswrapper[4809]: I1206 06:13:03.858273 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-l8clc"] Dec 06 06:13:03 crc kubenswrapper[4809]: I1206 06:13:03.866868 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-j4c62"] Dec 06 06:13:03 crc kubenswrapper[4809]: I1206 06:13:03.868859 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-j4c62" Dec 06 06:13:03 crc kubenswrapper[4809]: I1206 06:13:03.872103 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-8bk4m" Dec 06 06:13:03 crc kubenswrapper[4809]: I1206 06:13:03.873749 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-j4c62"] Dec 06 06:13:03 crc kubenswrapper[4809]: I1206 06:13:03.901195 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-h69gw"] Dec 06 06:13:03 crc kubenswrapper[4809]: I1206 06:13:03.903062 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-h69gw" Dec 06 06:13:03 crc kubenswrapper[4809]: I1206 06:13:03.910161 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-dg6wm" Dec 06 06:13:03 crc kubenswrapper[4809]: I1206 06:13:03.919824 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-h69gw"] Dec 06 06:13:03 crc kubenswrapper[4809]: I1206 06:13:03.951517 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2mrz\" (UniqueName: \"kubernetes.io/projected/44c524db-76cf-49c3-8c6e-1ad181d553ae-kube-api-access-q2mrz\") pod \"barbican-operator-controller-manager-7d9dfd778-l8clc\" (UID: \"44c524db-76cf-49c3-8c6e-1ad181d553ae\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-l8clc" Dec 06 06:13:03 crc kubenswrapper[4809]: I1206 06:13:03.956675 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-szvjh"] Dec 06 06:13:03 crc kubenswrapper[4809]: I1206 06:13:03.958720 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-szvjh" Dec 06 06:13:03 crc kubenswrapper[4809]: I1206 06:13:03.964061 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-p4q5k" Dec 06 06:13:03 crc kubenswrapper[4809]: I1206 06:13:03.993150 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4s24n"] Dec 06 06:13:03 crc kubenswrapper[4809]: I1206 06:13:03.995578 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4s24n" Dec 06 06:13:03 crc kubenswrapper[4809]: I1206 06:13:03.998584 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-l4q6s" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.009134 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-szvjh"] Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.026274 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-9gfg2"] Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.041477 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-9gfg2" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.048871 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-gpt5t" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.051636 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4s24n"] Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.054226 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtpln\" (UniqueName: \"kubernetes.io/projected/8be493fb-5378-467d-9fde-d5154cda3207-kube-api-access-xtpln\") pod \"designate-operator-controller-manager-78b4bc895b-h69gw\" (UID: \"8be493fb-5378-467d-9fde-d5154cda3207\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-h69gw" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.054286 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2mrz\" (UniqueName: \"kubernetes.io/projected/44c524db-76cf-49c3-8c6e-1ad181d553ae-kube-api-access-q2mrz\") pod \"barbican-operator-controller-manager-7d9dfd778-l8clc\" (UID: \"44c524db-76cf-49c3-8c6e-1ad181d553ae\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-l8clc" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.054310 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqllb\" (UniqueName: \"kubernetes.io/projected/7233cfab-96cb-4459-a507-3031cac37720-kube-api-access-gqllb\") pod \"cinder-operator-controller-manager-859b6ccc6-j4c62\" (UID: \"7233cfab-96cb-4459-a507-3031cac37720\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-j4c62" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.129954 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2mrz\" (UniqueName: \"kubernetes.io/projected/44c524db-76cf-49c3-8c6e-1ad181d553ae-kube-api-access-q2mrz\") pod \"barbican-operator-controller-manager-7d9dfd778-l8clc\" (UID: \"44c524db-76cf-49c3-8c6e-1ad181d553ae\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-l8clc" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.143485 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-9gfg2"] Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.155790 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bf52\" (UniqueName: \"kubernetes.io/projected/204981d9-37b3-4584-b4bc-57df854143cc-kube-api-access-7bf52\") pod \"glance-operator-controller-manager-77987cd8cd-szvjh\" (UID: \"204981d9-37b3-4584-b4bc-57df854143cc\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-szvjh" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.155867 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6g9g\" (UniqueName: \"kubernetes.io/projected/ee834860-9609-482d-95db-82fa1a1a6941-kube-api-access-m6g9g\") pod \"horizon-operator-controller-manager-68c6d99b8f-9gfg2\" (UID: \"ee834860-9609-482d-95db-82fa1a1a6941\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-9gfg2" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.155912 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6czls\" (UniqueName: \"kubernetes.io/projected/35b44606-4fdd-4662-9550-adc5ed2478b3-kube-api-access-6czls\") pod \"heat-operator-controller-manager-5f64f6f8bb-4s24n\" (UID: \"35b44606-4fdd-4662-9550-adc5ed2478b3\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4s24n" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.156006 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtpln\" (UniqueName: \"kubernetes.io/projected/8be493fb-5378-467d-9fde-d5154cda3207-kube-api-access-xtpln\") pod \"designate-operator-controller-manager-78b4bc895b-h69gw\" (UID: \"8be493fb-5378-467d-9fde-d5154cda3207\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-h69gw" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.156046 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqllb\" (UniqueName: \"kubernetes.io/projected/7233cfab-96cb-4459-a507-3031cac37720-kube-api-access-gqllb\") pod \"cinder-operator-controller-manager-859b6ccc6-j4c62\" (UID: \"7233cfab-96cb-4459-a507-3031cac37720\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-j4c62" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.169118 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-7trbt"] Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.171602 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7trbt" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.177697 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-tbzl2" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.178627 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-4b9l8"] Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.180775 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-4b9l8" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.188874 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-gzv6k"] Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.189897 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-pc84p" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.190125 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.190833 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-gzv6k" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.220780 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-l8clc" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.232035 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-d8kxw" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.237858 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtpln\" (UniqueName: \"kubernetes.io/projected/8be493fb-5378-467d-9fde-d5154cda3207-kube-api-access-xtpln\") pod \"designate-operator-controller-manager-78b4bc895b-h69gw\" (UID: \"8be493fb-5378-467d-9fde-d5154cda3207\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-h69gw" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.247754 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqllb\" (UniqueName: \"kubernetes.io/projected/7233cfab-96cb-4459-a507-3031cac37720-kube-api-access-gqllb\") pod \"cinder-operator-controller-manager-859b6ccc6-j4c62\" (UID: \"7233cfab-96cb-4459-a507-3031cac37720\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-j4c62" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.250557 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-7trbt"] Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.254394 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-h69gw" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.266826 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgkrh\" (UniqueName: \"kubernetes.io/projected/29c6d2cb-27fc-4b79-969a-462be41663e3-kube-api-access-hgkrh\") pod \"infra-operator-controller-manager-57548d458d-4b9l8\" (UID: \"29c6d2cb-27fc-4b79-969a-462be41663e3\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-4b9l8" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.272325 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdksd\" (UniqueName: \"kubernetes.io/projected/45f87079-e852-4b47-a240-6f2bcf247509-kube-api-access-sdksd\") pod \"ironic-operator-controller-manager-6c548fd776-7trbt\" (UID: \"45f87079-e852-4b47-a240-6f2bcf247509\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7trbt" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.272424 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/29c6d2cb-27fc-4b79-969a-462be41663e3-cert\") pod \"infra-operator-controller-manager-57548d458d-4b9l8\" (UID: \"29c6d2cb-27fc-4b79-969a-462be41663e3\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-4b9l8" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.272457 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4txtd\" (UniqueName: \"kubernetes.io/projected/ced7591a-f80d-46a0-b8ea-024c2e37ae9f-kube-api-access-4txtd\") pod \"keystone-operator-controller-manager-7765d96ddf-gzv6k\" (UID: \"ced7591a-f80d-46a0-b8ea-024c2e37ae9f\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-gzv6k" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.272512 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bf52\" (UniqueName: \"kubernetes.io/projected/204981d9-37b3-4584-b4bc-57df854143cc-kube-api-access-7bf52\") pod \"glance-operator-controller-manager-77987cd8cd-szvjh\" (UID: \"204981d9-37b3-4584-b4bc-57df854143cc\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-szvjh" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.294978 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6g9g\" (UniqueName: \"kubernetes.io/projected/ee834860-9609-482d-95db-82fa1a1a6941-kube-api-access-m6g9g\") pod \"horizon-operator-controller-manager-68c6d99b8f-9gfg2\" (UID: \"ee834860-9609-482d-95db-82fa1a1a6941\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-9gfg2" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.295497 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6czls\" (UniqueName: \"kubernetes.io/projected/35b44606-4fdd-4662-9550-adc5ed2478b3-kube-api-access-6czls\") pod \"heat-operator-controller-manager-5f64f6f8bb-4s24n\" (UID: \"35b44606-4fdd-4662-9550-adc5ed2478b3\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4s24n" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.306724 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-vh8d4"] Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.319304 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-vh8d4" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.332170 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-v896w" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.335080 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6czls\" (UniqueName: \"kubernetes.io/projected/35b44606-4fdd-4662-9550-adc5ed2478b3-kube-api-access-6czls\") pod \"heat-operator-controller-manager-5f64f6f8bb-4s24n\" (UID: \"35b44606-4fdd-4662-9550-adc5ed2478b3\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4s24n" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.342486 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bf52\" (UniqueName: \"kubernetes.io/projected/204981d9-37b3-4584-b4bc-57df854143cc-kube-api-access-7bf52\") pod \"glance-operator-controller-manager-77987cd8cd-szvjh\" (UID: \"204981d9-37b3-4584-b4bc-57df854143cc\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-szvjh" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.352270 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-gzv6k"] Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.357215 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6g9g\" (UniqueName: \"kubernetes.io/projected/ee834860-9609-482d-95db-82fa1a1a6941-kube-api-access-m6g9g\") pod \"horizon-operator-controller-manager-68c6d99b8f-9gfg2\" (UID: \"ee834860-9609-482d-95db-82fa1a1a6941\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-9gfg2" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.370059 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-9gfg2" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.377613 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-vh8d4"] Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.390057 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-4b9l8"] Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.400124 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgkrh\" (UniqueName: \"kubernetes.io/projected/29c6d2cb-27fc-4b79-969a-462be41663e3-kube-api-access-hgkrh\") pod \"infra-operator-controller-manager-57548d458d-4b9l8\" (UID: \"29c6d2cb-27fc-4b79-969a-462be41663e3\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-4b9l8" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.400187 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gm7l\" (UniqueName: \"kubernetes.io/projected/93397035-a1a7-4c92-bdeb-5d0091dea181-kube-api-access-5gm7l\") pod \"manila-operator-controller-manager-7c79b5df47-vh8d4\" (UID: \"93397035-a1a7-4c92-bdeb-5d0091dea181\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-vh8d4" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.400246 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdksd\" (UniqueName: \"kubernetes.io/projected/45f87079-e852-4b47-a240-6f2bcf247509-kube-api-access-sdksd\") pod \"ironic-operator-controller-manager-6c548fd776-7trbt\" (UID: \"45f87079-e852-4b47-a240-6f2bcf247509\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7trbt" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.400272 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/29c6d2cb-27fc-4b79-969a-462be41663e3-cert\") pod \"infra-operator-controller-manager-57548d458d-4b9l8\" (UID: \"29c6d2cb-27fc-4b79-969a-462be41663e3\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-4b9l8" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.400292 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4txtd\" (UniqueName: \"kubernetes.io/projected/ced7591a-f80d-46a0-b8ea-024c2e37ae9f-kube-api-access-4txtd\") pod \"keystone-operator-controller-manager-7765d96ddf-gzv6k\" (UID: \"ced7591a-f80d-46a0-b8ea-024c2e37ae9f\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-gzv6k" Dec 06 06:13:04 crc kubenswrapper[4809]: E1206 06:13:04.400740 4809 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 06 06:13:04 crc kubenswrapper[4809]: E1206 06:13:04.400860 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/29c6d2cb-27fc-4b79-969a-462be41663e3-cert podName:29c6d2cb-27fc-4b79-969a-462be41663e3 nodeName:}" failed. No retries permitted until 2025-12-06 06:13:04.90083907 +0000 UTC m=+1309.789822012 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/29c6d2cb-27fc-4b79-969a-462be41663e3-cert") pod "infra-operator-controller-manager-57548d458d-4b9l8" (UID: "29c6d2cb-27fc-4b79-969a-462be41663e3") : secret "infra-operator-webhook-server-cert" not found Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.404983 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-kcp8k"] Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.406872 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-kcp8k" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.409376 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-jhqhx" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.420105 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-sxcr7"] Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.421732 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-sxcr7" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.423496 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-kcp8k"] Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.425544 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgkrh\" (UniqueName: \"kubernetes.io/projected/29c6d2cb-27fc-4b79-969a-462be41663e3-kube-api-access-hgkrh\") pod \"infra-operator-controller-manager-57548d458d-4b9l8\" (UID: \"29c6d2cb-27fc-4b79-969a-462be41663e3\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-4b9l8" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.425872 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-r8tjv" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.445010 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-sxcr7"] Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.451809 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4txtd\" (UniqueName: \"kubernetes.io/projected/ced7591a-f80d-46a0-b8ea-024c2e37ae9f-kube-api-access-4txtd\") pod \"keystone-operator-controller-manager-7765d96ddf-gzv6k\" (UID: \"ced7591a-f80d-46a0-b8ea-024c2e37ae9f\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-gzv6k" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.461005 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-27fdn"] Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.463318 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-27fdn" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.467226 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdksd\" (UniqueName: \"kubernetes.io/projected/45f87079-e852-4b47-a240-6f2bcf247509-kube-api-access-sdksd\") pod \"ironic-operator-controller-manager-6c548fd776-7trbt\" (UID: \"45f87079-e852-4b47-a240-6f2bcf247509\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7trbt" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.471767 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-blzwz" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.475407 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-27fdn"] Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.522059 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-dfs6c"] Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.528547 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddbp9\" (UniqueName: \"kubernetes.io/projected/f736a8a4-ff7e-4179-951f-66405b1fd933-kube-api-access-ddbp9\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-kcp8k\" (UID: \"f736a8a4-ff7e-4179-951f-66405b1fd933\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-kcp8k" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.528621 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qxhk\" (UniqueName: \"kubernetes.io/projected/c37c165f-5cd1-496f-b092-41efdfcd7eca-kube-api-access-7qxhk\") pod \"mariadb-operator-controller-manager-56bbcc9d85-sxcr7\" (UID: \"c37c165f-5cd1-496f-b092-41efdfcd7eca\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-sxcr7" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.528660 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gm7l\" (UniqueName: \"kubernetes.io/projected/93397035-a1a7-4c92-bdeb-5d0091dea181-kube-api-access-5gm7l\") pod \"manila-operator-controller-manager-7c79b5df47-vh8d4\" (UID: \"93397035-a1a7-4c92-bdeb-5d0091dea181\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-vh8d4" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.528689 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2892s\" (UniqueName: \"kubernetes.io/projected/e6e0283a-55e9-4613-b74d-81fbb840690e-kube-api-access-2892s\") pod \"nova-operator-controller-manager-697bc559fc-27fdn\" (UID: \"e6e0283a-55e9-4613-b74d-81fbb840690e\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-27fdn" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.537581 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-dfs6c"] Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.537658 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd462lbf"] Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.538018 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-dfs6c" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.542857 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-j4c62" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.544469 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-vfvwz" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.545240 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd462lbf" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.552382 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.552473 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-8fdp6" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.569546 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gm7l\" (UniqueName: \"kubernetes.io/projected/93397035-a1a7-4c92-bdeb-5d0091dea181-kube-api-access-5gm7l\") pod \"manila-operator-controller-manager-7c79b5df47-vh8d4\" (UID: \"93397035-a1a7-4c92-bdeb-5d0091dea181\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-vh8d4" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.580654 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-szvjh" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.615712 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-ghwtq"] Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.624004 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4s24n" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.628066 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7trbt" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.630076 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddbp9\" (UniqueName: \"kubernetes.io/projected/f736a8a4-ff7e-4179-951f-66405b1fd933-kube-api-access-ddbp9\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-kcp8k\" (UID: \"f736a8a4-ff7e-4179-951f-66405b1fd933\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-kcp8k" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.634319 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qxhk\" (UniqueName: \"kubernetes.io/projected/c37c165f-5cd1-496f-b092-41efdfcd7eca-kube-api-access-7qxhk\") pod \"mariadb-operator-controller-manager-56bbcc9d85-sxcr7\" (UID: \"c37c165f-5cd1-496f-b092-41efdfcd7eca\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-sxcr7" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.634571 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jlwg\" (UniqueName: \"kubernetes.io/projected/2ac245f3-8f64-4b6f-a51f-09e4b5811412-kube-api-access-6jlwg\") pod \"octavia-operator-controller-manager-998648c74-dfs6c\" (UID: \"2ac245f3-8f64-4b6f-a51f-09e4b5811412\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-dfs6c" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.634718 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2892s\" (UniqueName: \"kubernetes.io/projected/e6e0283a-55e9-4613-b74d-81fbb840690e-kube-api-access-2892s\") pod \"nova-operator-controller-manager-697bc559fc-27fdn\" (UID: \"e6e0283a-55e9-4613-b74d-81fbb840690e\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-27fdn" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.635066 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6fbac224-e82d-482e-aeea-c83a9f978775-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd462lbf\" (UID: \"6fbac224-e82d-482e-aeea-c83a9f978775\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd462lbf" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.635177 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpfcv\" (UniqueName: \"kubernetes.io/projected/6fbac224-e82d-482e-aeea-c83a9f978775-kube-api-access-fpfcv\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd462lbf\" (UID: \"6fbac224-e82d-482e-aeea-c83a9f978775\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd462lbf" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.653788 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-ghwtq"] Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.653997 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-ghwtq" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.658190 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-mg9mm" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.663505 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qxhk\" (UniqueName: \"kubernetes.io/projected/c37c165f-5cd1-496f-b092-41efdfcd7eca-kube-api-access-7qxhk\") pod \"mariadb-operator-controller-manager-56bbcc9d85-sxcr7\" (UID: \"c37c165f-5cd1-496f-b092-41efdfcd7eca\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-sxcr7" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.664133 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2892s\" (UniqueName: \"kubernetes.io/projected/e6e0283a-55e9-4613-b74d-81fbb840690e-kube-api-access-2892s\") pod \"nova-operator-controller-manager-697bc559fc-27fdn\" (UID: \"e6e0283a-55e9-4613-b74d-81fbb840690e\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-27fdn" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.686838 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddbp9\" (UniqueName: \"kubernetes.io/projected/f736a8a4-ff7e-4179-951f-66405b1fd933-kube-api-access-ddbp9\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-kcp8k\" (UID: \"f736a8a4-ff7e-4179-951f-66405b1fd933\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-kcp8k" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.690051 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd462lbf"] Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.695538 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-gzv6k" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.736535 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-vqvvr"] Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.738602 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jlwg\" (UniqueName: \"kubernetes.io/projected/2ac245f3-8f64-4b6f-a51f-09e4b5811412-kube-api-access-6jlwg\") pod \"octavia-operator-controller-manager-998648c74-dfs6c\" (UID: \"2ac245f3-8f64-4b6f-a51f-09e4b5811412\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-dfs6c" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.738735 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfkgj\" (UniqueName: \"kubernetes.io/projected/b379307b-aafd-42da-b761-38a556a0d143-kube-api-access-vfkgj\") pod \"ovn-operator-controller-manager-b6456fdb6-ghwtq\" (UID: \"b379307b-aafd-42da-b761-38a556a0d143\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-ghwtq" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.738872 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6fbac224-e82d-482e-aeea-c83a9f978775-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd462lbf\" (UID: \"6fbac224-e82d-482e-aeea-c83a9f978775\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd462lbf" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.738918 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpfcv\" (UniqueName: \"kubernetes.io/projected/6fbac224-e82d-482e-aeea-c83a9f978775-kube-api-access-fpfcv\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd462lbf\" (UID: \"6fbac224-e82d-482e-aeea-c83a9f978775\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd462lbf" Dec 06 06:13:04 crc kubenswrapper[4809]: E1206 06:13:04.739565 4809 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 06:13:04 crc kubenswrapper[4809]: E1206 06:13:04.739633 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6fbac224-e82d-482e-aeea-c83a9f978775-cert podName:6fbac224-e82d-482e-aeea-c83a9f978775 nodeName:}" failed. No retries permitted until 2025-12-06 06:13:05.2396126 +0000 UTC m=+1310.128595732 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6fbac224-e82d-482e-aeea-c83a9f978775-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd462lbf" (UID: "6fbac224-e82d-482e-aeea-c83a9f978775") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.739791 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-vqvvr" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.741894 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-vzjzz" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.763395 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-zz2mm"] Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.764790 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-vh8d4" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.765703 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-zz2mm" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.768500 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jlwg\" (UniqueName: \"kubernetes.io/projected/2ac245f3-8f64-4b6f-a51f-09e4b5811412-kube-api-access-6jlwg\") pod \"octavia-operator-controller-manager-998648c74-dfs6c\" (UID: \"2ac245f3-8f64-4b6f-a51f-09e4b5811412\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-dfs6c" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.769713 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-vqvvr"] Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.769418 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-8qpmj" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.774780 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpfcv\" (UniqueName: \"kubernetes.io/projected/6fbac224-e82d-482e-aeea-c83a9f978775-kube-api-access-fpfcv\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd462lbf\" (UID: \"6fbac224-e82d-482e-aeea-c83a9f978775\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd462lbf" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.781556 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-kcp8k" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.786396 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-zz2mm"] Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.794233 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-b4644b845-zxwvr"] Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.796104 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-b4644b845-zxwvr" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.799313 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-ckmvr" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.801109 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-sxcr7" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.805706 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-b4644b845-zxwvr"] Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.831424 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-lqkvl"] Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.833156 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-lqkvl" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.836644 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-plcw4" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.843538 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfkgj\" (UniqueName: \"kubernetes.io/projected/b379307b-aafd-42da-b761-38a556a0d143-kube-api-access-vfkgj\") pod \"ovn-operator-controller-manager-b6456fdb6-ghwtq\" (UID: \"b379307b-aafd-42da-b761-38a556a0d143\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-ghwtq" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.843712 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r48nh\" (UniqueName: \"kubernetes.io/projected/803d54f1-fbac-4ec5-b486-efae1981b46d-kube-api-access-r48nh\") pod \"swift-operator-controller-manager-5f8c65bbfc-vqvvr\" (UID: \"803d54f1-fbac-4ec5-b486-efae1981b46d\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-vqvvr" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.843821 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2lnv\" (UniqueName: \"kubernetes.io/projected/b55361fa-8a7c-46ff-92f6-9dd72dbc0536-kube-api-access-b2lnv\") pod \"placement-operator-controller-manager-78f8948974-zz2mm\" (UID: \"b55361fa-8a7c-46ff-92f6-9dd72dbc0536\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-zz2mm" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.849605 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-lqkvl"] Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.867044 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-bhbcl"] Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.868783 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-27fdn" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.874547 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-bhbcl" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.899533 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-t78tc" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.924961 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfkgj\" (UniqueName: \"kubernetes.io/projected/b379307b-aafd-42da-b761-38a556a0d143-kube-api-access-vfkgj\") pod \"ovn-operator-controller-manager-b6456fdb6-ghwtq\" (UID: \"b379307b-aafd-42da-b761-38a556a0d143\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-ghwtq" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.938519 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-dfs6c" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.947179 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2lnv\" (UniqueName: \"kubernetes.io/projected/b55361fa-8a7c-46ff-92f6-9dd72dbc0536-kube-api-access-b2lnv\") pod \"placement-operator-controller-manager-78f8948974-zz2mm\" (UID: \"b55361fa-8a7c-46ff-92f6-9dd72dbc0536\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-zz2mm" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.947322 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6kvf\" (UniqueName: \"kubernetes.io/projected/fd7dd0a5-5e39-4812-b93f-a27c39284d33-kube-api-access-m6kvf\") pod \"watcher-operator-controller-manager-769dc69bc-bhbcl\" (UID: \"fd7dd0a5-5e39-4812-b93f-a27c39284d33\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-bhbcl" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.947401 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/29c6d2cb-27fc-4b79-969a-462be41663e3-cert\") pod \"infra-operator-controller-manager-57548d458d-4b9l8\" (UID: \"29c6d2cb-27fc-4b79-969a-462be41663e3\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-4b9l8" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.947465 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r48nh\" (UniqueName: \"kubernetes.io/projected/803d54f1-fbac-4ec5-b486-efae1981b46d-kube-api-access-r48nh\") pod \"swift-operator-controller-manager-5f8c65bbfc-vqvvr\" (UID: \"803d54f1-fbac-4ec5-b486-efae1981b46d\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-vqvvr" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.947512 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rg67\" (UniqueName: \"kubernetes.io/projected/b376f569-0116-41f4-8a57-b5e5769e6a10-kube-api-access-6rg67\") pod \"test-operator-controller-manager-5854674fcc-lqkvl\" (UID: \"b376f569-0116-41f4-8a57-b5e5769e6a10\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-lqkvl" Dec 06 06:13:04 crc kubenswrapper[4809]: I1206 06:13:04.947551 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4r4dh\" (UniqueName: \"kubernetes.io/projected/24aed9fd-d792-4bce-82a8-3ea4d70094b4-kube-api-access-4r4dh\") pod \"telemetry-operator-controller-manager-b4644b845-zxwvr\" (UID: \"24aed9fd-d792-4bce-82a8-3ea4d70094b4\") " pod="openstack-operators/telemetry-operator-controller-manager-b4644b845-zxwvr" Dec 06 06:13:04 crc kubenswrapper[4809]: E1206 06:13:04.948301 4809 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 06 06:13:04 crc kubenswrapper[4809]: E1206 06:13:04.948380 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/29c6d2cb-27fc-4b79-969a-462be41663e3-cert podName:29c6d2cb-27fc-4b79-969a-462be41663e3 nodeName:}" failed. No retries permitted until 2025-12-06 06:13:05.948347106 +0000 UTC m=+1310.837330048 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/29c6d2cb-27fc-4b79-969a-462be41663e3-cert") pod "infra-operator-controller-manager-57548d458d-4b9l8" (UID: "29c6d2cb-27fc-4b79-969a-462be41663e3") : secret "infra-operator-webhook-server-cert" not found Dec 06 06:13:05 crc kubenswrapper[4809]: I1206 06:13:05.036436 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-ghwtq" Dec 06 06:13:05 crc kubenswrapper[4809]: I1206 06:13:05.051652 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6kvf\" (UniqueName: \"kubernetes.io/projected/fd7dd0a5-5e39-4812-b93f-a27c39284d33-kube-api-access-m6kvf\") pod \"watcher-operator-controller-manager-769dc69bc-bhbcl\" (UID: \"fd7dd0a5-5e39-4812-b93f-a27c39284d33\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-bhbcl" Dec 06 06:13:05 crc kubenswrapper[4809]: I1206 06:13:05.051822 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rg67\" (UniqueName: \"kubernetes.io/projected/b376f569-0116-41f4-8a57-b5e5769e6a10-kube-api-access-6rg67\") pod \"test-operator-controller-manager-5854674fcc-lqkvl\" (UID: \"b376f569-0116-41f4-8a57-b5e5769e6a10\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-lqkvl" Dec 06 06:13:05 crc kubenswrapper[4809]: I1206 06:13:05.051867 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4r4dh\" (UniqueName: \"kubernetes.io/projected/24aed9fd-d792-4bce-82a8-3ea4d70094b4-kube-api-access-4r4dh\") pod \"telemetry-operator-controller-manager-b4644b845-zxwvr\" (UID: \"24aed9fd-d792-4bce-82a8-3ea4d70094b4\") " pod="openstack-operators/telemetry-operator-controller-manager-b4644b845-zxwvr" Dec 06 06:13:05 crc kubenswrapper[4809]: I1206 06:13:05.056814 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r48nh\" (UniqueName: \"kubernetes.io/projected/803d54f1-fbac-4ec5-b486-efae1981b46d-kube-api-access-r48nh\") pod \"swift-operator-controller-manager-5f8c65bbfc-vqvvr\" (UID: \"803d54f1-fbac-4ec5-b486-efae1981b46d\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-vqvvr" Dec 06 06:13:05 crc kubenswrapper[4809]: I1206 06:13:05.058726 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2lnv\" (UniqueName: \"kubernetes.io/projected/b55361fa-8a7c-46ff-92f6-9dd72dbc0536-kube-api-access-b2lnv\") pod \"placement-operator-controller-manager-78f8948974-zz2mm\" (UID: \"b55361fa-8a7c-46ff-92f6-9dd72dbc0536\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-zz2mm" Dec 06 06:13:05 crc kubenswrapper[4809]: I1206 06:13:05.062024 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-bhbcl"] Dec 06 06:13:05 crc kubenswrapper[4809]: I1206 06:13:05.095649 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-vqvvr" Dec 06 06:13:05 crc kubenswrapper[4809]: I1206 06:13:05.106461 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4r4dh\" (UniqueName: \"kubernetes.io/projected/24aed9fd-d792-4bce-82a8-3ea4d70094b4-kube-api-access-4r4dh\") pod \"telemetry-operator-controller-manager-b4644b845-zxwvr\" (UID: \"24aed9fd-d792-4bce-82a8-3ea4d70094b4\") " pod="openstack-operators/telemetry-operator-controller-manager-b4644b845-zxwvr" Dec 06 06:13:05 crc kubenswrapper[4809]: I1206 06:13:05.106889 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-zz2mm" Dec 06 06:13:05 crc kubenswrapper[4809]: I1206 06:13:05.118601 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rg67\" (UniqueName: \"kubernetes.io/projected/b376f569-0116-41f4-8a57-b5e5769e6a10-kube-api-access-6rg67\") pod \"test-operator-controller-manager-5854674fcc-lqkvl\" (UID: \"b376f569-0116-41f4-8a57-b5e5769e6a10\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-lqkvl" Dec 06 06:13:05 crc kubenswrapper[4809]: I1206 06:13:05.124452 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-b4644b845-zxwvr" Dec 06 06:13:05 crc kubenswrapper[4809]: I1206 06:13:05.125638 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6kvf\" (UniqueName: \"kubernetes.io/projected/fd7dd0a5-5e39-4812-b93f-a27c39284d33-kube-api-access-m6kvf\") pod \"watcher-operator-controller-manager-769dc69bc-bhbcl\" (UID: \"fd7dd0a5-5e39-4812-b93f-a27c39284d33\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-bhbcl" Dec 06 06:13:05 crc kubenswrapper[4809]: I1206 06:13:05.165501 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-lqkvl" Dec 06 06:13:05 crc kubenswrapper[4809]: I1206 06:13:05.270462 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6fbac224-e82d-482e-aeea-c83a9f978775-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd462lbf\" (UID: \"6fbac224-e82d-482e-aeea-c83a9f978775\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd462lbf" Dec 06 06:13:05 crc kubenswrapper[4809]: E1206 06:13:05.270655 4809 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 06:13:05 crc kubenswrapper[4809]: E1206 06:13:05.270753 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6fbac224-e82d-482e-aeea-c83a9f978775-cert podName:6fbac224-e82d-482e-aeea-c83a9f978775 nodeName:}" failed. No retries permitted until 2025-12-06 06:13:06.270726065 +0000 UTC m=+1311.159709007 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6fbac224-e82d-482e-aeea-c83a9f978775-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd462lbf" (UID: "6fbac224-e82d-482e-aeea-c83a9f978775") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 06:13:05 crc kubenswrapper[4809]: I1206 06:13:05.276219 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-bhbcl" Dec 06 06:13:05 crc kubenswrapper[4809]: I1206 06:13:05.281437 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5c4b598d95-mbx56"] Dec 06 06:13:05 crc kubenswrapper[4809]: I1206 06:13:05.282982 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-5c4b598d95-mbx56" Dec 06 06:13:05 crc kubenswrapper[4809]: I1206 06:13:05.298345 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 06 06:13:05 crc kubenswrapper[4809]: I1206 06:13:05.303007 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-kgrq2" Dec 06 06:13:05 crc kubenswrapper[4809]: I1206 06:13:05.303249 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 06 06:13:05 crc kubenswrapper[4809]: I1206 06:13:05.332709 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5c4b598d95-mbx56"] Dec 06 06:13:05 crc kubenswrapper[4809]: I1206 06:13:05.341257 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-848xh"] Dec 06 06:13:05 crc kubenswrapper[4809]: I1206 06:13:05.342891 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-848xh" Dec 06 06:13:05 crc kubenswrapper[4809]: I1206 06:13:05.346681 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-bgqnw" Dec 06 06:13:05 crc kubenswrapper[4809]: I1206 06:13:05.351154 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-848xh"] Dec 06 06:13:05 crc kubenswrapper[4809]: I1206 06:13:05.372534 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zx4c4\" (UniqueName: \"kubernetes.io/projected/4dd349b6-b792-451e-9cab-7c52a5768a3d-kube-api-access-zx4c4\") pod \"openstack-operator-controller-manager-5c4b598d95-mbx56\" (UID: \"4dd349b6-b792-451e-9cab-7c52a5768a3d\") " pod="openstack-operators/openstack-operator-controller-manager-5c4b598d95-mbx56" Dec 06 06:13:05 crc kubenswrapper[4809]: I1206 06:13:05.372792 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4dd349b6-b792-451e-9cab-7c52a5768a3d-metrics-certs\") pod \"openstack-operator-controller-manager-5c4b598d95-mbx56\" (UID: \"4dd349b6-b792-451e-9cab-7c52a5768a3d\") " pod="openstack-operators/openstack-operator-controller-manager-5c4b598d95-mbx56" Dec 06 06:13:05 crc kubenswrapper[4809]: I1206 06:13:05.372892 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4dd349b6-b792-451e-9cab-7c52a5768a3d-webhook-certs\") pod \"openstack-operator-controller-manager-5c4b598d95-mbx56\" (UID: \"4dd349b6-b792-451e-9cab-7c52a5768a3d\") " pod="openstack-operators/openstack-operator-controller-manager-5c4b598d95-mbx56" Dec 06 06:13:05 crc kubenswrapper[4809]: I1206 06:13:05.372989 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvj9h\" (UniqueName: \"kubernetes.io/projected/0acb7c70-8d76-4bef-9d66-cd1213802a85-kube-api-access-wvj9h\") pod \"rabbitmq-cluster-operator-manager-668c99d594-848xh\" (UID: \"0acb7c70-8d76-4bef-9d66-cd1213802a85\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-848xh" Dec 06 06:13:05 crc kubenswrapper[4809]: W1206 06:13:05.429820 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod44c524db_76cf_49c3_8c6e_1ad181d553ae.slice/crio-4c3ac4e7add5dcc826d204d6500f7da6f7f7caa13c477dcd61d9d28d0a9b54b6 WatchSource:0}: Error finding container 4c3ac4e7add5dcc826d204d6500f7da6f7f7caa13c477dcd61d9d28d0a9b54b6: Status 404 returned error can't find the container with id 4c3ac4e7add5dcc826d204d6500f7da6f7f7caa13c477dcd61d9d28d0a9b54b6 Dec 06 06:13:05 crc kubenswrapper[4809]: I1206 06:13:05.435893 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-h69gw"] Dec 06 06:13:05 crc kubenswrapper[4809]: I1206 06:13:05.456251 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-l8clc"] Dec 06 06:13:05 crc kubenswrapper[4809]: I1206 06:13:05.474313 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zx4c4\" (UniqueName: \"kubernetes.io/projected/4dd349b6-b792-451e-9cab-7c52a5768a3d-kube-api-access-zx4c4\") pod \"openstack-operator-controller-manager-5c4b598d95-mbx56\" (UID: \"4dd349b6-b792-451e-9cab-7c52a5768a3d\") " pod="openstack-operators/openstack-operator-controller-manager-5c4b598d95-mbx56" Dec 06 06:13:05 crc kubenswrapper[4809]: I1206 06:13:05.474467 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4dd349b6-b792-451e-9cab-7c52a5768a3d-metrics-certs\") pod \"openstack-operator-controller-manager-5c4b598d95-mbx56\" (UID: \"4dd349b6-b792-451e-9cab-7c52a5768a3d\") " pod="openstack-operators/openstack-operator-controller-manager-5c4b598d95-mbx56" Dec 06 06:13:05 crc kubenswrapper[4809]: I1206 06:13:05.474554 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4dd349b6-b792-451e-9cab-7c52a5768a3d-webhook-certs\") pod \"openstack-operator-controller-manager-5c4b598d95-mbx56\" (UID: \"4dd349b6-b792-451e-9cab-7c52a5768a3d\") " pod="openstack-operators/openstack-operator-controller-manager-5c4b598d95-mbx56" Dec 06 06:13:05 crc kubenswrapper[4809]: I1206 06:13:05.474613 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvj9h\" (UniqueName: \"kubernetes.io/projected/0acb7c70-8d76-4bef-9d66-cd1213802a85-kube-api-access-wvj9h\") pod \"rabbitmq-cluster-operator-manager-668c99d594-848xh\" (UID: \"0acb7c70-8d76-4bef-9d66-cd1213802a85\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-848xh" Dec 06 06:13:05 crc kubenswrapper[4809]: E1206 06:13:05.475138 4809 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 06 06:13:05 crc kubenswrapper[4809]: E1206 06:13:05.475207 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4dd349b6-b792-451e-9cab-7c52a5768a3d-metrics-certs podName:4dd349b6-b792-451e-9cab-7c52a5768a3d nodeName:}" failed. No retries permitted until 2025-12-06 06:13:05.975183515 +0000 UTC m=+1310.864166457 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4dd349b6-b792-451e-9cab-7c52a5768a3d-metrics-certs") pod "openstack-operator-controller-manager-5c4b598d95-mbx56" (UID: "4dd349b6-b792-451e-9cab-7c52a5768a3d") : secret "metrics-server-cert" not found Dec 06 06:13:05 crc kubenswrapper[4809]: E1206 06:13:05.475561 4809 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 06 06:13:05 crc kubenswrapper[4809]: E1206 06:13:05.475599 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4dd349b6-b792-451e-9cab-7c52a5768a3d-webhook-certs podName:4dd349b6-b792-451e-9cab-7c52a5768a3d nodeName:}" failed. No retries permitted until 2025-12-06 06:13:05.975579105 +0000 UTC m=+1310.864562047 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/4dd349b6-b792-451e-9cab-7c52a5768a3d-webhook-certs") pod "openstack-operator-controller-manager-5c4b598d95-mbx56" (UID: "4dd349b6-b792-451e-9cab-7c52a5768a3d") : secret "webhook-server-cert" not found Dec 06 06:13:05 crc kubenswrapper[4809]: I1206 06:13:05.511895 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zx4c4\" (UniqueName: \"kubernetes.io/projected/4dd349b6-b792-451e-9cab-7c52a5768a3d-kube-api-access-zx4c4\") pod \"openstack-operator-controller-manager-5c4b598d95-mbx56\" (UID: \"4dd349b6-b792-451e-9cab-7c52a5768a3d\") " pod="openstack-operators/openstack-operator-controller-manager-5c4b598d95-mbx56" Dec 06 06:13:05 crc kubenswrapper[4809]: I1206 06:13:05.512128 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvj9h\" (UniqueName: \"kubernetes.io/projected/0acb7c70-8d76-4bef-9d66-cd1213802a85-kube-api-access-wvj9h\") pod \"rabbitmq-cluster-operator-manager-668c99d594-848xh\" (UID: \"0acb7c70-8d76-4bef-9d66-cd1213802a85\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-848xh" Dec 06 06:13:05 crc kubenswrapper[4809]: I1206 06:13:05.621209 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-9gfg2"] Dec 06 06:13:05 crc kubenswrapper[4809]: W1206 06:13:05.661259 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podee834860_9609_482d_95db_82fa1a1a6941.slice/crio-270a39262e4ad743d8317ff71f7b6d369cc45a759502c152d998cdab670ed310 WatchSource:0}: Error finding container 270a39262e4ad743d8317ff71f7b6d369cc45a759502c152d998cdab670ed310: Status 404 returned error can't find the container with id 270a39262e4ad743d8317ff71f7b6d369cc45a759502c152d998cdab670ed310 Dec 06 06:13:05 crc kubenswrapper[4809]: I1206 06:13:05.722399 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-848xh" Dec 06 06:13:05 crc kubenswrapper[4809]: I1206 06:13:05.895431 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-j4c62"] Dec 06 06:13:05 crc kubenswrapper[4809]: W1206 06:13:05.909695 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7233cfab_96cb_4459_a507_3031cac37720.slice/crio-381da71acdc97d89a568e93a55d7a6ddb3e1ce524e0f11534f0f439dd329701f WatchSource:0}: Error finding container 381da71acdc97d89a568e93a55d7a6ddb3e1ce524e0f11534f0f439dd329701f: Status 404 returned error can't find the container with id 381da71acdc97d89a568e93a55d7a6ddb3e1ce524e0f11534f0f439dd329701f Dec 06 06:13:05 crc kubenswrapper[4809]: I1206 06:13:05.919035 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-szvjh"] Dec 06 06:13:05 crc kubenswrapper[4809]: W1206 06:13:05.921098 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod204981d9_37b3_4584_b4bc_57df854143cc.slice/crio-f2e7054cdda15aff0918d743aabf723145df6bca5164df8d6d81430231799911 WatchSource:0}: Error finding container f2e7054cdda15aff0918d743aabf723145df6bca5164df8d6d81430231799911: Status 404 returned error can't find the container with id f2e7054cdda15aff0918d743aabf723145df6bca5164df8d6d81430231799911 Dec 06 06:13:05 crc kubenswrapper[4809]: I1206 06:13:05.989694 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4dd349b6-b792-451e-9cab-7c52a5768a3d-metrics-certs\") pod \"openstack-operator-controller-manager-5c4b598d95-mbx56\" (UID: \"4dd349b6-b792-451e-9cab-7c52a5768a3d\") " pod="openstack-operators/openstack-operator-controller-manager-5c4b598d95-mbx56" Dec 06 06:13:05 crc kubenswrapper[4809]: I1206 06:13:05.989809 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4dd349b6-b792-451e-9cab-7c52a5768a3d-webhook-certs\") pod \"openstack-operator-controller-manager-5c4b598d95-mbx56\" (UID: \"4dd349b6-b792-451e-9cab-7c52a5768a3d\") " pod="openstack-operators/openstack-operator-controller-manager-5c4b598d95-mbx56" Dec 06 06:13:05 crc kubenswrapper[4809]: E1206 06:13:05.989912 4809 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 06 06:13:05 crc kubenswrapper[4809]: E1206 06:13:05.990029 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4dd349b6-b792-451e-9cab-7c52a5768a3d-metrics-certs podName:4dd349b6-b792-451e-9cab-7c52a5768a3d nodeName:}" failed. No retries permitted until 2025-12-06 06:13:06.990005049 +0000 UTC m=+1311.878987991 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4dd349b6-b792-451e-9cab-7c52a5768a3d-metrics-certs") pod "openstack-operator-controller-manager-5c4b598d95-mbx56" (UID: "4dd349b6-b792-451e-9cab-7c52a5768a3d") : secret "metrics-server-cert" not found Dec 06 06:13:05 crc kubenswrapper[4809]: I1206 06:13:05.990127 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/29c6d2cb-27fc-4b79-969a-462be41663e3-cert\") pod \"infra-operator-controller-manager-57548d458d-4b9l8\" (UID: \"29c6d2cb-27fc-4b79-969a-462be41663e3\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-4b9l8" Dec 06 06:13:05 crc kubenswrapper[4809]: E1206 06:13:05.990128 4809 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 06 06:13:05 crc kubenswrapper[4809]: E1206 06:13:05.990201 4809 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 06 06:13:05 crc kubenswrapper[4809]: E1206 06:13:05.990309 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4dd349b6-b792-451e-9cab-7c52a5768a3d-webhook-certs podName:4dd349b6-b792-451e-9cab-7c52a5768a3d nodeName:}" failed. No retries permitted until 2025-12-06 06:13:06.990286797 +0000 UTC m=+1311.879269739 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/4dd349b6-b792-451e-9cab-7c52a5768a3d-webhook-certs") pod "openstack-operator-controller-manager-5c4b598d95-mbx56" (UID: "4dd349b6-b792-451e-9cab-7c52a5768a3d") : secret "webhook-server-cert" not found Dec 06 06:13:05 crc kubenswrapper[4809]: E1206 06:13:05.990359 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/29c6d2cb-27fc-4b79-969a-462be41663e3-cert podName:29c6d2cb-27fc-4b79-969a-462be41663e3 nodeName:}" failed. No retries permitted until 2025-12-06 06:13:07.990334698 +0000 UTC m=+1312.879317640 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/29c6d2cb-27fc-4b79-969a-462be41663e3-cert") pod "infra-operator-controller-manager-57548d458d-4b9l8" (UID: "29c6d2cb-27fc-4b79-969a-462be41663e3") : secret "infra-operator-webhook-server-cert" not found Dec 06 06:13:06 crc kubenswrapper[4809]: I1206 06:13:06.229079 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-l8clc" event={"ID":"44c524db-76cf-49c3-8c6e-1ad181d553ae","Type":"ContainerStarted","Data":"4c3ac4e7add5dcc826d204d6500f7da6f7f7caa13c477dcd61d9d28d0a9b54b6"} Dec 06 06:13:06 crc kubenswrapper[4809]: I1206 06:13:06.230614 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-h69gw" event={"ID":"8be493fb-5378-467d-9fde-d5154cda3207","Type":"ContainerStarted","Data":"67c5ccd89f1cbd558614ccb18806f66b48726256e0523d527b1979cce61d4fc6"} Dec 06 06:13:06 crc kubenswrapper[4809]: I1206 06:13:06.231900 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-9gfg2" event={"ID":"ee834860-9609-482d-95db-82fa1a1a6941","Type":"ContainerStarted","Data":"270a39262e4ad743d8317ff71f7b6d369cc45a759502c152d998cdab670ed310"} Dec 06 06:13:06 crc kubenswrapper[4809]: I1206 06:13:06.233436 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-j4c62" event={"ID":"7233cfab-96cb-4459-a507-3031cac37720","Type":"ContainerStarted","Data":"381da71acdc97d89a568e93a55d7a6ddb3e1ce524e0f11534f0f439dd329701f"} Dec 06 06:13:06 crc kubenswrapper[4809]: I1206 06:13:06.235431 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-szvjh" event={"ID":"204981d9-37b3-4584-b4bc-57df854143cc","Type":"ContainerStarted","Data":"f2e7054cdda15aff0918d743aabf723145df6bca5164df8d6d81430231799911"} Dec 06 06:13:06 crc kubenswrapper[4809]: I1206 06:13:06.296134 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6fbac224-e82d-482e-aeea-c83a9f978775-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd462lbf\" (UID: \"6fbac224-e82d-482e-aeea-c83a9f978775\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd462lbf" Dec 06 06:13:06 crc kubenswrapper[4809]: E1206 06:13:06.296391 4809 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 06:13:06 crc kubenswrapper[4809]: E1206 06:13:06.296608 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6fbac224-e82d-482e-aeea-c83a9f978775-cert podName:6fbac224-e82d-482e-aeea-c83a9f978775 nodeName:}" failed. No retries permitted until 2025-12-06 06:13:08.296582662 +0000 UTC m=+1313.185565604 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6fbac224-e82d-482e-aeea-c83a9f978775-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd462lbf" (UID: "6fbac224-e82d-482e-aeea-c83a9f978775") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 06:13:06 crc kubenswrapper[4809]: W1206 06:13:06.415598 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod93397035_a1a7_4c92_bdeb_5d0091dea181.slice/crio-157d2599a51a49504b02c129021812b75e6900ee08551140fe2eb9683e7afefc WatchSource:0}: Error finding container 157d2599a51a49504b02c129021812b75e6900ee08551140fe2eb9683e7afefc: Status 404 returned error can't find the container with id 157d2599a51a49504b02c129021812b75e6900ee08551140fe2eb9683e7afefc Dec 06 06:13:06 crc kubenswrapper[4809]: I1206 06:13:06.415685 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-vh8d4"] Dec 06 06:13:06 crc kubenswrapper[4809]: W1206 06:13:06.425161 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podced7591a_f80d_46a0_b8ea_024c2e37ae9f.slice/crio-699006b6715df3d5101aaabc256ba8e2f27c2b7ef07b8e3d0175661d9bc052dd WatchSource:0}: Error finding container 699006b6715df3d5101aaabc256ba8e2f27c2b7ef07b8e3d0175661d9bc052dd: Status 404 returned error can't find the container with id 699006b6715df3d5101aaabc256ba8e2f27c2b7ef07b8e3d0175661d9bc052dd Dec 06 06:13:06 crc kubenswrapper[4809]: I1206 06:13:06.437025 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-gzv6k"] Dec 06 06:13:06 crc kubenswrapper[4809]: I1206 06:13:06.472635 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-7trbt"] Dec 06 06:13:06 crc kubenswrapper[4809]: W1206 06:13:06.481655 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod35b44606_4fdd_4662_9550_adc5ed2478b3.slice/crio-139d70514675315bab247ca1de6dc0267d0dc89fc328345ca4d78d0a431b75a2 WatchSource:0}: Error finding container 139d70514675315bab247ca1de6dc0267d0dc89fc328345ca4d78d0a431b75a2: Status 404 returned error can't find the container with id 139d70514675315bab247ca1de6dc0267d0dc89fc328345ca4d78d0a431b75a2 Dec 06 06:13:06 crc kubenswrapper[4809]: W1206 06:13:06.483662 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc37c165f_5cd1_496f_b092_41efdfcd7eca.slice/crio-b39b5e67028083755345f9f33d9c9d3d5e083ba825d80f2851d0454af1cdbad0 WatchSource:0}: Error finding container b39b5e67028083755345f9f33d9c9d3d5e083ba825d80f2851d0454af1cdbad0: Status 404 returned error can't find the container with id b39b5e67028083755345f9f33d9c9d3d5e083ba825d80f2851d0454af1cdbad0 Dec 06 06:13:06 crc kubenswrapper[4809]: I1206 06:13:06.484059 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4s24n"] Dec 06 06:13:06 crc kubenswrapper[4809]: W1206 06:13:06.484800 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod45f87079_e852_4b47_a240_6f2bcf247509.slice/crio-00c95316606600d4a3e98fa371ee36940fb93932dfed2c62c3ce00ff88df381e WatchSource:0}: Error finding container 00c95316606600d4a3e98fa371ee36940fb93932dfed2c62c3ce00ff88df381e: Status 404 returned error can't find the container with id 00c95316606600d4a3e98fa371ee36940fb93932dfed2c62c3ce00ff88df381e Dec 06 06:13:06 crc kubenswrapper[4809]: I1206 06:13:06.499685 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-sxcr7"] Dec 06 06:13:06 crc kubenswrapper[4809]: I1206 06:13:06.509842 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-kcp8k"] Dec 06 06:13:07 crc kubenswrapper[4809]: I1206 06:13:07.016084 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-b4644b845-zxwvr"] Dec 06 06:13:07 crc kubenswrapper[4809]: I1206 06:13:07.016554 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4dd349b6-b792-451e-9cab-7c52a5768a3d-metrics-certs\") pod \"openstack-operator-controller-manager-5c4b598d95-mbx56\" (UID: \"4dd349b6-b792-451e-9cab-7c52a5768a3d\") " pod="openstack-operators/openstack-operator-controller-manager-5c4b598d95-mbx56" Dec 06 06:13:07 crc kubenswrapper[4809]: I1206 06:13:07.016592 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4dd349b6-b792-451e-9cab-7c52a5768a3d-webhook-certs\") pod \"openstack-operator-controller-manager-5c4b598d95-mbx56\" (UID: \"4dd349b6-b792-451e-9cab-7c52a5768a3d\") " pod="openstack-operators/openstack-operator-controller-manager-5c4b598d95-mbx56" Dec 06 06:13:07 crc kubenswrapper[4809]: E1206 06:13:07.016783 4809 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 06 06:13:07 crc kubenswrapper[4809]: E1206 06:13:07.016843 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4dd349b6-b792-451e-9cab-7c52a5768a3d-webhook-certs podName:4dd349b6-b792-451e-9cab-7c52a5768a3d nodeName:}" failed. No retries permitted until 2025-12-06 06:13:09.016824924 +0000 UTC m=+1313.905807866 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/4dd349b6-b792-451e-9cab-7c52a5768a3d-webhook-certs") pod "openstack-operator-controller-manager-5c4b598d95-mbx56" (UID: "4dd349b6-b792-451e-9cab-7c52a5768a3d") : secret "webhook-server-cert" not found Dec 06 06:13:07 crc kubenswrapper[4809]: E1206 06:13:07.018275 4809 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 06 06:13:07 crc kubenswrapper[4809]: E1206 06:13:07.018307 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4dd349b6-b792-451e-9cab-7c52a5768a3d-metrics-certs podName:4dd349b6-b792-451e-9cab-7c52a5768a3d nodeName:}" failed. No retries permitted until 2025-12-06 06:13:09.018298684 +0000 UTC m=+1313.907281626 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4dd349b6-b792-451e-9cab-7c52a5768a3d-metrics-certs") pod "openstack-operator-controller-manager-5c4b598d95-mbx56" (UID: "4dd349b6-b792-451e-9cab-7c52a5768a3d") : secret "metrics-server-cert" not found Dec 06 06:13:07 crc kubenswrapper[4809]: I1206 06:13:07.036802 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-bhbcl"] Dec 06 06:13:07 crc kubenswrapper[4809]: W1206 06:13:07.044675 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfd7dd0a5_5e39_4812_b93f_a27c39284d33.slice/crio-9971fb5e9ba957570a03facec758e4803129523eff3262691f5bf9e007d63de6 WatchSource:0}: Error finding container 9971fb5e9ba957570a03facec758e4803129523eff3262691f5bf9e007d63de6: Status 404 returned error can't find the container with id 9971fb5e9ba957570a03facec758e4803129523eff3262691f5bf9e007d63de6 Dec 06 06:13:07 crc kubenswrapper[4809]: I1206 06:13:07.049966 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-dfs6c"] Dec 06 06:13:07 crc kubenswrapper[4809]: W1206 06:13:07.059140 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod803d54f1_fbac_4ec5_b486_efae1981b46d.slice/crio-9b84df5b9e7f23e9ea0b0fa07f5f078e1cd3db37f85619eee5987869bc890647 WatchSource:0}: Error finding container 9b84df5b9e7f23e9ea0b0fa07f5f078e1cd3db37f85619eee5987869bc890647: Status 404 returned error can't find the container with id 9b84df5b9e7f23e9ea0b0fa07f5f078e1cd3db37f85619eee5987869bc890647 Dec 06 06:13:07 crc kubenswrapper[4809]: I1206 06:13:07.064566 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-zz2mm"] Dec 06 06:13:07 crc kubenswrapper[4809]: I1206 06:13:07.072621 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-27fdn"] Dec 06 06:13:07 crc kubenswrapper[4809]: E1206 06:13:07.085680 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6rg67,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-lqkvl_openstack-operators(b376f569-0116-41f4-8a57-b5e5769e6a10): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 06 06:13:07 crc kubenswrapper[4809]: I1206 06:13:07.085751 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-ghwtq"] Dec 06 06:13:07 crc kubenswrapper[4809]: E1206 06:13:07.087803 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6rg67,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-lqkvl_openstack-operators(b376f569-0116-41f4-8a57-b5e5769e6a10): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 06 06:13:07 crc kubenswrapper[4809]: E1206 06:13:07.089641 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-lqkvl" podUID="b376f569-0116-41f4-8a57-b5e5769e6a10" Dec 06 06:13:07 crc kubenswrapper[4809]: I1206 06:13:07.106507 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-vqvvr"] Dec 06 06:13:07 crc kubenswrapper[4809]: I1206 06:13:07.108713 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-lqkvl"] Dec 06 06:13:07 crc kubenswrapper[4809]: I1206 06:13:07.115052 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-848xh"] Dec 06 06:13:07 crc kubenswrapper[4809]: E1206 06:13:07.124324 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wvj9h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-848xh_openstack-operators(0acb7c70-8d76-4bef-9d66-cd1213802a85): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 06 06:13:07 crc kubenswrapper[4809]: E1206 06:13:07.125918 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-848xh" podUID="0acb7c70-8d76-4bef-9d66-cd1213802a85" Dec 06 06:13:07 crc kubenswrapper[4809]: I1206 06:13:07.245918 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-27fdn" event={"ID":"e6e0283a-55e9-4613-b74d-81fbb840690e","Type":"ContainerStarted","Data":"66071a6bc9cc5dd07b39423072830d9d9e866118fab2103258cc21a31ee6d9ec"} Dec 06 06:13:07 crc kubenswrapper[4809]: I1206 06:13:07.248292 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7trbt" event={"ID":"45f87079-e852-4b47-a240-6f2bcf247509","Type":"ContainerStarted","Data":"00c95316606600d4a3e98fa371ee36940fb93932dfed2c62c3ce00ff88df381e"} Dec 06 06:13:07 crc kubenswrapper[4809]: I1206 06:13:07.250637 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4s24n" event={"ID":"35b44606-4fdd-4662-9550-adc5ed2478b3","Type":"ContainerStarted","Data":"139d70514675315bab247ca1de6dc0267d0dc89fc328345ca4d78d0a431b75a2"} Dec 06 06:13:07 crc kubenswrapper[4809]: I1206 06:13:07.252038 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-dfs6c" event={"ID":"2ac245f3-8f64-4b6f-a51f-09e4b5811412","Type":"ContainerStarted","Data":"365da07cf87af067c8857d234f82fbc4a58cb7d75a3435c0be159e2c272adcb0"} Dec 06 06:13:07 crc kubenswrapper[4809]: I1206 06:13:07.254320 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-ghwtq" event={"ID":"b379307b-aafd-42da-b761-38a556a0d143","Type":"ContainerStarted","Data":"279c8bdfee46be8264844b63a77d2d261becea1c4673f81479176e900b83bc63"} Dec 06 06:13:07 crc kubenswrapper[4809]: I1206 06:13:07.255488 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-gzv6k" event={"ID":"ced7591a-f80d-46a0-b8ea-024c2e37ae9f","Type":"ContainerStarted","Data":"699006b6715df3d5101aaabc256ba8e2f27c2b7ef07b8e3d0175661d9bc052dd"} Dec 06 06:13:07 crc kubenswrapper[4809]: I1206 06:13:07.256756 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-vqvvr" event={"ID":"803d54f1-fbac-4ec5-b486-efae1981b46d","Type":"ContainerStarted","Data":"9b84df5b9e7f23e9ea0b0fa07f5f078e1cd3db37f85619eee5987869bc890647"} Dec 06 06:13:07 crc kubenswrapper[4809]: I1206 06:13:07.257782 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-vh8d4" event={"ID":"93397035-a1a7-4c92-bdeb-5d0091dea181","Type":"ContainerStarted","Data":"157d2599a51a49504b02c129021812b75e6900ee08551140fe2eb9683e7afefc"} Dec 06 06:13:07 crc kubenswrapper[4809]: I1206 06:13:07.259450 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-kcp8k" event={"ID":"f736a8a4-ff7e-4179-951f-66405b1fd933","Type":"ContainerStarted","Data":"804be7383b6d61b0168ec0c805210d019f8e8a9113d571c46bac36e071719a0a"} Dec 06 06:13:07 crc kubenswrapper[4809]: I1206 06:13:07.260651 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-sxcr7" event={"ID":"c37c165f-5cd1-496f-b092-41efdfcd7eca","Type":"ContainerStarted","Data":"b39b5e67028083755345f9f33d9c9d3d5e083ba825d80f2851d0454af1cdbad0"} Dec 06 06:13:07 crc kubenswrapper[4809]: I1206 06:13:07.262337 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-848xh" event={"ID":"0acb7c70-8d76-4bef-9d66-cd1213802a85","Type":"ContainerStarted","Data":"167d2d9447ea9a3111edf2788c316e722b538c344b5620d19083d57d605dadc3"} Dec 06 06:13:07 crc kubenswrapper[4809]: I1206 06:13:07.264273 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-lqkvl" event={"ID":"b376f569-0116-41f4-8a57-b5e5769e6a10","Type":"ContainerStarted","Data":"ac65d23a131412accfcf1014c695d3434c6f631a30cd6f3d8a490a338dab2726"} Dec 06 06:13:07 crc kubenswrapper[4809]: E1206 06:13:07.265120 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-848xh" podUID="0acb7c70-8d76-4bef-9d66-cd1213802a85" Dec 06 06:13:07 crc kubenswrapper[4809]: I1206 06:13:07.266674 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-b4644b845-zxwvr" event={"ID":"24aed9fd-d792-4bce-82a8-3ea4d70094b4","Type":"ContainerStarted","Data":"3c9ae26c98e6c5b5b76181146bc319da85c18fc464f9bd10c4f4e25fe7ad7d93"} Dec 06 06:13:07 crc kubenswrapper[4809]: I1206 06:13:07.268885 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-zz2mm" event={"ID":"b55361fa-8a7c-46ff-92f6-9dd72dbc0536","Type":"ContainerStarted","Data":"e9800064b12c87919e4a708a9239c6bcd56bc733d986f648b0e17e6535dcf1ad"} Dec 06 06:13:07 crc kubenswrapper[4809]: E1206 06:13:07.268995 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-lqkvl" podUID="b376f569-0116-41f4-8a57-b5e5769e6a10" Dec 06 06:13:07 crc kubenswrapper[4809]: I1206 06:13:07.271106 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-bhbcl" event={"ID":"fd7dd0a5-5e39-4812-b93f-a27c39284d33","Type":"ContainerStarted","Data":"9971fb5e9ba957570a03facec758e4803129523eff3262691f5bf9e007d63de6"} Dec 06 06:13:08 crc kubenswrapper[4809]: I1206 06:13:08.034868 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/29c6d2cb-27fc-4b79-969a-462be41663e3-cert\") pod \"infra-operator-controller-manager-57548d458d-4b9l8\" (UID: \"29c6d2cb-27fc-4b79-969a-462be41663e3\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-4b9l8" Dec 06 06:13:08 crc kubenswrapper[4809]: E1206 06:13:08.035173 4809 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 06 06:13:08 crc kubenswrapper[4809]: E1206 06:13:08.035243 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/29c6d2cb-27fc-4b79-969a-462be41663e3-cert podName:29c6d2cb-27fc-4b79-969a-462be41663e3 nodeName:}" failed. No retries permitted until 2025-12-06 06:13:12.03522425 +0000 UTC m=+1316.924207192 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/29c6d2cb-27fc-4b79-969a-462be41663e3-cert") pod "infra-operator-controller-manager-57548d458d-4b9l8" (UID: "29c6d2cb-27fc-4b79-969a-462be41663e3") : secret "infra-operator-webhook-server-cert" not found Dec 06 06:13:08 crc kubenswrapper[4809]: E1206 06:13:08.289845 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-848xh" podUID="0acb7c70-8d76-4bef-9d66-cd1213802a85" Dec 06 06:13:08 crc kubenswrapper[4809]: E1206 06:13:08.290830 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-lqkvl" podUID="b376f569-0116-41f4-8a57-b5e5769e6a10" Dec 06 06:13:08 crc kubenswrapper[4809]: I1206 06:13:08.342158 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6fbac224-e82d-482e-aeea-c83a9f978775-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd462lbf\" (UID: \"6fbac224-e82d-482e-aeea-c83a9f978775\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd462lbf" Dec 06 06:13:08 crc kubenswrapper[4809]: E1206 06:13:08.342340 4809 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 06:13:08 crc kubenswrapper[4809]: E1206 06:13:08.342493 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6fbac224-e82d-482e-aeea-c83a9f978775-cert podName:6fbac224-e82d-482e-aeea-c83a9f978775 nodeName:}" failed. No retries permitted until 2025-12-06 06:13:12.342379818 +0000 UTC m=+1317.231362760 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6fbac224-e82d-482e-aeea-c83a9f978775-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd462lbf" (UID: "6fbac224-e82d-482e-aeea-c83a9f978775") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 06:13:09 crc kubenswrapper[4809]: I1206 06:13:09.059269 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4dd349b6-b792-451e-9cab-7c52a5768a3d-metrics-certs\") pod \"openstack-operator-controller-manager-5c4b598d95-mbx56\" (UID: \"4dd349b6-b792-451e-9cab-7c52a5768a3d\") " pod="openstack-operators/openstack-operator-controller-manager-5c4b598d95-mbx56" Dec 06 06:13:09 crc kubenswrapper[4809]: I1206 06:13:09.059665 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4dd349b6-b792-451e-9cab-7c52a5768a3d-webhook-certs\") pod \"openstack-operator-controller-manager-5c4b598d95-mbx56\" (UID: \"4dd349b6-b792-451e-9cab-7c52a5768a3d\") " pod="openstack-operators/openstack-operator-controller-manager-5c4b598d95-mbx56" Dec 06 06:13:09 crc kubenswrapper[4809]: E1206 06:13:09.060023 4809 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 06 06:13:09 crc kubenswrapper[4809]: E1206 06:13:09.060073 4809 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 06 06:13:09 crc kubenswrapper[4809]: E1206 06:13:09.060099 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4dd349b6-b792-451e-9cab-7c52a5768a3d-metrics-certs podName:4dd349b6-b792-451e-9cab-7c52a5768a3d nodeName:}" failed. No retries permitted until 2025-12-06 06:13:13.060077481 +0000 UTC m=+1317.949060423 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4dd349b6-b792-451e-9cab-7c52a5768a3d-metrics-certs") pod "openstack-operator-controller-manager-5c4b598d95-mbx56" (UID: "4dd349b6-b792-451e-9cab-7c52a5768a3d") : secret "metrics-server-cert" not found Dec 06 06:13:09 crc kubenswrapper[4809]: E1206 06:13:09.060113 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4dd349b6-b792-451e-9cab-7c52a5768a3d-webhook-certs podName:4dd349b6-b792-451e-9cab-7c52a5768a3d nodeName:}" failed. No retries permitted until 2025-12-06 06:13:13.060107792 +0000 UTC m=+1317.949090734 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/4dd349b6-b792-451e-9cab-7c52a5768a3d-webhook-certs") pod "openstack-operator-controller-manager-5c4b598d95-mbx56" (UID: "4dd349b6-b792-451e-9cab-7c52a5768a3d") : secret "webhook-server-cert" not found Dec 06 06:13:12 crc kubenswrapper[4809]: I1206 06:13:12.062173 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/29c6d2cb-27fc-4b79-969a-462be41663e3-cert\") pod \"infra-operator-controller-manager-57548d458d-4b9l8\" (UID: \"29c6d2cb-27fc-4b79-969a-462be41663e3\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-4b9l8" Dec 06 06:13:12 crc kubenswrapper[4809]: E1206 06:13:12.062455 4809 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 06 06:13:12 crc kubenswrapper[4809]: E1206 06:13:12.062540 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/29c6d2cb-27fc-4b79-969a-462be41663e3-cert podName:29c6d2cb-27fc-4b79-969a-462be41663e3 nodeName:}" failed. No retries permitted until 2025-12-06 06:13:20.06251763 +0000 UTC m=+1324.951500572 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/29c6d2cb-27fc-4b79-969a-462be41663e3-cert") pod "infra-operator-controller-manager-57548d458d-4b9l8" (UID: "29c6d2cb-27fc-4b79-969a-462be41663e3") : secret "infra-operator-webhook-server-cert" not found Dec 06 06:13:12 crc kubenswrapper[4809]: I1206 06:13:12.369610 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6fbac224-e82d-482e-aeea-c83a9f978775-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd462lbf\" (UID: \"6fbac224-e82d-482e-aeea-c83a9f978775\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd462lbf" Dec 06 06:13:12 crc kubenswrapper[4809]: E1206 06:13:12.369813 4809 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 06:13:12 crc kubenswrapper[4809]: E1206 06:13:12.369916 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6fbac224-e82d-482e-aeea-c83a9f978775-cert podName:6fbac224-e82d-482e-aeea-c83a9f978775 nodeName:}" failed. No retries permitted until 2025-12-06 06:13:20.369889314 +0000 UTC m=+1325.258872446 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6fbac224-e82d-482e-aeea-c83a9f978775-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd462lbf" (UID: "6fbac224-e82d-482e-aeea-c83a9f978775") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 06:13:13 crc kubenswrapper[4809]: I1206 06:13:13.083701 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4dd349b6-b792-451e-9cab-7c52a5768a3d-metrics-certs\") pod \"openstack-operator-controller-manager-5c4b598d95-mbx56\" (UID: \"4dd349b6-b792-451e-9cab-7c52a5768a3d\") " pod="openstack-operators/openstack-operator-controller-manager-5c4b598d95-mbx56" Dec 06 06:13:13 crc kubenswrapper[4809]: I1206 06:13:13.084153 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4dd349b6-b792-451e-9cab-7c52a5768a3d-webhook-certs\") pod \"openstack-operator-controller-manager-5c4b598d95-mbx56\" (UID: \"4dd349b6-b792-451e-9cab-7c52a5768a3d\") " pod="openstack-operators/openstack-operator-controller-manager-5c4b598d95-mbx56" Dec 06 06:13:13 crc kubenswrapper[4809]: E1206 06:13:13.083999 4809 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 06 06:13:13 crc kubenswrapper[4809]: E1206 06:13:13.084365 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4dd349b6-b792-451e-9cab-7c52a5768a3d-metrics-certs podName:4dd349b6-b792-451e-9cab-7c52a5768a3d nodeName:}" failed. No retries permitted until 2025-12-06 06:13:21.084341349 +0000 UTC m=+1325.973324291 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4dd349b6-b792-451e-9cab-7c52a5768a3d-metrics-certs") pod "openstack-operator-controller-manager-5c4b598d95-mbx56" (UID: "4dd349b6-b792-451e-9cab-7c52a5768a3d") : secret "metrics-server-cert" not found Dec 06 06:13:13 crc kubenswrapper[4809]: E1206 06:13:13.084483 4809 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 06 06:13:13 crc kubenswrapper[4809]: E1206 06:13:13.084531 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4dd349b6-b792-451e-9cab-7c52a5768a3d-webhook-certs podName:4dd349b6-b792-451e-9cab-7c52a5768a3d nodeName:}" failed. No retries permitted until 2025-12-06 06:13:21.084525304 +0000 UTC m=+1325.973508246 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/4dd349b6-b792-451e-9cab-7c52a5768a3d-webhook-certs") pod "openstack-operator-controller-manager-5c4b598d95-mbx56" (UID: "4dd349b6-b792-451e-9cab-7c52a5768a3d") : secret "webhook-server-cert" not found Dec 06 06:13:20 crc kubenswrapper[4809]: I1206 06:13:20.160002 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/29c6d2cb-27fc-4b79-969a-462be41663e3-cert\") pod \"infra-operator-controller-manager-57548d458d-4b9l8\" (UID: \"29c6d2cb-27fc-4b79-969a-462be41663e3\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-4b9l8" Dec 06 06:13:20 crc kubenswrapper[4809]: I1206 06:13:20.167349 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/29c6d2cb-27fc-4b79-969a-462be41663e3-cert\") pod \"infra-operator-controller-manager-57548d458d-4b9l8\" (UID: \"29c6d2cb-27fc-4b79-969a-462be41663e3\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-4b9l8" Dec 06 06:13:20 crc kubenswrapper[4809]: I1206 06:13:20.284895 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-pc84p" Dec 06 06:13:20 crc kubenswrapper[4809]: I1206 06:13:20.293234 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-4b9l8" Dec 06 06:13:20 crc kubenswrapper[4809]: I1206 06:13:20.467151 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6fbac224-e82d-482e-aeea-c83a9f978775-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd462lbf\" (UID: \"6fbac224-e82d-482e-aeea-c83a9f978775\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd462lbf" Dec 06 06:13:20 crc kubenswrapper[4809]: I1206 06:13:20.475545 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6fbac224-e82d-482e-aeea-c83a9f978775-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd462lbf\" (UID: \"6fbac224-e82d-482e-aeea-c83a9f978775\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd462lbf" Dec 06 06:13:20 crc kubenswrapper[4809]: I1206 06:13:20.549140 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-8fdp6" Dec 06 06:13:20 crc kubenswrapper[4809]: I1206 06:13:20.556613 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd462lbf" Dec 06 06:13:21 crc kubenswrapper[4809]: I1206 06:13:21.094540 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4dd349b6-b792-451e-9cab-7c52a5768a3d-webhook-certs\") pod \"openstack-operator-controller-manager-5c4b598d95-mbx56\" (UID: \"4dd349b6-b792-451e-9cab-7c52a5768a3d\") " pod="openstack-operators/openstack-operator-controller-manager-5c4b598d95-mbx56" Dec 06 06:13:21 crc kubenswrapper[4809]: I1206 06:13:21.094911 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4dd349b6-b792-451e-9cab-7c52a5768a3d-metrics-certs\") pod \"openstack-operator-controller-manager-5c4b598d95-mbx56\" (UID: \"4dd349b6-b792-451e-9cab-7c52a5768a3d\") " pod="openstack-operators/openstack-operator-controller-manager-5c4b598d95-mbx56" Dec 06 06:13:21 crc kubenswrapper[4809]: I1206 06:13:21.102414 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4dd349b6-b792-451e-9cab-7c52a5768a3d-webhook-certs\") pod \"openstack-operator-controller-manager-5c4b598d95-mbx56\" (UID: \"4dd349b6-b792-451e-9cab-7c52a5768a3d\") " pod="openstack-operators/openstack-operator-controller-manager-5c4b598d95-mbx56" Dec 06 06:13:21 crc kubenswrapper[4809]: I1206 06:13:21.110640 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4dd349b6-b792-451e-9cab-7c52a5768a3d-metrics-certs\") pod \"openstack-operator-controller-manager-5c4b598d95-mbx56\" (UID: \"4dd349b6-b792-451e-9cab-7c52a5768a3d\") " pod="openstack-operators/openstack-operator-controller-manager-5c4b598d95-mbx56" Dec 06 06:13:21 crc kubenswrapper[4809]: I1206 06:13:21.301597 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-kgrq2" Dec 06 06:13:21 crc kubenswrapper[4809]: I1206 06:13:21.309739 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-5c4b598d95-mbx56" Dec 06 06:13:33 crc kubenswrapper[4809]: E1206 06:13:33.397243 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:abdb733b01e92ac17f565762f30f1d075b44c16421bd06e557f6bb3c319e1809" Dec 06 06:13:33 crc kubenswrapper[4809]: E1206 06:13:33.399390 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:abdb733b01e92ac17f565762f30f1d075b44c16421bd06e557f6bb3c319e1809,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7bf52,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-77987cd8cd-szvjh_openstack-operators(204981d9-37b3-4584-b4bc-57df854143cc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 06:13:33 crc kubenswrapper[4809]: E1206 06:13:33.403020 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621" Dec 06 06:13:33 crc kubenswrapper[4809]: E1206 06:13:33.403317 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-m6kvf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-bhbcl_openstack-operators(fd7dd0a5-5e39-4812-b93f-a27c39284d33): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 06:13:35 crc kubenswrapper[4809]: E1206 06:13:35.083589 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168" Dec 06 06:13:35 crc kubenswrapper[4809]: E1206 06:13:35.084048 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6jlwg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-dfs6c_openstack-operators(2ac245f3-8f64-4b6f-a51f-09e4b5811412): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 06:13:35 crc kubenswrapper[4809]: E1206 06:13:35.695504 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d" Dec 06 06:13:35 crc kubenswrapper[4809]: E1206 06:13:35.696085 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-r48nh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-vqvvr_openstack-operators(803d54f1-fbac-4ec5-b486-efae1981b46d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 06:13:36 crc kubenswrapper[4809]: E1206 06:13:36.177599 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7" Dec 06 06:13:36 crc kubenswrapper[4809]: E1206 06:13:36.177906 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7qxhk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-56bbcc9d85-sxcr7_openstack-operators(c37c165f-5cd1-496f-b092-41efdfcd7eca): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 06:13:36 crc kubenswrapper[4809]: E1206 06:13:36.849052 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:0f523b7e2fa9e86fef986acf07d0c42d5658c475d565f11eaea926ebffcb6530" Dec 06 06:13:36 crc kubenswrapper[4809]: E1206 06:13:36.849382 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:0f523b7e2fa9e86fef986acf07d0c42d5658c475d565f11eaea926ebffcb6530,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-sdksd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-6c548fd776-7trbt_openstack-operators(45f87079-e852-4b47-a240-6f2bcf247509): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 06:13:40 crc kubenswrapper[4809]: E1206 06:13:40.620865 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.113:5001/openstack-k8s-operators/telemetry-operator:28cdb051e5b48bc8368c80adce71d0f7a4b620fe" Dec 06 06:13:40 crc kubenswrapper[4809]: E1206 06:13:40.621564 4809 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.113:5001/openstack-k8s-operators/telemetry-operator:28cdb051e5b48bc8368c80adce71d0f7a4b620fe" Dec 06 06:13:40 crc kubenswrapper[4809]: E1206 06:13:40.621944 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.113:5001/openstack-k8s-operators/telemetry-operator:28cdb051e5b48bc8368c80adce71d0f7a4b620fe,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4r4dh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-b4644b845-zxwvr_openstack-operators(24aed9fd-d792-4bce-82a8-3ea4d70094b4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 06:13:47 crc kubenswrapper[4809]: E1206 06:13:47.221029 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7" Dec 06 06:13:47 crc kubenswrapper[4809]: E1206 06:13:47.222076 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4txtd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-gzv6k_openstack-operators(ced7591a-f80d-46a0-b8ea-024c2e37ae9f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 06:13:47 crc kubenswrapper[4809]: E1206 06:13:47.851958 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670" Dec 06 06:13:47 crc kubenswrapper[4809]: E1206 06:13:47.852437 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2892s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-27fdn_openstack-operators(e6e0283a-55e9-4613-b74d-81fbb840690e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 06:13:48 crc kubenswrapper[4809]: I1206 06:13:48.286577 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd462lbf"] Dec 06 06:13:48 crc kubenswrapper[4809]: E1206 06:13:48.494725 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Dec 06 06:13:48 crc kubenswrapper[4809]: E1206 06:13:48.495166 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wvj9h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-848xh_openstack-operators(0acb7c70-8d76-4bef-9d66-cd1213802a85): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 06:13:48 crc kubenswrapper[4809]: E1206 06:13:48.496482 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-848xh" podUID="0acb7c70-8d76-4bef-9d66-cd1213802a85" Dec 06 06:13:48 crc kubenswrapper[4809]: I1206 06:13:48.963005 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-4b9l8"] Dec 06 06:13:49 crc kubenswrapper[4809]: I1206 06:13:49.128198 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5c4b598d95-mbx56"] Dec 06 06:13:49 crc kubenswrapper[4809]: I1206 06:13:49.684104 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd462lbf" event={"ID":"6fbac224-e82d-482e-aeea-c83a9f978775","Type":"ContainerStarted","Data":"892dc4b32ded4b3a53a88a5aca6b40e51092716ca5bb79fc2eac2959e48b7e79"} Dec 06 06:13:49 crc kubenswrapper[4809]: I1206 06:13:49.686383 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-l8clc" event={"ID":"44c524db-76cf-49c3-8c6e-1ad181d553ae","Type":"ContainerStarted","Data":"ad18e72add2ee4eca473ce71997c45898b84af749769671b40dee67d3f0b2c2e"} Dec 06 06:13:51 crc kubenswrapper[4809]: W1206 06:13:51.453327 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod29c6d2cb_27fc_4b79_969a_462be41663e3.slice/crio-73f02370829e421eca239bffdfd26d6104cedaac110fca53d134ca16d39db8cf WatchSource:0}: Error finding container 73f02370829e421eca239bffdfd26d6104cedaac110fca53d134ca16d39db8cf: Status 404 returned error can't find the container with id 73f02370829e421eca239bffdfd26d6104cedaac110fca53d134ca16d39db8cf Dec 06 06:13:51 crc kubenswrapper[4809]: W1206 06:13:51.454926 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4dd349b6_b792_451e_9cab_7c52a5768a3d.slice/crio-9972b473013311c99a911092f171d15af9cfce5138555e4cb7680c7c0f37d99e WatchSource:0}: Error finding container 9972b473013311c99a911092f171d15af9cfce5138555e4cb7680c7c0f37d99e: Status 404 returned error can't find the container with id 9972b473013311c99a911092f171d15af9cfce5138555e4cb7680c7c0f37d99e Dec 06 06:13:51 crc kubenswrapper[4809]: I1206 06:13:51.811164 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5c4b598d95-mbx56" event={"ID":"4dd349b6-b792-451e-9cab-7c52a5768a3d","Type":"ContainerStarted","Data":"9972b473013311c99a911092f171d15af9cfce5138555e4cb7680c7c0f37d99e"} Dec 06 06:13:51 crc kubenswrapper[4809]: I1206 06:13:51.812824 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-4b9l8" event={"ID":"29c6d2cb-27fc-4b79-969a-462be41663e3","Type":"ContainerStarted","Data":"73f02370829e421eca239bffdfd26d6104cedaac110fca53d134ca16d39db8cf"} Dec 06 06:13:52 crc kubenswrapper[4809]: I1206 06:13:52.835677 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-j4c62" event={"ID":"7233cfab-96cb-4459-a507-3031cac37720","Type":"ContainerStarted","Data":"d20d298e57955cccf19fa084f6aeae922f81ada2afd6542bd2143d5e9ca7613f"} Dec 06 06:13:52 crc kubenswrapper[4809]: I1206 06:13:52.838326 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-h69gw" event={"ID":"8be493fb-5378-467d-9fde-d5154cda3207","Type":"ContainerStarted","Data":"a3214ebb66e68e38eb792c553ad61e9da9c775af6a8739d8639029d919fc6ce0"} Dec 06 06:13:52 crc kubenswrapper[4809]: I1206 06:13:52.842065 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-ghwtq" event={"ID":"b379307b-aafd-42da-b761-38a556a0d143","Type":"ContainerStarted","Data":"6e4b94abdd189b59672919fa34380fde5189c4bafa6296237fccfa667d66f71f"} Dec 06 06:13:52 crc kubenswrapper[4809]: I1206 06:13:52.853264 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-zz2mm" event={"ID":"b55361fa-8a7c-46ff-92f6-9dd72dbc0536","Type":"ContainerStarted","Data":"fc52cb51241d075b457507adc84ffc36825c1ca57c47728ba7a6d596a95ad040"} Dec 06 06:13:52 crc kubenswrapper[4809]: I1206 06:13:52.874088 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-9gfg2" event={"ID":"ee834860-9609-482d-95db-82fa1a1a6941","Type":"ContainerStarted","Data":"a209f3828abfd84eda0251c8b70491c767ef48d092f4733be7097a419cfc654f"} Dec 06 06:13:53 crc kubenswrapper[4809]: I1206 06:13:53.912240 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-vh8d4" event={"ID":"93397035-a1a7-4c92-bdeb-5d0091dea181","Type":"ContainerStarted","Data":"85c24317a1d6f5c268814b6de11f058bfd0a8d6f0f3690b4ec06506eb42d0e06"} Dec 06 06:13:53 crc kubenswrapper[4809]: I1206 06:13:53.923455 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5c4b598d95-mbx56" event={"ID":"4dd349b6-b792-451e-9cab-7c52a5768a3d","Type":"ContainerStarted","Data":"e87662d6eddbc944f89fefa51e4f80fbdb462d3050f6ca27b4037946f08c7b04"} Dec 06 06:13:53 crc kubenswrapper[4809]: I1206 06:13:53.925069 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-5c4b598d95-mbx56" Dec 06 06:13:53 crc kubenswrapper[4809]: I1206 06:13:53.931653 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-kcp8k" event={"ID":"f736a8a4-ff7e-4179-951f-66405b1fd933","Type":"ContainerStarted","Data":"cff5ece0643bcde0532ad2e75ff008bbf9979218c500e580a1c47ddace4e304a"} Dec 06 06:13:53 crc kubenswrapper[4809]: I1206 06:13:53.938505 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4s24n" event={"ID":"35b44606-4fdd-4662-9550-adc5ed2478b3","Type":"ContainerStarted","Data":"d18047b19465a1cdc32402a05024df2387b6b3cfc5429634eb959419fbc6048c"} Dec 06 06:13:55 crc kubenswrapper[4809]: I1206 06:13:55.432061 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-5c4b598d95-mbx56" podStartSLOduration=51.432029456 podStartE2EDuration="51.432029456s" podCreationTimestamp="2025-12-06 06:13:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:13:53.962431859 +0000 UTC m=+1358.851414791" watchObservedRunningTime="2025-12-06 06:13:55.432029456 +0000 UTC m=+1360.321012398" Dec 06 06:13:56 crc kubenswrapper[4809]: E1206 06:13:56.254323 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 06 06:13:56 crc kubenswrapper[4809]: E1206 06:13:56.255383 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7bf52,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-77987cd8cd-szvjh_openstack-operators(204981d9-37b3-4584-b4bc-57df854143cc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 06:13:56 crc kubenswrapper[4809]: E1206 06:13:56.256778 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-szvjh" podUID="204981d9-37b3-4584-b4bc-57df854143cc" Dec 06 06:13:56 crc kubenswrapper[4809]: E1206 06:13:56.504991 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-998648c74-dfs6c" podUID="2ac245f3-8f64-4b6f-a51f-09e4b5811412" Dec 06 06:13:56 crc kubenswrapper[4809]: I1206 06:13:56.975825 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-lqkvl" event={"ID":"b376f569-0116-41f4-8a57-b5e5769e6a10","Type":"ContainerStarted","Data":"5c7db629ed3edf74098384a0a0132f41fe046c98a11cfe620d66c8f600323dc5"} Dec 06 06:13:56 crc kubenswrapper[4809]: I1206 06:13:56.981922 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-dfs6c" event={"ID":"2ac245f3-8f64-4b6f-a51f-09e4b5811412","Type":"ContainerStarted","Data":"5e1663a2661b5b1ef76f9c8d487c592083728b788ab35acbdd3cc16e580c04e9"} Dec 06 06:14:00 crc kubenswrapper[4809]: E1206 06:14:00.066192 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-vqvvr" podUID="803d54f1-fbac-4ec5-b486-efae1981b46d" Dec 06 06:14:00 crc kubenswrapper[4809]: E1206 06:14:00.290838 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-27fdn" podUID="e6e0283a-55e9-4613-b74d-81fbb840690e" Dec 06 06:14:00 crc kubenswrapper[4809]: E1206 06:14:00.397197 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-848xh" podUID="0acb7c70-8d76-4bef-9d66-cd1213802a85" Dec 06 06:14:00 crc kubenswrapper[4809]: E1206 06:14:00.773344 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-gzv6k" podUID="ced7591a-f80d-46a0-b8ea-024c2e37ae9f" Dec 06 06:14:00 crc kubenswrapper[4809]: E1206 06:14:00.775441 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-sxcr7" podUID="c37c165f-5cd1-496f-b092-41efdfcd7eca" Dec 06 06:14:00 crc kubenswrapper[4809]: E1206 06:14:00.894014 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-bhbcl" podUID="fd7dd0a5-5e39-4812-b93f-a27c39284d33" Dec 06 06:14:00 crc kubenswrapper[4809]: E1206 06:14:00.942402 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7trbt" podUID="45f87079-e852-4b47-a240-6f2bcf247509" Dec 06 06:14:00 crc kubenswrapper[4809]: E1206 06:14:00.965368 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-b4644b845-zxwvr" podUID="24aed9fd-d792-4bce-82a8-3ea4d70094b4" Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.060305 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-vh8d4" event={"ID":"93397035-a1a7-4c92-bdeb-5d0091dea181","Type":"ContainerStarted","Data":"e2c19725cd4eed63f5ff830e377126aad5c306d99cf0ad975bb902af9ecb6b53"} Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.069802 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd462lbf" event={"ID":"6fbac224-e82d-482e-aeea-c83a9f978775","Type":"ContainerStarted","Data":"e0d1203b2b4ea6eef7f2343a750bed29c71c8cb6e7a2fb1fa571c99d2d50700f"} Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.084960 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-j4c62" event={"ID":"7233cfab-96cb-4459-a507-3031cac37720","Type":"ContainerStarted","Data":"69cf5c5267e139bffab66f8d88ca0abee39e3d8ef2718cdad2e2ebe2eed82ce2"} Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.087854 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-j4c62" Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.097878 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-j4c62" Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.106203 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-27fdn" event={"ID":"e6e0283a-55e9-4613-b74d-81fbb840690e","Type":"ContainerStarted","Data":"4985a42deaad9910df91450094dfd03fe913173af8fec187c7e09900fa57eeda"} Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.135531 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-lqkvl" event={"ID":"b376f569-0116-41f4-8a57-b5e5769e6a10","Type":"ContainerStarted","Data":"ff96bc458c8e645732c023244b36c5a705c8a01e72faf52b4c60717b93881dc6"} Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.137203 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-lqkvl" Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.139879 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-lqkvl" Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.152036 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7trbt" event={"ID":"45f87079-e852-4b47-a240-6f2bcf247509","Type":"ContainerStarted","Data":"1c2c55f37283a52d8d7c1c5e7e1c4a8e5be119b273e0e8742435980b599bf9b6"} Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.153070 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-j4c62" podStartSLOduration=4.373963638 podStartE2EDuration="58.153039202s" podCreationTimestamp="2025-12-06 06:13:03 +0000 UTC" firstStartedPulling="2025-12-06 06:13:05.912196893 +0000 UTC m=+1310.801179835" lastFinishedPulling="2025-12-06 06:13:59.691272457 +0000 UTC m=+1364.580255399" observedRunningTime="2025-12-06 06:14:01.117409732 +0000 UTC m=+1366.006392694" watchObservedRunningTime="2025-12-06 06:14:01.153039202 +0000 UTC m=+1366.042022144" Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.167589 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-kcp8k" event={"ID":"f736a8a4-ff7e-4179-951f-66405b1fd933","Type":"ContainerStarted","Data":"373c46960c463bceb268c83ef5d88519903a56e97b9efc3faab494b449622af3"} Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.168307 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-kcp8k" Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.170355 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-kcp8k" Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.172489 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-szvjh" event={"ID":"204981d9-37b3-4584-b4bc-57df854143cc","Type":"ContainerStarted","Data":"9550b98ec46ad6f507f22ddf2775e84bc37708bd25ecae7d0b054cdf02676059"} Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.173433 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-4b9l8" event={"ID":"29c6d2cb-27fc-4b79-969a-462be41663e3","Type":"ContainerStarted","Data":"58b3abc20e667a1fcfa7e8dd2925566cfe5ccf91513884a14368e407258af682"} Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.175103 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-sxcr7" event={"ID":"c37c165f-5cd1-496f-b092-41efdfcd7eca","Type":"ContainerStarted","Data":"9f4685d530c3fde45f80a7abe0f280a8200b15021f692cf2df9634b3300cd260"} Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.209276 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-gzv6k" event={"ID":"ced7591a-f80d-46a0-b8ea-024c2e37ae9f","Type":"ContainerStarted","Data":"fc84a3e9f312f66bd0aec4f936e31dc7ac4c18c59f8de1061142e776a299a9b4"} Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.231458 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-bhbcl" event={"ID":"fd7dd0a5-5e39-4812-b93f-a27c39284d33","Type":"ContainerStarted","Data":"5693d9193425bde449fbcfe4bf0aaa1f4a214e0a4ec7ce44c4c24331df43d0d4"} Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.261743 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-lqkvl" podStartSLOduration=12.91127524 podStartE2EDuration="57.261707342s" podCreationTimestamp="2025-12-06 06:13:04 +0000 UTC" firstStartedPulling="2025-12-06 06:13:07.085440613 +0000 UTC m=+1311.974423555" lastFinishedPulling="2025-12-06 06:13:51.435872715 +0000 UTC m=+1356.324855657" observedRunningTime="2025-12-06 06:14:01.217323925 +0000 UTC m=+1366.106306867" watchObservedRunningTime="2025-12-06 06:14:01.261707342 +0000 UTC m=+1366.150690284" Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.272805 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-ghwtq" event={"ID":"b379307b-aafd-42da-b761-38a556a0d143","Type":"ContainerStarted","Data":"0577d07c5f34e25068226112c42fc703e8ec5a8ee3ef9bb8c6b67a0c93c93395"} Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.274411 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-ghwtq" Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.295156 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-ghwtq" Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.344374 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-5c4b598d95-mbx56" Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.358221 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-kcp8k" podStartSLOduration=4.277784839 podStartE2EDuration="57.358196094s" podCreationTimestamp="2025-12-06 06:13:04 +0000 UTC" firstStartedPulling="2025-12-06 06:13:06.46013231 +0000 UTC m=+1311.349115252" lastFinishedPulling="2025-12-06 06:13:59.540543555 +0000 UTC m=+1364.429526507" observedRunningTime="2025-12-06 06:14:01.314562838 +0000 UTC m=+1366.203545780" watchObservedRunningTime="2025-12-06 06:14:01.358196094 +0000 UTC m=+1366.247179036" Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.361696 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4s24n" Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.362851 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4s24n" event={"ID":"35b44606-4fdd-4662-9550-adc5ed2478b3","Type":"ContainerStarted","Data":"44b61ae44e97a49a0427520f14a0e85ac6ab7c856faf55fcf269e06d6dc50f0f"} Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.365767 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-b4644b845-zxwvr" event={"ID":"24aed9fd-d792-4bce-82a8-3ea4d70094b4","Type":"ContainerStarted","Data":"1d609c8672b7c792c21efbf6c23fafa175e4ebe3d0e56b252edfa2fa986a6dbe"} Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.366321 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4s24n" Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.373970 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-zz2mm" event={"ID":"b55361fa-8a7c-46ff-92f6-9dd72dbc0536","Type":"ContainerStarted","Data":"82537242d30944356c3ab051602a9c3e51bd8f671ca0d94d59e090a0c2469ffc"} Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.376794 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-zz2mm" Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.381232 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-zz2mm" Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.399955 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-ghwtq" podStartSLOduration=5.156842401 podStartE2EDuration="57.399915839s" podCreationTimestamp="2025-12-06 06:13:04 +0000 UTC" firstStartedPulling="2025-12-06 06:13:07.075993428 +0000 UTC m=+1311.964976370" lastFinishedPulling="2025-12-06 06:13:59.319066866 +0000 UTC m=+1364.208049808" observedRunningTime="2025-12-06 06:14:01.380755622 +0000 UTC m=+1366.269738564" watchObservedRunningTime="2025-12-06 06:14:01.399915839 +0000 UTC m=+1366.288898781" Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.477098 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-9gfg2" event={"ID":"ee834860-9609-482d-95db-82fa1a1a6941","Type":"ContainerStarted","Data":"bf94ed7f8f16134e7d10e443e1e724bbbad644d19c7e67763d27bdde650b4a5a"} Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.477156 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-9gfg2" Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.477195 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-9gfg2" Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.502256 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-l8clc" event={"ID":"44c524db-76cf-49c3-8c6e-1ad181d553ae","Type":"ContainerStarted","Data":"dab64c546e09e2ac2f3c3ecf9a24e639252eaccf2d71be16a0894fc56abcc144"} Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.503092 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-l8clc" Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.516290 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-h69gw" event={"ID":"8be493fb-5378-467d-9fde-d5154cda3207","Type":"ContainerStarted","Data":"6f13c743907271c0e70be219c589a13fefac4110714d42c6c75c1f88b367cade"} Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.517069 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-h69gw" Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.522092 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-l8clc" Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.535416 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-h69gw" Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.547577 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-dfs6c" event={"ID":"2ac245f3-8f64-4b6f-a51f-09e4b5811412","Type":"ContainerStarted","Data":"897bc62b81014e1ac53bcd7adeac999904fe75e456f5893225163ffb73130006"} Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.548918 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-dfs6c" Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.550878 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-vqvvr" event={"ID":"803d54f1-fbac-4ec5-b486-efae1981b46d","Type":"ContainerStarted","Data":"bc4a58482d3edf7a7a76f83c76b9a1c8e8d74de12f86e0762b1c4d6c8b048b37"} Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.598259 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-9gfg2" podStartSLOduration=4.9880349630000005 podStartE2EDuration="58.598222646s" podCreationTimestamp="2025-12-06 06:13:03 +0000 UTC" firstStartedPulling="2025-12-06 06:13:05.663770457 +0000 UTC m=+1310.552753389" lastFinishedPulling="2025-12-06 06:13:59.27395812 +0000 UTC m=+1364.162941072" observedRunningTime="2025-12-06 06:14:01.546258355 +0000 UTC m=+1366.435241307" watchObservedRunningTime="2025-12-06 06:14:01.598222646 +0000 UTC m=+1366.487205598" Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.599541 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-zz2mm" podStartSLOduration=4.969396591 podStartE2EDuration="57.599531761s" podCreationTimestamp="2025-12-06 06:13:04 +0000 UTC" firstStartedPulling="2025-12-06 06:13:07.063641645 +0000 UTC m=+1311.952624587" lastFinishedPulling="2025-12-06 06:13:59.693776815 +0000 UTC m=+1364.582759757" observedRunningTime="2025-12-06 06:14:01.584521107 +0000 UTC m=+1366.473504049" watchObservedRunningTime="2025-12-06 06:14:01.599531761 +0000 UTC m=+1366.488514713" Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.633409 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4s24n" podStartSLOduration=5.441479435 podStartE2EDuration="58.633387844s" podCreationTimestamp="2025-12-06 06:13:03 +0000 UTC" firstStartedPulling="2025-12-06 06:13:06.48497577 +0000 UTC m=+1311.373958712" lastFinishedPulling="2025-12-06 06:13:59.676884179 +0000 UTC m=+1364.565867121" observedRunningTime="2025-12-06 06:14:01.630442955 +0000 UTC m=+1366.519425897" watchObservedRunningTime="2025-12-06 06:14:01.633387844 +0000 UTC m=+1366.522370786" Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.683160 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-l8clc" podStartSLOduration=4.429795978 podStartE2EDuration="58.683122285s" podCreationTimestamp="2025-12-06 06:13:03 +0000 UTC" firstStartedPulling="2025-12-06 06:13:05.441053795 +0000 UTC m=+1310.330036737" lastFinishedPulling="2025-12-06 06:13:59.694380102 +0000 UTC m=+1364.583363044" observedRunningTime="2025-12-06 06:14:01.664597106 +0000 UTC m=+1366.553580068" watchObservedRunningTime="2025-12-06 06:14:01.683122285 +0000 UTC m=+1366.572105227" Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.747677 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-h69gw" podStartSLOduration=4.447416265 podStartE2EDuration="58.747647845s" podCreationTimestamp="2025-12-06 06:13:03 +0000 UTC" firstStartedPulling="2025-12-06 06:13:05.361019018 +0000 UTC m=+1310.250001960" lastFinishedPulling="2025-12-06 06:13:59.661250598 +0000 UTC m=+1364.550233540" observedRunningTime="2025-12-06 06:14:01.711142521 +0000 UTC m=+1366.600125463" watchObservedRunningTime="2025-12-06 06:14:01.747647845 +0000 UTC m=+1366.636630787" Dec 06 06:14:01 crc kubenswrapper[4809]: I1206 06:14:01.772502 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-dfs6c" podStartSLOduration=5.125144371 podStartE2EDuration="57.772474715s" podCreationTimestamp="2025-12-06 06:13:04 +0000 UTC" firstStartedPulling="2025-12-06 06:13:07.044648163 +0000 UTC m=+1311.933631105" lastFinishedPulling="2025-12-06 06:13:59.691978497 +0000 UTC m=+1364.580961449" observedRunningTime="2025-12-06 06:14:01.74150155 +0000 UTC m=+1366.630484502" watchObservedRunningTime="2025-12-06 06:14:01.772474715 +0000 UTC m=+1366.661457657" Dec 06 06:14:02 crc kubenswrapper[4809]: I1206 06:14:02.561515 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-4b9l8" event={"ID":"29c6d2cb-27fc-4b79-969a-462be41663e3","Type":"ContainerStarted","Data":"ff02cdd267ef39cd14565ebc2af68b85aff005bd43e22a0f53a0eab3e3bbb54f"} Dec 06 06:14:02 crc kubenswrapper[4809]: I1206 06:14:02.562015 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-4b9l8" Dec 06 06:14:02 crc kubenswrapper[4809]: I1206 06:14:02.563121 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-b4644b845-zxwvr" event={"ID":"24aed9fd-d792-4bce-82a8-3ea4d70094b4","Type":"ContainerStarted","Data":"83a8da67da6ae849e0e421f160c79a00273036172a22b6fc8c0226d385d3c08b"} Dec 06 06:14:02 crc kubenswrapper[4809]: I1206 06:14:02.564042 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-b4644b845-zxwvr" Dec 06 06:14:02 crc kubenswrapper[4809]: I1206 06:14:02.565098 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd462lbf" event={"ID":"6fbac224-e82d-482e-aeea-c83a9f978775","Type":"ContainerStarted","Data":"2fcd2d85b3eb4b71b7f614d2a079af5cbc5edc014d6d47a434c6be2e06a8ea4c"} Dec 06 06:14:02 crc kubenswrapper[4809]: I1206 06:14:02.565126 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd462lbf" Dec 06 06:14:02 crc kubenswrapper[4809]: I1206 06:14:02.567589 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-szvjh" event={"ID":"204981d9-37b3-4584-b4bc-57df854143cc","Type":"ContainerStarted","Data":"0038b713482e4271bf786e05fe30eb19204fe7dedf2b9daccc92a93fd778c3b9"} Dec 06 06:14:02 crc kubenswrapper[4809]: I1206 06:14:02.591487 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-57548d458d-4b9l8" podStartSLOduration=51.388293784 podStartE2EDuration="59.591457248s" podCreationTimestamp="2025-12-06 06:13:03 +0000 UTC" firstStartedPulling="2025-12-06 06:13:51.458469424 +0000 UTC m=+1356.347452366" lastFinishedPulling="2025-12-06 06:13:59.661632888 +0000 UTC m=+1364.550615830" observedRunningTime="2025-12-06 06:14:02.577942133 +0000 UTC m=+1367.466925075" watchObservedRunningTime="2025-12-06 06:14:02.591457248 +0000 UTC m=+1367.480440190" Dec 06 06:14:02 crc kubenswrapper[4809]: I1206 06:14:02.623696 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd462lbf" podStartSLOduration=47.663470537 podStartE2EDuration="58.623663076s" podCreationTimestamp="2025-12-06 06:13:04 +0000 UTC" firstStartedPulling="2025-12-06 06:13:48.700216166 +0000 UTC m=+1353.589199108" lastFinishedPulling="2025-12-06 06:13:59.660408715 +0000 UTC m=+1364.549391647" observedRunningTime="2025-12-06 06:14:02.61082336 +0000 UTC m=+1367.499806312" watchObservedRunningTime="2025-12-06 06:14:02.623663076 +0000 UTC m=+1367.512646008" Dec 06 06:14:02 crc kubenswrapper[4809]: I1206 06:14:02.639407 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-vh8d4" podStartSLOduration=5.370479834 podStartE2EDuration="58.63937306s" podCreationTimestamp="2025-12-06 06:13:04 +0000 UTC" firstStartedPulling="2025-12-06 06:13:06.42416586 +0000 UTC m=+1311.313148802" lastFinishedPulling="2025-12-06 06:13:59.693059086 +0000 UTC m=+1364.582042028" observedRunningTime="2025-12-06 06:14:02.63641775 +0000 UTC m=+1367.525400722" watchObservedRunningTime="2025-12-06 06:14:02.63937306 +0000 UTC m=+1367.528356012" Dec 06 06:14:02 crc kubenswrapper[4809]: I1206 06:14:02.662654 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-szvjh" podStartSLOduration=5.893133461 podStartE2EDuration="59.662626037s" podCreationTimestamp="2025-12-06 06:13:03 +0000 UTC" firstStartedPulling="2025-12-06 06:13:05.924995048 +0000 UTC m=+1310.813977980" lastFinishedPulling="2025-12-06 06:13:59.694487614 +0000 UTC m=+1364.583470556" observedRunningTime="2025-12-06 06:14:02.650595202 +0000 UTC m=+1367.539578144" watchObservedRunningTime="2025-12-06 06:14:02.662626037 +0000 UTC m=+1367.551608979" Dec 06 06:14:02 crc kubenswrapper[4809]: I1206 06:14:02.675224 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-b4644b845-zxwvr" podStartSLOduration=3.747082344 podStartE2EDuration="58.675199725s" podCreationTimestamp="2025-12-06 06:13:04 +0000 UTC" firstStartedPulling="2025-12-06 06:13:07.038442786 +0000 UTC m=+1311.927425728" lastFinishedPulling="2025-12-06 06:14:01.966560167 +0000 UTC m=+1366.855543109" observedRunningTime="2025-12-06 06:14:02.673983453 +0000 UTC m=+1367.562966395" watchObservedRunningTime="2025-12-06 06:14:02.675199725 +0000 UTC m=+1367.564182657" Dec 06 06:14:03 crc kubenswrapper[4809]: I1206 06:14:03.600125 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-vqvvr" event={"ID":"803d54f1-fbac-4ec5-b486-efae1981b46d","Type":"ContainerStarted","Data":"c315c94378e3dbfd1a01d387b95539c36326672328e93b7da85ac0bf89583043"} Dec 06 06:14:03 crc kubenswrapper[4809]: I1206 06:14:03.602253 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-vqvvr" Dec 06 06:14:03 crc kubenswrapper[4809]: I1206 06:14:03.633652 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-gzv6k" event={"ID":"ced7591a-f80d-46a0-b8ea-024c2e37ae9f","Type":"ContainerStarted","Data":"947c1f66dda10ade413913dcc8673297d5618b8f9370eff2d725a12a84d4b238"} Dec 06 06:14:03 crc kubenswrapper[4809]: I1206 06:14:03.635028 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-gzv6k" Dec 06 06:14:03 crc kubenswrapper[4809]: I1206 06:14:03.653548 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-bhbcl" event={"ID":"fd7dd0a5-5e39-4812-b93f-a27c39284d33","Type":"ContainerStarted","Data":"8ae35726489025911e54c463aca49b4867c80c782fee9c12fdc71b7233053771"} Dec 06 06:14:03 crc kubenswrapper[4809]: I1206 06:14:03.654823 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-bhbcl" Dec 06 06:14:03 crc kubenswrapper[4809]: I1206 06:14:03.661622 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-27fdn" event={"ID":"e6e0283a-55e9-4613-b74d-81fbb840690e","Type":"ContainerStarted","Data":"d8000d64b8e3c303bbcb8031f5aa11caca5df8c447f2da020aa04a793431d29f"} Dec 06 06:14:03 crc kubenswrapper[4809]: I1206 06:14:03.662102 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-27fdn" Dec 06 06:14:03 crc kubenswrapper[4809]: I1206 06:14:03.677839 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7trbt" event={"ID":"45f87079-e852-4b47-a240-6f2bcf247509","Type":"ContainerStarted","Data":"be1ecfef1755d34a69c3ac07b04e729bc32095ffdca7c1d32de77e47839fe82a"} Dec 06 06:14:03 crc kubenswrapper[4809]: I1206 06:14:03.679266 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7trbt" Dec 06 06:14:03 crc kubenswrapper[4809]: I1206 06:14:03.679989 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-gzv6k" podStartSLOduration=3.3364514 podStartE2EDuration="59.679972458s" podCreationTimestamp="2025-12-06 06:13:04 +0000 UTC" firstStartedPulling="2025-12-06 06:13:06.434652754 +0000 UTC m=+1311.323635696" lastFinishedPulling="2025-12-06 06:14:02.778173812 +0000 UTC m=+1367.667156754" observedRunningTime="2025-12-06 06:14:03.672264081 +0000 UTC m=+1368.561247023" watchObservedRunningTime="2025-12-06 06:14:03.679972458 +0000 UTC m=+1368.568955400" Dec 06 06:14:03 crc kubenswrapper[4809]: I1206 06:14:03.694613 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-vqvvr" podStartSLOduration=4.002247225 podStartE2EDuration="59.694585003s" podCreationTimestamp="2025-12-06 06:13:04 +0000 UTC" firstStartedPulling="2025-12-06 06:13:07.065166316 +0000 UTC m=+1311.954149258" lastFinishedPulling="2025-12-06 06:14:02.757504094 +0000 UTC m=+1367.646487036" observedRunningTime="2025-12-06 06:14:03.637729829 +0000 UTC m=+1368.526712781" watchObservedRunningTime="2025-12-06 06:14:03.694585003 +0000 UTC m=+1368.583567935" Dec 06 06:14:03 crc kubenswrapper[4809]: I1206 06:14:03.695504 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-sxcr7" event={"ID":"c37c165f-5cd1-496f-b092-41efdfcd7eca","Type":"ContainerStarted","Data":"938fa47565f71087c3323653b9e66e5e62af1fedce8aa73c476bc9220a1ef829"} Dec 06 06:14:03 crc kubenswrapper[4809]: I1206 06:14:03.722195 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-szvjh" Dec 06 06:14:03 crc kubenswrapper[4809]: I1206 06:14:03.729529 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-bhbcl" podStartSLOduration=4.005306075 podStartE2EDuration="59.729500613s" podCreationTimestamp="2025-12-06 06:13:04 +0000 UTC" firstStartedPulling="2025-12-06 06:13:07.058611139 +0000 UTC m=+1311.947594081" lastFinishedPulling="2025-12-06 06:14:02.782805677 +0000 UTC m=+1367.671788619" observedRunningTime="2025-12-06 06:14:03.719558285 +0000 UTC m=+1368.608541237" watchObservedRunningTime="2025-12-06 06:14:03.729500613 +0000 UTC m=+1368.618483545" Dec 06 06:14:03 crc kubenswrapper[4809]: I1206 06:14:03.830955 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-27fdn" podStartSLOduration=4.128119137 podStartE2EDuration="59.830907738s" podCreationTimestamp="2025-12-06 06:13:04 +0000 UTC" firstStartedPulling="2025-12-06 06:13:07.078264449 +0000 UTC m=+1311.967247391" lastFinishedPulling="2025-12-06 06:14:02.78105305 +0000 UTC m=+1367.670035992" observedRunningTime="2025-12-06 06:14:03.79278341 +0000 UTC m=+1368.681766372" watchObservedRunningTime="2025-12-06 06:14:03.830907738 +0000 UTC m=+1368.719890680" Dec 06 06:14:03 crc kubenswrapper[4809]: I1206 06:14:03.866536 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-sxcr7" podStartSLOduration=3.590249422 podStartE2EDuration="59.866493927s" podCreationTimestamp="2025-12-06 06:13:04 +0000 UTC" firstStartedPulling="2025-12-06 06:13:06.490760386 +0000 UTC m=+1311.379743328" lastFinishedPulling="2025-12-06 06:14:02.767004891 +0000 UTC m=+1367.655987833" observedRunningTime="2025-12-06 06:14:03.817525377 +0000 UTC m=+1368.706508329" watchObservedRunningTime="2025-12-06 06:14:03.866493927 +0000 UTC m=+1368.755476869" Dec 06 06:14:03 crc kubenswrapper[4809]: I1206 06:14:03.890763 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7trbt" podStartSLOduration=4.552782342 podStartE2EDuration="1m0.890729861s" podCreationTimestamp="2025-12-06 06:13:03 +0000 UTC" firstStartedPulling="2025-12-06 06:13:06.491059493 +0000 UTC m=+1311.380042435" lastFinishedPulling="2025-12-06 06:14:02.829007012 +0000 UTC m=+1367.717989954" observedRunningTime="2025-12-06 06:14:03.847430384 +0000 UTC m=+1368.736413326" watchObservedRunningTime="2025-12-06 06:14:03.890729861 +0000 UTC m=+1368.779712803" Dec 06 06:14:04 crc kubenswrapper[4809]: I1206 06:14:04.496606 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:14:04 crc kubenswrapper[4809]: I1206 06:14:04.496692 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:14:04 crc kubenswrapper[4809]: I1206 06:14:04.709330 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-sxcr7" Dec 06 06:14:04 crc kubenswrapper[4809]: I1206 06:14:04.765948 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-vh8d4" Dec 06 06:14:04 crc kubenswrapper[4809]: I1206 06:14:04.768097 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-vh8d4" Dec 06 06:14:10 crc kubenswrapper[4809]: I1206 06:14:10.299577 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-4b9l8" Dec 06 06:14:10 crc kubenswrapper[4809]: I1206 06:14:10.563492 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd462lbf" Dec 06 06:14:14 crc kubenswrapper[4809]: I1206 06:14:14.585398 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-szvjh" Dec 06 06:14:14 crc kubenswrapper[4809]: I1206 06:14:14.632850 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7trbt" Dec 06 06:14:14 crc kubenswrapper[4809]: I1206 06:14:14.700485 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-gzv6k" Dec 06 06:14:14 crc kubenswrapper[4809]: I1206 06:14:14.806103 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-sxcr7" Dec 06 06:14:14 crc kubenswrapper[4809]: I1206 06:14:14.902127 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-27fdn" Dec 06 06:14:14 crc kubenswrapper[4809]: I1206 06:14:14.943065 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-dfs6c" Dec 06 06:14:15 crc kubenswrapper[4809]: I1206 06:14:15.102512 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-vqvvr" Dec 06 06:14:15 crc kubenswrapper[4809]: I1206 06:14:15.138007 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-b4644b845-zxwvr" Dec 06 06:14:15 crc kubenswrapper[4809]: I1206 06:14:15.279872 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-bhbcl" Dec 06 06:14:15 crc kubenswrapper[4809]: I1206 06:14:15.836145 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-848xh" event={"ID":"0acb7c70-8d76-4bef-9d66-cd1213802a85","Type":"ContainerStarted","Data":"5b7d701ff49943c6dae8ce192b7b90fb8863b7174186553533fc5c1c827c92d7"} Dec 06 06:14:15 crc kubenswrapper[4809]: I1206 06:14:15.860464 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-848xh" podStartSLOduration=3.864661904 podStartE2EDuration="1m11.860435662s" podCreationTimestamp="2025-12-06 06:13:04 +0000 UTC" firstStartedPulling="2025-12-06 06:13:07.124182657 +0000 UTC m=+1312.013165599" lastFinishedPulling="2025-12-06 06:14:15.119956415 +0000 UTC m=+1380.008939357" observedRunningTime="2025-12-06 06:14:15.858188431 +0000 UTC m=+1380.747171363" watchObservedRunningTime="2025-12-06 06:14:15.860435662 +0000 UTC m=+1380.749418604" Dec 06 06:14:16 crc kubenswrapper[4809]: I1206 06:14:16.329332 4809 scope.go:117] "RemoveContainer" containerID="3eafd7a93b4c2781932a6a5633692e5a998e2267bd814a6fab9555c1cfe27a50" Dec 06 06:14:16 crc kubenswrapper[4809]: I1206 06:14:16.354023 4809 scope.go:117] "RemoveContainer" containerID="f14f6cc8a1629492e948858156ba6278de7f6860843677daa792447867714126" Dec 06 06:14:16 crc kubenswrapper[4809]: I1206 06:14:16.389827 4809 scope.go:117] "RemoveContainer" containerID="3766b2cb5f760f6460c334541c09b657c21dfe1d8c5958c91ee89455ef9506f2" Dec 06 06:14:29 crc kubenswrapper[4809]: I1206 06:14:29.610444 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-xm4sd"] Dec 06 06:14:29 crc kubenswrapper[4809]: I1206 06:14:29.613165 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-xm4sd" Dec 06 06:14:29 crc kubenswrapper[4809]: I1206 06:14:29.616311 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 06 06:14:29 crc kubenswrapper[4809]: I1206 06:14:29.616711 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 06 06:14:29 crc kubenswrapper[4809]: I1206 06:14:29.617189 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 06 06:14:29 crc kubenswrapper[4809]: I1206 06:14:29.620812 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-nwtmr" Dec 06 06:14:29 crc kubenswrapper[4809]: I1206 06:14:29.625750 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-xm4sd"] Dec 06 06:14:29 crc kubenswrapper[4809]: I1206 06:14:29.694348 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tr9lq\" (UniqueName: \"kubernetes.io/projected/68b27fd9-c5ea-4e7a-9639-afb6adecf768-kube-api-access-tr9lq\") pod \"dnsmasq-dns-675f4bcbfc-xm4sd\" (UID: \"68b27fd9-c5ea-4e7a-9639-afb6adecf768\") " pod="openstack/dnsmasq-dns-675f4bcbfc-xm4sd" Dec 06 06:14:29 crc kubenswrapper[4809]: I1206 06:14:29.694543 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68b27fd9-c5ea-4e7a-9639-afb6adecf768-config\") pod \"dnsmasq-dns-675f4bcbfc-xm4sd\" (UID: \"68b27fd9-c5ea-4e7a-9639-afb6adecf768\") " pod="openstack/dnsmasq-dns-675f4bcbfc-xm4sd" Dec 06 06:14:29 crc kubenswrapper[4809]: I1206 06:14:29.702835 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-9bft5"] Dec 06 06:14:29 crc kubenswrapper[4809]: I1206 06:14:29.705445 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-9bft5" Dec 06 06:14:29 crc kubenswrapper[4809]: I1206 06:14:29.707986 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 06 06:14:29 crc kubenswrapper[4809]: I1206 06:14:29.731676 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-9bft5"] Dec 06 06:14:29 crc kubenswrapper[4809]: I1206 06:14:29.797108 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5lhx\" (UniqueName: \"kubernetes.io/projected/78537fe2-3d4b-4659-82fc-1c8fa02d0f66-kube-api-access-b5lhx\") pod \"dnsmasq-dns-78dd6ddcc-9bft5\" (UID: \"78537fe2-3d4b-4659-82fc-1c8fa02d0f66\") " pod="openstack/dnsmasq-dns-78dd6ddcc-9bft5" Dec 06 06:14:29 crc kubenswrapper[4809]: I1206 06:14:29.797180 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78537fe2-3d4b-4659-82fc-1c8fa02d0f66-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-9bft5\" (UID: \"78537fe2-3d4b-4659-82fc-1c8fa02d0f66\") " pod="openstack/dnsmasq-dns-78dd6ddcc-9bft5" Dec 06 06:14:29 crc kubenswrapper[4809]: I1206 06:14:29.797307 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68b27fd9-c5ea-4e7a-9639-afb6adecf768-config\") pod \"dnsmasq-dns-675f4bcbfc-xm4sd\" (UID: \"68b27fd9-c5ea-4e7a-9639-afb6adecf768\") " pod="openstack/dnsmasq-dns-675f4bcbfc-xm4sd" Dec 06 06:14:29 crc kubenswrapper[4809]: I1206 06:14:29.797347 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78537fe2-3d4b-4659-82fc-1c8fa02d0f66-config\") pod \"dnsmasq-dns-78dd6ddcc-9bft5\" (UID: \"78537fe2-3d4b-4659-82fc-1c8fa02d0f66\") " pod="openstack/dnsmasq-dns-78dd6ddcc-9bft5" Dec 06 06:14:29 crc kubenswrapper[4809]: I1206 06:14:29.797492 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tr9lq\" (UniqueName: \"kubernetes.io/projected/68b27fd9-c5ea-4e7a-9639-afb6adecf768-kube-api-access-tr9lq\") pod \"dnsmasq-dns-675f4bcbfc-xm4sd\" (UID: \"68b27fd9-c5ea-4e7a-9639-afb6adecf768\") " pod="openstack/dnsmasq-dns-675f4bcbfc-xm4sd" Dec 06 06:14:29 crc kubenswrapper[4809]: I1206 06:14:29.798754 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68b27fd9-c5ea-4e7a-9639-afb6adecf768-config\") pod \"dnsmasq-dns-675f4bcbfc-xm4sd\" (UID: \"68b27fd9-c5ea-4e7a-9639-afb6adecf768\") " pod="openstack/dnsmasq-dns-675f4bcbfc-xm4sd" Dec 06 06:14:29 crc kubenswrapper[4809]: I1206 06:14:29.817767 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tr9lq\" (UniqueName: \"kubernetes.io/projected/68b27fd9-c5ea-4e7a-9639-afb6adecf768-kube-api-access-tr9lq\") pod \"dnsmasq-dns-675f4bcbfc-xm4sd\" (UID: \"68b27fd9-c5ea-4e7a-9639-afb6adecf768\") " pod="openstack/dnsmasq-dns-675f4bcbfc-xm4sd" Dec 06 06:14:29 crc kubenswrapper[4809]: I1206 06:14:29.901316 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78537fe2-3d4b-4659-82fc-1c8fa02d0f66-config\") pod \"dnsmasq-dns-78dd6ddcc-9bft5\" (UID: \"78537fe2-3d4b-4659-82fc-1c8fa02d0f66\") " pod="openstack/dnsmasq-dns-78dd6ddcc-9bft5" Dec 06 06:14:29 crc kubenswrapper[4809]: I1206 06:14:29.901504 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5lhx\" (UniqueName: \"kubernetes.io/projected/78537fe2-3d4b-4659-82fc-1c8fa02d0f66-kube-api-access-b5lhx\") pod \"dnsmasq-dns-78dd6ddcc-9bft5\" (UID: \"78537fe2-3d4b-4659-82fc-1c8fa02d0f66\") " pod="openstack/dnsmasq-dns-78dd6ddcc-9bft5" Dec 06 06:14:29 crc kubenswrapper[4809]: I1206 06:14:29.901536 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78537fe2-3d4b-4659-82fc-1c8fa02d0f66-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-9bft5\" (UID: \"78537fe2-3d4b-4659-82fc-1c8fa02d0f66\") " pod="openstack/dnsmasq-dns-78dd6ddcc-9bft5" Dec 06 06:14:29 crc kubenswrapper[4809]: I1206 06:14:29.903081 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78537fe2-3d4b-4659-82fc-1c8fa02d0f66-config\") pod \"dnsmasq-dns-78dd6ddcc-9bft5\" (UID: \"78537fe2-3d4b-4659-82fc-1c8fa02d0f66\") " pod="openstack/dnsmasq-dns-78dd6ddcc-9bft5" Dec 06 06:14:29 crc kubenswrapper[4809]: I1206 06:14:29.903694 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78537fe2-3d4b-4659-82fc-1c8fa02d0f66-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-9bft5\" (UID: \"78537fe2-3d4b-4659-82fc-1c8fa02d0f66\") " pod="openstack/dnsmasq-dns-78dd6ddcc-9bft5" Dec 06 06:14:29 crc kubenswrapper[4809]: I1206 06:14:29.922122 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5lhx\" (UniqueName: \"kubernetes.io/projected/78537fe2-3d4b-4659-82fc-1c8fa02d0f66-kube-api-access-b5lhx\") pod \"dnsmasq-dns-78dd6ddcc-9bft5\" (UID: \"78537fe2-3d4b-4659-82fc-1c8fa02d0f66\") " pod="openstack/dnsmasq-dns-78dd6ddcc-9bft5" Dec 06 06:14:29 crc kubenswrapper[4809]: I1206 06:14:29.943243 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-xm4sd" Dec 06 06:14:30 crc kubenswrapper[4809]: I1206 06:14:30.045594 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-9bft5" Dec 06 06:14:30 crc kubenswrapper[4809]: W1206 06:14:30.423315 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod68b27fd9_c5ea_4e7a_9639_afb6adecf768.slice/crio-7a4833bc6acf76d02c3fae3b9da7758e13cd8e7e8048b3b234324c9fc01d7359 WatchSource:0}: Error finding container 7a4833bc6acf76d02c3fae3b9da7758e13cd8e7e8048b3b234324c9fc01d7359: Status 404 returned error can't find the container with id 7a4833bc6acf76d02c3fae3b9da7758e13cd8e7e8048b3b234324c9fc01d7359 Dec 06 06:14:30 crc kubenswrapper[4809]: I1206 06:14:30.424524 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-xm4sd"] Dec 06 06:14:30 crc kubenswrapper[4809]: I1206 06:14:30.612646 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-9bft5"] Dec 06 06:14:30 crc kubenswrapper[4809]: I1206 06:14:30.998832 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-9bft5" event={"ID":"78537fe2-3d4b-4659-82fc-1c8fa02d0f66","Type":"ContainerStarted","Data":"07f0f5ce67102fa261a1d47b13ff3c8824183d2e6e07d1d15afbe5c667a1c656"} Dec 06 06:14:31 crc kubenswrapper[4809]: I1206 06:14:31.000620 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-xm4sd" event={"ID":"68b27fd9-c5ea-4e7a-9639-afb6adecf768","Type":"ContainerStarted","Data":"7a4833bc6acf76d02c3fae3b9da7758e13cd8e7e8048b3b234324c9fc01d7359"} Dec 06 06:14:32 crc kubenswrapper[4809]: I1206 06:14:32.657162 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-xm4sd"] Dec 06 06:14:32 crc kubenswrapper[4809]: I1206 06:14:32.680102 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-k6s2d"] Dec 06 06:14:32 crc kubenswrapper[4809]: I1206 06:14:32.682233 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-k6s2d" Dec 06 06:14:32 crc kubenswrapper[4809]: I1206 06:14:32.718058 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-k6s2d"] Dec 06 06:14:32 crc kubenswrapper[4809]: I1206 06:14:32.780725 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de481cc6-bc8d-4a07-991b-72cad738f1d5-dns-svc\") pod \"dnsmasq-dns-666b6646f7-k6s2d\" (UID: \"de481cc6-bc8d-4a07-991b-72cad738f1d5\") " pod="openstack/dnsmasq-dns-666b6646f7-k6s2d" Dec 06 06:14:32 crc kubenswrapper[4809]: I1206 06:14:32.780796 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de481cc6-bc8d-4a07-991b-72cad738f1d5-config\") pod \"dnsmasq-dns-666b6646f7-k6s2d\" (UID: \"de481cc6-bc8d-4a07-991b-72cad738f1d5\") " pod="openstack/dnsmasq-dns-666b6646f7-k6s2d" Dec 06 06:14:32 crc kubenswrapper[4809]: I1206 06:14:32.780857 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sc5mn\" (UniqueName: \"kubernetes.io/projected/de481cc6-bc8d-4a07-991b-72cad738f1d5-kube-api-access-sc5mn\") pod \"dnsmasq-dns-666b6646f7-k6s2d\" (UID: \"de481cc6-bc8d-4a07-991b-72cad738f1d5\") " pod="openstack/dnsmasq-dns-666b6646f7-k6s2d" Dec 06 06:14:32 crc kubenswrapper[4809]: I1206 06:14:32.885459 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de481cc6-bc8d-4a07-991b-72cad738f1d5-dns-svc\") pod \"dnsmasq-dns-666b6646f7-k6s2d\" (UID: \"de481cc6-bc8d-4a07-991b-72cad738f1d5\") " pod="openstack/dnsmasq-dns-666b6646f7-k6s2d" Dec 06 06:14:32 crc kubenswrapper[4809]: I1206 06:14:32.885541 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de481cc6-bc8d-4a07-991b-72cad738f1d5-config\") pod \"dnsmasq-dns-666b6646f7-k6s2d\" (UID: \"de481cc6-bc8d-4a07-991b-72cad738f1d5\") " pod="openstack/dnsmasq-dns-666b6646f7-k6s2d" Dec 06 06:14:32 crc kubenswrapper[4809]: I1206 06:14:32.885590 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sc5mn\" (UniqueName: \"kubernetes.io/projected/de481cc6-bc8d-4a07-991b-72cad738f1d5-kube-api-access-sc5mn\") pod \"dnsmasq-dns-666b6646f7-k6s2d\" (UID: \"de481cc6-bc8d-4a07-991b-72cad738f1d5\") " pod="openstack/dnsmasq-dns-666b6646f7-k6s2d" Dec 06 06:14:32 crc kubenswrapper[4809]: I1206 06:14:32.887472 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de481cc6-bc8d-4a07-991b-72cad738f1d5-dns-svc\") pod \"dnsmasq-dns-666b6646f7-k6s2d\" (UID: \"de481cc6-bc8d-4a07-991b-72cad738f1d5\") " pod="openstack/dnsmasq-dns-666b6646f7-k6s2d" Dec 06 06:14:32 crc kubenswrapper[4809]: I1206 06:14:32.888872 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de481cc6-bc8d-4a07-991b-72cad738f1d5-config\") pod \"dnsmasq-dns-666b6646f7-k6s2d\" (UID: \"de481cc6-bc8d-4a07-991b-72cad738f1d5\") " pod="openstack/dnsmasq-dns-666b6646f7-k6s2d" Dec 06 06:14:32 crc kubenswrapper[4809]: I1206 06:14:32.941688 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sc5mn\" (UniqueName: \"kubernetes.io/projected/de481cc6-bc8d-4a07-991b-72cad738f1d5-kube-api-access-sc5mn\") pod \"dnsmasq-dns-666b6646f7-k6s2d\" (UID: \"de481cc6-bc8d-4a07-991b-72cad738f1d5\") " pod="openstack/dnsmasq-dns-666b6646f7-k6s2d" Dec 06 06:14:33 crc kubenswrapper[4809]: I1206 06:14:33.002345 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-9bft5"] Dec 06 06:14:33 crc kubenswrapper[4809]: I1206 06:14:33.020109 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-k6s2d" Dec 06 06:14:33 crc kubenswrapper[4809]: I1206 06:14:33.046615 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-gr99p"] Dec 06 06:14:33 crc kubenswrapper[4809]: I1206 06:14:33.048746 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-gr99p" Dec 06 06:14:33 crc kubenswrapper[4809]: I1206 06:14:33.065826 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-gr99p"] Dec 06 06:14:33 crc kubenswrapper[4809]: I1206 06:14:33.093148 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a8bfcbfb-3946-420c-b90f-f390887145f0-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-gr99p\" (UID: \"a8bfcbfb-3946-420c-b90f-f390887145f0\") " pod="openstack/dnsmasq-dns-57d769cc4f-gr99p" Dec 06 06:14:33 crc kubenswrapper[4809]: I1206 06:14:33.093445 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8bfcbfb-3946-420c-b90f-f390887145f0-config\") pod \"dnsmasq-dns-57d769cc4f-gr99p\" (UID: \"a8bfcbfb-3946-420c-b90f-f390887145f0\") " pod="openstack/dnsmasq-dns-57d769cc4f-gr99p" Dec 06 06:14:33 crc kubenswrapper[4809]: I1206 06:14:33.093680 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zt2b\" (UniqueName: \"kubernetes.io/projected/a8bfcbfb-3946-420c-b90f-f390887145f0-kube-api-access-4zt2b\") pod \"dnsmasq-dns-57d769cc4f-gr99p\" (UID: \"a8bfcbfb-3946-420c-b90f-f390887145f0\") " pod="openstack/dnsmasq-dns-57d769cc4f-gr99p" Dec 06 06:14:33 crc kubenswrapper[4809]: I1206 06:14:33.197172 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zt2b\" (UniqueName: \"kubernetes.io/projected/a8bfcbfb-3946-420c-b90f-f390887145f0-kube-api-access-4zt2b\") pod \"dnsmasq-dns-57d769cc4f-gr99p\" (UID: \"a8bfcbfb-3946-420c-b90f-f390887145f0\") " pod="openstack/dnsmasq-dns-57d769cc4f-gr99p" Dec 06 06:14:33 crc kubenswrapper[4809]: I1206 06:14:33.197290 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a8bfcbfb-3946-420c-b90f-f390887145f0-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-gr99p\" (UID: \"a8bfcbfb-3946-420c-b90f-f390887145f0\") " pod="openstack/dnsmasq-dns-57d769cc4f-gr99p" Dec 06 06:14:33 crc kubenswrapper[4809]: I1206 06:14:33.197360 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8bfcbfb-3946-420c-b90f-f390887145f0-config\") pod \"dnsmasq-dns-57d769cc4f-gr99p\" (UID: \"a8bfcbfb-3946-420c-b90f-f390887145f0\") " pod="openstack/dnsmasq-dns-57d769cc4f-gr99p" Dec 06 06:14:33 crc kubenswrapper[4809]: I1206 06:14:33.199176 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a8bfcbfb-3946-420c-b90f-f390887145f0-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-gr99p\" (UID: \"a8bfcbfb-3946-420c-b90f-f390887145f0\") " pod="openstack/dnsmasq-dns-57d769cc4f-gr99p" Dec 06 06:14:33 crc kubenswrapper[4809]: I1206 06:14:33.205865 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8bfcbfb-3946-420c-b90f-f390887145f0-config\") pod \"dnsmasq-dns-57d769cc4f-gr99p\" (UID: \"a8bfcbfb-3946-420c-b90f-f390887145f0\") " pod="openstack/dnsmasq-dns-57d769cc4f-gr99p" Dec 06 06:14:33 crc kubenswrapper[4809]: I1206 06:14:33.230584 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zt2b\" (UniqueName: \"kubernetes.io/projected/a8bfcbfb-3946-420c-b90f-f390887145f0-kube-api-access-4zt2b\") pod \"dnsmasq-dns-57d769cc4f-gr99p\" (UID: \"a8bfcbfb-3946-420c-b90f-f390887145f0\") " pod="openstack/dnsmasq-dns-57d769cc4f-gr99p" Dec 06 06:14:33 crc kubenswrapper[4809]: I1206 06:14:33.392675 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-gr99p" Dec 06 06:14:33 crc kubenswrapper[4809]: I1206 06:14:33.761155 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-k6s2d"] Dec 06 06:14:33 crc kubenswrapper[4809]: I1206 06:14:33.832307 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 06:14:33 crc kubenswrapper[4809]: I1206 06:14:33.836896 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 06 06:14:33 crc kubenswrapper[4809]: I1206 06:14:33.843302 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 06 06:14:33 crc kubenswrapper[4809]: I1206 06:14:33.843471 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-r9k2s" Dec 06 06:14:33 crc kubenswrapper[4809]: I1206 06:14:33.843598 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 06 06:14:33 crc kubenswrapper[4809]: I1206 06:14:33.843836 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 06 06:14:33 crc kubenswrapper[4809]: I1206 06:14:33.843842 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 06 06:14:33 crc kubenswrapper[4809]: I1206 06:14:33.844055 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 06 06:14:33 crc kubenswrapper[4809]: I1206 06:14:33.844117 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 06 06:14:33 crc kubenswrapper[4809]: I1206 06:14:33.845638 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 06:14:33 crc kubenswrapper[4809]: I1206 06:14:33.918782 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/eb1bd4de-6456-4817-a85d-ab0176153f2c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"eb1bd4de-6456-4817-a85d-ab0176153f2c\") " pod="openstack/rabbitmq-server-0" Dec 06 06:14:33 crc kubenswrapper[4809]: I1206 06:14:33.918916 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/eb1bd4de-6456-4817-a85d-ab0176153f2c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"eb1bd4de-6456-4817-a85d-ab0176153f2c\") " pod="openstack/rabbitmq-server-0" Dec 06 06:14:33 crc kubenswrapper[4809]: I1206 06:14:33.919061 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/eb1bd4de-6456-4817-a85d-ab0176153f2c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"eb1bd4de-6456-4817-a85d-ab0176153f2c\") " pod="openstack/rabbitmq-server-0" Dec 06 06:14:33 crc kubenswrapper[4809]: I1206 06:14:33.919137 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/eb1bd4de-6456-4817-a85d-ab0176153f2c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"eb1bd4de-6456-4817-a85d-ab0176153f2c\") " pod="openstack/rabbitmq-server-0" Dec 06 06:14:33 crc kubenswrapper[4809]: I1206 06:14:33.919173 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/eb1bd4de-6456-4817-a85d-ab0176153f2c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"eb1bd4de-6456-4817-a85d-ab0176153f2c\") " pod="openstack/rabbitmq-server-0" Dec 06 06:14:33 crc kubenswrapper[4809]: I1206 06:14:33.919224 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"eb1bd4de-6456-4817-a85d-ab0176153f2c\") " pod="openstack/rabbitmq-server-0" Dec 06 06:14:33 crc kubenswrapper[4809]: I1206 06:14:33.919245 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/eb1bd4de-6456-4817-a85d-ab0176153f2c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"eb1bd4de-6456-4817-a85d-ab0176153f2c\") " pod="openstack/rabbitmq-server-0" Dec 06 06:14:33 crc kubenswrapper[4809]: I1206 06:14:33.919262 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/eb1bd4de-6456-4817-a85d-ab0176153f2c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"eb1bd4de-6456-4817-a85d-ab0176153f2c\") " pod="openstack/rabbitmq-server-0" Dec 06 06:14:33 crc kubenswrapper[4809]: I1206 06:14:33.919333 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/eb1bd4de-6456-4817-a85d-ab0176153f2c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"eb1bd4de-6456-4817-a85d-ab0176153f2c\") " pod="openstack/rabbitmq-server-0" Dec 06 06:14:33 crc kubenswrapper[4809]: I1206 06:14:33.919386 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfn22\" (UniqueName: \"kubernetes.io/projected/eb1bd4de-6456-4817-a85d-ab0176153f2c-kube-api-access-hfn22\") pod \"rabbitmq-server-0\" (UID: \"eb1bd4de-6456-4817-a85d-ab0176153f2c\") " pod="openstack/rabbitmq-server-0" Dec 06 06:14:33 crc kubenswrapper[4809]: I1206 06:14:33.919458 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eb1bd4de-6456-4817-a85d-ab0176153f2c-config-data\") pod \"rabbitmq-server-0\" (UID: \"eb1bd4de-6456-4817-a85d-ab0176153f2c\") " pod="openstack/rabbitmq-server-0" Dec 06 06:14:33 crc kubenswrapper[4809]: I1206 06:14:33.965749 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-gr99p"] Dec 06 06:14:33 crc kubenswrapper[4809]: W1206 06:14:33.968787 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda8bfcbfb_3946_420c_b90f_f390887145f0.slice/crio-ffd1005a8f86fc583dc6e3396dc2eb2e0c68d026f81484a8edef9ff0ac96ce7f WatchSource:0}: Error finding container ffd1005a8f86fc583dc6e3396dc2eb2e0c68d026f81484a8edef9ff0ac96ce7f: Status 404 returned error can't find the container with id ffd1005a8f86fc583dc6e3396dc2eb2e0c68d026f81484a8edef9ff0ac96ce7f Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.022555 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/eb1bd4de-6456-4817-a85d-ab0176153f2c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"eb1bd4de-6456-4817-a85d-ab0176153f2c\") " pod="openstack/rabbitmq-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.022648 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/eb1bd4de-6456-4817-a85d-ab0176153f2c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"eb1bd4de-6456-4817-a85d-ab0176153f2c\") " pod="openstack/rabbitmq-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.022705 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/eb1bd4de-6456-4817-a85d-ab0176153f2c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"eb1bd4de-6456-4817-a85d-ab0176153f2c\") " pod="openstack/rabbitmq-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.022744 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"eb1bd4de-6456-4817-a85d-ab0176153f2c\") " pod="openstack/rabbitmq-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.022766 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/eb1bd4de-6456-4817-a85d-ab0176153f2c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"eb1bd4de-6456-4817-a85d-ab0176153f2c\") " pod="openstack/rabbitmq-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.022788 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/eb1bd4de-6456-4817-a85d-ab0176153f2c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"eb1bd4de-6456-4817-a85d-ab0176153f2c\") " pod="openstack/rabbitmq-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.022855 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/eb1bd4de-6456-4817-a85d-ab0176153f2c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"eb1bd4de-6456-4817-a85d-ab0176153f2c\") " pod="openstack/rabbitmq-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.023190 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfn22\" (UniqueName: \"kubernetes.io/projected/eb1bd4de-6456-4817-a85d-ab0176153f2c-kube-api-access-hfn22\") pod \"rabbitmq-server-0\" (UID: \"eb1bd4de-6456-4817-a85d-ab0176153f2c\") " pod="openstack/rabbitmq-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.023239 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eb1bd4de-6456-4817-a85d-ab0176153f2c-config-data\") pod \"rabbitmq-server-0\" (UID: \"eb1bd4de-6456-4817-a85d-ab0176153f2c\") " pod="openstack/rabbitmq-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.023321 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/eb1bd4de-6456-4817-a85d-ab0176153f2c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"eb1bd4de-6456-4817-a85d-ab0176153f2c\") " pod="openstack/rabbitmq-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.023348 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/eb1bd4de-6456-4817-a85d-ab0176153f2c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"eb1bd4de-6456-4817-a85d-ab0176153f2c\") " pod="openstack/rabbitmq-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.024259 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/eb1bd4de-6456-4817-a85d-ab0176153f2c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"eb1bd4de-6456-4817-a85d-ab0176153f2c\") " pod="openstack/rabbitmq-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.024377 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"eb1bd4de-6456-4817-a85d-ab0176153f2c\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.025326 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eb1bd4de-6456-4817-a85d-ab0176153f2c-config-data\") pod \"rabbitmq-server-0\" (UID: \"eb1bd4de-6456-4817-a85d-ab0176153f2c\") " pod="openstack/rabbitmq-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.025549 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/eb1bd4de-6456-4817-a85d-ab0176153f2c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"eb1bd4de-6456-4817-a85d-ab0176153f2c\") " pod="openstack/rabbitmq-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.025699 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/eb1bd4de-6456-4817-a85d-ab0176153f2c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"eb1bd4de-6456-4817-a85d-ab0176153f2c\") " pod="openstack/rabbitmq-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.026200 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/eb1bd4de-6456-4817-a85d-ab0176153f2c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"eb1bd4de-6456-4817-a85d-ab0176153f2c\") " pod="openstack/rabbitmq-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.040869 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/eb1bd4de-6456-4817-a85d-ab0176153f2c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"eb1bd4de-6456-4817-a85d-ab0176153f2c\") " pod="openstack/rabbitmq-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.040986 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/eb1bd4de-6456-4817-a85d-ab0176153f2c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"eb1bd4de-6456-4817-a85d-ab0176153f2c\") " pod="openstack/rabbitmq-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.041309 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/eb1bd4de-6456-4817-a85d-ab0176153f2c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"eb1bd4de-6456-4817-a85d-ab0176153f2c\") " pod="openstack/rabbitmq-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.043867 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfn22\" (UniqueName: \"kubernetes.io/projected/eb1bd4de-6456-4817-a85d-ab0176153f2c-kube-api-access-hfn22\") pod \"rabbitmq-server-0\" (UID: \"eb1bd4de-6456-4817-a85d-ab0176153f2c\") " pod="openstack/rabbitmq-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.045652 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/eb1bd4de-6456-4817-a85d-ab0176153f2c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"eb1bd4de-6456-4817-a85d-ab0176153f2c\") " pod="openstack/rabbitmq-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.068483 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"eb1bd4de-6456-4817-a85d-ab0176153f2c\") " pod="openstack/rabbitmq-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.075630 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-k6s2d" event={"ID":"de481cc6-bc8d-4a07-991b-72cad738f1d5","Type":"ContainerStarted","Data":"2c8fde9c0e29b0ed8ca9b413a966026f7c882190f281e14ed5c48d80d16cf1f4"} Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.078389 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-gr99p" event={"ID":"a8bfcbfb-3946-420c-b90f-f390887145f0","Type":"ContainerStarted","Data":"ffd1005a8f86fc583dc6e3396dc2eb2e0c68d026f81484a8edef9ff0ac96ce7f"} Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.170539 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.260817 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.265140 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.271101 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.271748 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.271994 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.272220 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.272342 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-tqjgv" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.272450 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.272581 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.319955 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.334630 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1ceabc13-38c5-4425-8d85-c3afa8935243-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"1ceabc13-38c5-4425-8d85-c3afa8935243\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.334705 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1ceabc13-38c5-4425-8d85-c3afa8935243-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"1ceabc13-38c5-4425-8d85-c3afa8935243\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.334746 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1ceabc13-38c5-4425-8d85-c3afa8935243\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.334765 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1ceabc13-38c5-4425-8d85-c3afa8935243-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1ceabc13-38c5-4425-8d85-c3afa8935243\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.334793 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1ceabc13-38c5-4425-8d85-c3afa8935243-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"1ceabc13-38c5-4425-8d85-c3afa8935243\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.334864 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gf4mc\" (UniqueName: \"kubernetes.io/projected/1ceabc13-38c5-4425-8d85-c3afa8935243-kube-api-access-gf4mc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1ceabc13-38c5-4425-8d85-c3afa8935243\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.334913 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1ceabc13-38c5-4425-8d85-c3afa8935243-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"1ceabc13-38c5-4425-8d85-c3afa8935243\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.334949 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1ceabc13-38c5-4425-8d85-c3afa8935243-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1ceabc13-38c5-4425-8d85-c3afa8935243\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.334967 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1ceabc13-38c5-4425-8d85-c3afa8935243-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"1ceabc13-38c5-4425-8d85-c3afa8935243\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.334984 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1ceabc13-38c5-4425-8d85-c3afa8935243-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"1ceabc13-38c5-4425-8d85-c3afa8935243\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.335200 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1ceabc13-38c5-4425-8d85-c3afa8935243-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"1ceabc13-38c5-4425-8d85-c3afa8935243\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.437351 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1ceabc13-38c5-4425-8d85-c3afa8935243-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"1ceabc13-38c5-4425-8d85-c3afa8935243\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.437437 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1ceabc13-38c5-4425-8d85-c3afa8935243-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"1ceabc13-38c5-4425-8d85-c3afa8935243\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.437466 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1ceabc13-38c5-4425-8d85-c3afa8935243-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"1ceabc13-38c5-4425-8d85-c3afa8935243\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.437502 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1ceabc13-38c5-4425-8d85-c3afa8935243\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.437521 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1ceabc13-38c5-4425-8d85-c3afa8935243-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1ceabc13-38c5-4425-8d85-c3afa8935243\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.437541 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1ceabc13-38c5-4425-8d85-c3afa8935243-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"1ceabc13-38c5-4425-8d85-c3afa8935243\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.437586 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gf4mc\" (UniqueName: \"kubernetes.io/projected/1ceabc13-38c5-4425-8d85-c3afa8935243-kube-api-access-gf4mc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1ceabc13-38c5-4425-8d85-c3afa8935243\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.437621 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1ceabc13-38c5-4425-8d85-c3afa8935243-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"1ceabc13-38c5-4425-8d85-c3afa8935243\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.437644 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1ceabc13-38c5-4425-8d85-c3afa8935243-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1ceabc13-38c5-4425-8d85-c3afa8935243\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.437661 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1ceabc13-38c5-4425-8d85-c3afa8935243-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"1ceabc13-38c5-4425-8d85-c3afa8935243\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.437676 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1ceabc13-38c5-4425-8d85-c3afa8935243-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"1ceabc13-38c5-4425-8d85-c3afa8935243\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.439280 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1ceabc13-38c5-4425-8d85-c3afa8935243-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"1ceabc13-38c5-4425-8d85-c3afa8935243\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.439752 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1ceabc13-38c5-4425-8d85-c3afa8935243-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1ceabc13-38c5-4425-8d85-c3afa8935243\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.440056 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1ceabc13-38c5-4425-8d85-c3afa8935243-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"1ceabc13-38c5-4425-8d85-c3afa8935243\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.443596 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1ceabc13-38c5-4425-8d85-c3afa8935243\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.444395 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1ceabc13-38c5-4425-8d85-c3afa8935243-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1ceabc13-38c5-4425-8d85-c3afa8935243\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.445800 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1ceabc13-38c5-4425-8d85-c3afa8935243-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"1ceabc13-38c5-4425-8d85-c3afa8935243\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.453547 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1ceabc13-38c5-4425-8d85-c3afa8935243-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"1ceabc13-38c5-4425-8d85-c3afa8935243\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.456190 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1ceabc13-38c5-4425-8d85-c3afa8935243-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"1ceabc13-38c5-4425-8d85-c3afa8935243\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.457309 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1ceabc13-38c5-4425-8d85-c3afa8935243-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"1ceabc13-38c5-4425-8d85-c3afa8935243\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.464993 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1ceabc13-38c5-4425-8d85-c3afa8935243-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"1ceabc13-38c5-4425-8d85-c3afa8935243\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.478060 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gf4mc\" (UniqueName: \"kubernetes.io/projected/1ceabc13-38c5-4425-8d85-c3afa8935243-kube-api-access-gf4mc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1ceabc13-38c5-4425-8d85-c3afa8935243\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.498917 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.499240 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.508115 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1ceabc13-38c5-4425-8d85-c3afa8935243\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.606902 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:14:34 crc kubenswrapper[4809]: I1206 06:14:34.821553 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 06:14:35 crc kubenswrapper[4809]: I1206 06:14:35.109162 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"eb1bd4de-6456-4817-a85d-ab0176153f2c","Type":"ContainerStarted","Data":"e04ab83414709672446678fb1f5632ec7fd7e0a434fdd4fe2bad2914e8502fa3"} Dec 06 06:14:35 crc kubenswrapper[4809]: I1206 06:14:35.142796 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 06:14:35 crc kubenswrapper[4809]: I1206 06:14:35.586246 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 06 06:14:35 crc kubenswrapper[4809]: I1206 06:14:35.589290 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 06 06:14:35 crc kubenswrapper[4809]: I1206 06:14:35.608065 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 06 06:14:35 crc kubenswrapper[4809]: I1206 06:14:35.609091 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-lfhr5" Dec 06 06:14:35 crc kubenswrapper[4809]: I1206 06:14:35.610676 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 06 06:14:35 crc kubenswrapper[4809]: I1206 06:14:35.614098 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 06 06:14:35 crc kubenswrapper[4809]: I1206 06:14:35.619403 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 06 06:14:35 crc kubenswrapper[4809]: I1206 06:14:35.622885 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 06 06:14:35 crc kubenswrapper[4809]: I1206 06:14:35.669655 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7c596e14-b96f-4e5f-a8ac-3dc66fa0276d-kolla-config\") pod \"openstack-galera-0\" (UID: \"7c596e14-b96f-4e5f-a8ac-3dc66fa0276d\") " pod="openstack/openstack-galera-0" Dec 06 06:14:35 crc kubenswrapper[4809]: I1206 06:14:35.669737 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7ch4\" (UniqueName: \"kubernetes.io/projected/7c596e14-b96f-4e5f-a8ac-3dc66fa0276d-kube-api-access-b7ch4\") pod \"openstack-galera-0\" (UID: \"7c596e14-b96f-4e5f-a8ac-3dc66fa0276d\") " pod="openstack/openstack-galera-0" Dec 06 06:14:35 crc kubenswrapper[4809]: I1206 06:14:35.669767 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"7c596e14-b96f-4e5f-a8ac-3dc66fa0276d\") " pod="openstack/openstack-galera-0" Dec 06 06:14:35 crc kubenswrapper[4809]: I1206 06:14:35.670078 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c596e14-b96f-4e5f-a8ac-3dc66fa0276d-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"7c596e14-b96f-4e5f-a8ac-3dc66fa0276d\") " pod="openstack/openstack-galera-0" Dec 06 06:14:35 crc kubenswrapper[4809]: I1206 06:14:35.670117 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c596e14-b96f-4e5f-a8ac-3dc66fa0276d-operator-scripts\") pod \"openstack-galera-0\" (UID: \"7c596e14-b96f-4e5f-a8ac-3dc66fa0276d\") " pod="openstack/openstack-galera-0" Dec 06 06:14:35 crc kubenswrapper[4809]: I1206 06:14:35.670156 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7c596e14-b96f-4e5f-a8ac-3dc66fa0276d-config-data-generated\") pod \"openstack-galera-0\" (UID: \"7c596e14-b96f-4e5f-a8ac-3dc66fa0276d\") " pod="openstack/openstack-galera-0" Dec 06 06:14:35 crc kubenswrapper[4809]: I1206 06:14:35.670184 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c596e14-b96f-4e5f-a8ac-3dc66fa0276d-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"7c596e14-b96f-4e5f-a8ac-3dc66fa0276d\") " pod="openstack/openstack-galera-0" Dec 06 06:14:35 crc kubenswrapper[4809]: I1206 06:14:35.670256 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7c596e14-b96f-4e5f-a8ac-3dc66fa0276d-config-data-default\") pod \"openstack-galera-0\" (UID: \"7c596e14-b96f-4e5f-a8ac-3dc66fa0276d\") " pod="openstack/openstack-galera-0" Dec 06 06:14:35 crc kubenswrapper[4809]: I1206 06:14:35.772374 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7c596e14-b96f-4e5f-a8ac-3dc66fa0276d-config-data-default\") pod \"openstack-galera-0\" (UID: \"7c596e14-b96f-4e5f-a8ac-3dc66fa0276d\") " pod="openstack/openstack-galera-0" Dec 06 06:14:35 crc kubenswrapper[4809]: I1206 06:14:35.772475 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7c596e14-b96f-4e5f-a8ac-3dc66fa0276d-kolla-config\") pod \"openstack-galera-0\" (UID: \"7c596e14-b96f-4e5f-a8ac-3dc66fa0276d\") " pod="openstack/openstack-galera-0" Dec 06 06:14:35 crc kubenswrapper[4809]: I1206 06:14:35.772537 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7ch4\" (UniqueName: \"kubernetes.io/projected/7c596e14-b96f-4e5f-a8ac-3dc66fa0276d-kube-api-access-b7ch4\") pod \"openstack-galera-0\" (UID: \"7c596e14-b96f-4e5f-a8ac-3dc66fa0276d\") " pod="openstack/openstack-galera-0" Dec 06 06:14:35 crc kubenswrapper[4809]: I1206 06:14:35.772575 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"7c596e14-b96f-4e5f-a8ac-3dc66fa0276d\") " pod="openstack/openstack-galera-0" Dec 06 06:14:35 crc kubenswrapper[4809]: I1206 06:14:35.772637 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c596e14-b96f-4e5f-a8ac-3dc66fa0276d-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"7c596e14-b96f-4e5f-a8ac-3dc66fa0276d\") " pod="openstack/openstack-galera-0" Dec 06 06:14:35 crc kubenswrapper[4809]: I1206 06:14:35.772680 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c596e14-b96f-4e5f-a8ac-3dc66fa0276d-operator-scripts\") pod \"openstack-galera-0\" (UID: \"7c596e14-b96f-4e5f-a8ac-3dc66fa0276d\") " pod="openstack/openstack-galera-0" Dec 06 06:14:35 crc kubenswrapper[4809]: I1206 06:14:35.772727 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7c596e14-b96f-4e5f-a8ac-3dc66fa0276d-config-data-generated\") pod \"openstack-galera-0\" (UID: \"7c596e14-b96f-4e5f-a8ac-3dc66fa0276d\") " pod="openstack/openstack-galera-0" Dec 06 06:14:35 crc kubenswrapper[4809]: I1206 06:14:35.772774 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c596e14-b96f-4e5f-a8ac-3dc66fa0276d-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"7c596e14-b96f-4e5f-a8ac-3dc66fa0276d\") " pod="openstack/openstack-galera-0" Dec 06 06:14:35 crc kubenswrapper[4809]: I1206 06:14:35.774336 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7c596e14-b96f-4e5f-a8ac-3dc66fa0276d-config-data-generated\") pod \"openstack-galera-0\" (UID: \"7c596e14-b96f-4e5f-a8ac-3dc66fa0276d\") " pod="openstack/openstack-galera-0" Dec 06 06:14:35 crc kubenswrapper[4809]: I1206 06:14:35.782289 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c596e14-b96f-4e5f-a8ac-3dc66fa0276d-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"7c596e14-b96f-4e5f-a8ac-3dc66fa0276d\") " pod="openstack/openstack-galera-0" Dec 06 06:14:35 crc kubenswrapper[4809]: I1206 06:14:35.782307 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7c596e14-b96f-4e5f-a8ac-3dc66fa0276d-config-data-default\") pod \"openstack-galera-0\" (UID: \"7c596e14-b96f-4e5f-a8ac-3dc66fa0276d\") " pod="openstack/openstack-galera-0" Dec 06 06:14:35 crc kubenswrapper[4809]: I1206 06:14:35.783054 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7c596e14-b96f-4e5f-a8ac-3dc66fa0276d-kolla-config\") pod \"openstack-galera-0\" (UID: \"7c596e14-b96f-4e5f-a8ac-3dc66fa0276d\") " pod="openstack/openstack-galera-0" Dec 06 06:14:35 crc kubenswrapper[4809]: I1206 06:14:35.785555 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"7c596e14-b96f-4e5f-a8ac-3dc66fa0276d\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/openstack-galera-0" Dec 06 06:14:35 crc kubenswrapper[4809]: I1206 06:14:35.787120 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c596e14-b96f-4e5f-a8ac-3dc66fa0276d-operator-scripts\") pod \"openstack-galera-0\" (UID: \"7c596e14-b96f-4e5f-a8ac-3dc66fa0276d\") " pod="openstack/openstack-galera-0" Dec 06 06:14:35 crc kubenswrapper[4809]: I1206 06:14:35.797409 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7ch4\" (UniqueName: \"kubernetes.io/projected/7c596e14-b96f-4e5f-a8ac-3dc66fa0276d-kube-api-access-b7ch4\") pod \"openstack-galera-0\" (UID: \"7c596e14-b96f-4e5f-a8ac-3dc66fa0276d\") " pod="openstack/openstack-galera-0" Dec 06 06:14:35 crc kubenswrapper[4809]: I1206 06:14:35.818497 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c596e14-b96f-4e5f-a8ac-3dc66fa0276d-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"7c596e14-b96f-4e5f-a8ac-3dc66fa0276d\") " pod="openstack/openstack-galera-0" Dec 06 06:14:35 crc kubenswrapper[4809]: I1206 06:14:35.833355 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"7c596e14-b96f-4e5f-a8ac-3dc66fa0276d\") " pod="openstack/openstack-galera-0" Dec 06 06:14:35 crc kubenswrapper[4809]: I1206 06:14:35.933701 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 06 06:14:36 crc kubenswrapper[4809]: I1206 06:14:36.968361 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 06 06:14:36 crc kubenswrapper[4809]: I1206 06:14:36.971156 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 06 06:14:36 crc kubenswrapper[4809]: I1206 06:14:36.976320 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 06 06:14:36 crc kubenswrapper[4809]: I1206 06:14:36.976572 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 06 06:14:36 crc kubenswrapper[4809]: I1206 06:14:36.977640 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-8tw9v" Dec 06 06:14:36 crc kubenswrapper[4809]: I1206 06:14:36.980482 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 06 06:14:36 crc kubenswrapper[4809]: I1206 06:14:36.985257 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.106392 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e7056938-6e36-4631-adfb-01f116eba651-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"e7056938-6e36-4631-adfb-01f116eba651\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.106541 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7056938-6e36-4631-adfb-01f116eba651-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"e7056938-6e36-4631-adfb-01f116eba651\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.106614 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7056938-6e36-4631-adfb-01f116eba651-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"e7056938-6e36-4631-adfb-01f116eba651\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.106644 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dhvx\" (UniqueName: \"kubernetes.io/projected/e7056938-6e36-4631-adfb-01f116eba651-kube-api-access-2dhvx\") pod \"openstack-cell1-galera-0\" (UID: \"e7056938-6e36-4631-adfb-01f116eba651\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.106687 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7056938-6e36-4631-adfb-01f116eba651-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"e7056938-6e36-4631-adfb-01f116eba651\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.106726 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"e7056938-6e36-4631-adfb-01f116eba651\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.106765 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e7056938-6e36-4631-adfb-01f116eba651-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"e7056938-6e36-4631-adfb-01f116eba651\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.106790 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e7056938-6e36-4631-adfb-01f116eba651-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"e7056938-6e36-4631-adfb-01f116eba651\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.208189 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e7056938-6e36-4631-adfb-01f116eba651-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"e7056938-6e36-4631-adfb-01f116eba651\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.208588 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e7056938-6e36-4631-adfb-01f116eba651-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"e7056938-6e36-4631-adfb-01f116eba651\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.208637 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e7056938-6e36-4631-adfb-01f116eba651-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"e7056938-6e36-4631-adfb-01f116eba651\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.208699 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7056938-6e36-4631-adfb-01f116eba651-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"e7056938-6e36-4631-adfb-01f116eba651\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.208753 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7056938-6e36-4631-adfb-01f116eba651-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"e7056938-6e36-4631-adfb-01f116eba651\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.208783 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dhvx\" (UniqueName: \"kubernetes.io/projected/e7056938-6e36-4631-adfb-01f116eba651-kube-api-access-2dhvx\") pod \"openstack-cell1-galera-0\" (UID: \"e7056938-6e36-4631-adfb-01f116eba651\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.208823 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7056938-6e36-4631-adfb-01f116eba651-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"e7056938-6e36-4631-adfb-01f116eba651\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.208857 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"e7056938-6e36-4631-adfb-01f116eba651\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.209287 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"e7056938-6e36-4631-adfb-01f116eba651\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/openstack-cell1-galera-0" Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.209842 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e7056938-6e36-4631-adfb-01f116eba651-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"e7056938-6e36-4631-adfb-01f116eba651\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.210470 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e7056938-6e36-4631-adfb-01f116eba651-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"e7056938-6e36-4631-adfb-01f116eba651\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.210723 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e7056938-6e36-4631-adfb-01f116eba651-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"e7056938-6e36-4631-adfb-01f116eba651\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.211900 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7056938-6e36-4631-adfb-01f116eba651-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"e7056938-6e36-4631-adfb-01f116eba651\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.217293 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7056938-6e36-4631-adfb-01f116eba651-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"e7056938-6e36-4631-adfb-01f116eba651\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.217566 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7056938-6e36-4631-adfb-01f116eba651-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"e7056938-6e36-4631-adfb-01f116eba651\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.233441 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dhvx\" (UniqueName: \"kubernetes.io/projected/e7056938-6e36-4631-adfb-01f116eba651-kube-api-access-2dhvx\") pod \"openstack-cell1-galera-0\" (UID: \"e7056938-6e36-4631-adfb-01f116eba651\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.234769 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"e7056938-6e36-4631-adfb-01f116eba651\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.327183 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.381680 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.383154 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.386362 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.386590 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-jvdrw" Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.387075 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.412208 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.514531 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d6c40a5-e2ff-4983-b9d8-fc483703870c-memcached-tls-certs\") pod \"memcached-0\" (UID: \"9d6c40a5-e2ff-4983-b9d8-fc483703870c\") " pod="openstack/memcached-0" Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.514605 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d6c40a5-e2ff-4983-b9d8-fc483703870c-combined-ca-bundle\") pod \"memcached-0\" (UID: \"9d6c40a5-e2ff-4983-b9d8-fc483703870c\") " pod="openstack/memcached-0" Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.514653 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5pxc\" (UniqueName: \"kubernetes.io/projected/9d6c40a5-e2ff-4983-b9d8-fc483703870c-kube-api-access-f5pxc\") pod \"memcached-0\" (UID: \"9d6c40a5-e2ff-4983-b9d8-fc483703870c\") " pod="openstack/memcached-0" Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.514727 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9d6c40a5-e2ff-4983-b9d8-fc483703870c-config-data\") pod \"memcached-0\" (UID: \"9d6c40a5-e2ff-4983-b9d8-fc483703870c\") " pod="openstack/memcached-0" Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.514749 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9d6c40a5-e2ff-4983-b9d8-fc483703870c-kolla-config\") pod \"memcached-0\" (UID: \"9d6c40a5-e2ff-4983-b9d8-fc483703870c\") " pod="openstack/memcached-0" Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.616453 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5pxc\" (UniqueName: \"kubernetes.io/projected/9d6c40a5-e2ff-4983-b9d8-fc483703870c-kube-api-access-f5pxc\") pod \"memcached-0\" (UID: \"9d6c40a5-e2ff-4983-b9d8-fc483703870c\") " pod="openstack/memcached-0" Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.616610 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9d6c40a5-e2ff-4983-b9d8-fc483703870c-config-data\") pod \"memcached-0\" (UID: \"9d6c40a5-e2ff-4983-b9d8-fc483703870c\") " pod="openstack/memcached-0" Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.616636 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9d6c40a5-e2ff-4983-b9d8-fc483703870c-kolla-config\") pod \"memcached-0\" (UID: \"9d6c40a5-e2ff-4983-b9d8-fc483703870c\") " pod="openstack/memcached-0" Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.616747 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d6c40a5-e2ff-4983-b9d8-fc483703870c-memcached-tls-certs\") pod \"memcached-0\" (UID: \"9d6c40a5-e2ff-4983-b9d8-fc483703870c\") " pod="openstack/memcached-0" Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.616778 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d6c40a5-e2ff-4983-b9d8-fc483703870c-combined-ca-bundle\") pod \"memcached-0\" (UID: \"9d6c40a5-e2ff-4983-b9d8-fc483703870c\") " pod="openstack/memcached-0" Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.618440 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9d6c40a5-e2ff-4983-b9d8-fc483703870c-kolla-config\") pod \"memcached-0\" (UID: \"9d6c40a5-e2ff-4983-b9d8-fc483703870c\") " pod="openstack/memcached-0" Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.618531 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9d6c40a5-e2ff-4983-b9d8-fc483703870c-config-data\") pod \"memcached-0\" (UID: \"9d6c40a5-e2ff-4983-b9d8-fc483703870c\") " pod="openstack/memcached-0" Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.623517 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d6c40a5-e2ff-4983-b9d8-fc483703870c-combined-ca-bundle\") pod \"memcached-0\" (UID: \"9d6c40a5-e2ff-4983-b9d8-fc483703870c\") " pod="openstack/memcached-0" Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.634029 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d6c40a5-e2ff-4983-b9d8-fc483703870c-memcached-tls-certs\") pod \"memcached-0\" (UID: \"9d6c40a5-e2ff-4983-b9d8-fc483703870c\") " pod="openstack/memcached-0" Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.641543 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5pxc\" (UniqueName: \"kubernetes.io/projected/9d6c40a5-e2ff-4983-b9d8-fc483703870c-kube-api-access-f5pxc\") pod \"memcached-0\" (UID: \"9d6c40a5-e2ff-4983-b9d8-fc483703870c\") " pod="openstack/memcached-0" Dec 06 06:14:37 crc kubenswrapper[4809]: I1206 06:14:37.707864 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 06 06:14:39 crc kubenswrapper[4809]: I1206 06:14:39.312460 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 06:14:39 crc kubenswrapper[4809]: I1206 06:14:39.315459 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 06 06:14:39 crc kubenswrapper[4809]: I1206 06:14:39.332978 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-cmv42" Dec 06 06:14:39 crc kubenswrapper[4809]: I1206 06:14:39.340188 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 06:14:39 crc kubenswrapper[4809]: I1206 06:14:39.456601 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85dpb\" (UniqueName: \"kubernetes.io/projected/cdf93098-585a-4391-8419-c5bc7ab1c567-kube-api-access-85dpb\") pod \"kube-state-metrics-0\" (UID: \"cdf93098-585a-4391-8419-c5bc7ab1c567\") " pod="openstack/kube-state-metrics-0" Dec 06 06:14:39 crc kubenswrapper[4809]: I1206 06:14:39.560365 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85dpb\" (UniqueName: \"kubernetes.io/projected/cdf93098-585a-4391-8419-c5bc7ab1c567-kube-api-access-85dpb\") pod \"kube-state-metrics-0\" (UID: \"cdf93098-585a-4391-8419-c5bc7ab1c567\") " pod="openstack/kube-state-metrics-0" Dec 06 06:14:39 crc kubenswrapper[4809]: I1206 06:14:39.590559 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85dpb\" (UniqueName: \"kubernetes.io/projected/cdf93098-585a-4391-8419-c5bc7ab1c567-kube-api-access-85dpb\") pod \"kube-state-metrics-0\" (UID: \"cdf93098-585a-4391-8419-c5bc7ab1c567\") " pod="openstack/kube-state-metrics-0" Dec 06 06:14:39 crc kubenswrapper[4809]: I1206 06:14:39.651126 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.026786 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-ui-dashboards-7d5fb4cbfb-sj6gs"] Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.029965 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-sj6gs" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.034479 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-ui-dashboards" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.038299 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-ui-dashboards-sa-dockercfg-h6dfs" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.042544 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-ui-dashboards-7d5fb4cbfb-sj6gs"] Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.070863 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85bqz\" (UniqueName: \"kubernetes.io/projected/6be1b144-2e51-42a1-9430-201ee08ce602-kube-api-access-85bqz\") pod \"observability-ui-dashboards-7d5fb4cbfb-sj6gs\" (UID: \"6be1b144-2e51-42a1-9430-201ee08ce602\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-sj6gs" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.070982 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6be1b144-2e51-42a1-9430-201ee08ce602-serving-cert\") pod \"observability-ui-dashboards-7d5fb4cbfb-sj6gs\" (UID: \"6be1b144-2e51-42a1-9430-201ee08ce602\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-sj6gs" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.173055 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85bqz\" (UniqueName: \"kubernetes.io/projected/6be1b144-2e51-42a1-9430-201ee08ce602-kube-api-access-85bqz\") pod \"observability-ui-dashboards-7d5fb4cbfb-sj6gs\" (UID: \"6be1b144-2e51-42a1-9430-201ee08ce602\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-sj6gs" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.173175 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6be1b144-2e51-42a1-9430-201ee08ce602-serving-cert\") pod \"observability-ui-dashboards-7d5fb4cbfb-sj6gs\" (UID: \"6be1b144-2e51-42a1-9430-201ee08ce602\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-sj6gs" Dec 06 06:14:40 crc kubenswrapper[4809]: E1206 06:14:40.173541 4809 secret.go:188] Couldn't get secret openshift-operators/observability-ui-dashboards: secret "observability-ui-dashboards" not found Dec 06 06:14:40 crc kubenswrapper[4809]: E1206 06:14:40.173654 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6be1b144-2e51-42a1-9430-201ee08ce602-serving-cert podName:6be1b144-2e51-42a1-9430-201ee08ce602 nodeName:}" failed. No retries permitted until 2025-12-06 06:14:40.673610429 +0000 UTC m=+1405.562593381 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/6be1b144-2e51-42a1-9430-201ee08ce602-serving-cert") pod "observability-ui-dashboards-7d5fb4cbfb-sj6gs" (UID: "6be1b144-2e51-42a1-9430-201ee08ce602") : secret "observability-ui-dashboards" not found Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.204645 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85bqz\" (UniqueName: \"kubernetes.io/projected/6be1b144-2e51-42a1-9430-201ee08ce602-kube-api-access-85bqz\") pod \"observability-ui-dashboards-7d5fb4cbfb-sj6gs\" (UID: \"6be1b144-2e51-42a1-9430-201ee08ce602\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-sj6gs" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.359974 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-5cccfc6d89-9j7zs"] Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.364800 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5cccfc6d89-9j7zs" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.371778 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5cccfc6d89-9j7zs"] Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.480994 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ac4de3b3-b463-4e5b-b935-ee6561519ddf-console-serving-cert\") pod \"console-5cccfc6d89-9j7zs\" (UID: \"ac4de3b3-b463-4e5b-b935-ee6561519ddf\") " pod="openshift-console/console-5cccfc6d89-9j7zs" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.481115 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ac4de3b3-b463-4e5b-b935-ee6561519ddf-oauth-serving-cert\") pod \"console-5cccfc6d89-9j7zs\" (UID: \"ac4de3b3-b463-4e5b-b935-ee6561519ddf\") " pod="openshift-console/console-5cccfc6d89-9j7zs" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.481570 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ac4de3b3-b463-4e5b-b935-ee6561519ddf-console-config\") pod \"console-5cccfc6d89-9j7zs\" (UID: \"ac4de3b3-b463-4e5b-b935-ee6561519ddf\") " pod="openshift-console/console-5cccfc6d89-9j7zs" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.481632 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ac4de3b3-b463-4e5b-b935-ee6561519ddf-console-oauth-config\") pod \"console-5cccfc6d89-9j7zs\" (UID: \"ac4de3b3-b463-4e5b-b935-ee6561519ddf\") " pod="openshift-console/console-5cccfc6d89-9j7zs" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.481650 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ac4de3b3-b463-4e5b-b935-ee6561519ddf-trusted-ca-bundle\") pod \"console-5cccfc6d89-9j7zs\" (UID: \"ac4de3b3-b463-4e5b-b935-ee6561519ddf\") " pod="openshift-console/console-5cccfc6d89-9j7zs" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.481713 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ac4de3b3-b463-4e5b-b935-ee6561519ddf-service-ca\") pod \"console-5cccfc6d89-9j7zs\" (UID: \"ac4de3b3-b463-4e5b-b935-ee6561519ddf\") " pod="openshift-console/console-5cccfc6d89-9j7zs" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.481748 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7bp9\" (UniqueName: \"kubernetes.io/projected/ac4de3b3-b463-4e5b-b935-ee6561519ddf-kube-api-access-t7bp9\") pod \"console-5cccfc6d89-9j7zs\" (UID: \"ac4de3b3-b463-4e5b-b935-ee6561519ddf\") " pod="openshift-console/console-5cccfc6d89-9j7zs" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.583432 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ac4de3b3-b463-4e5b-b935-ee6561519ddf-console-config\") pod \"console-5cccfc6d89-9j7zs\" (UID: \"ac4de3b3-b463-4e5b-b935-ee6561519ddf\") " pod="openshift-console/console-5cccfc6d89-9j7zs" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.583488 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ac4de3b3-b463-4e5b-b935-ee6561519ddf-console-oauth-config\") pod \"console-5cccfc6d89-9j7zs\" (UID: \"ac4de3b3-b463-4e5b-b935-ee6561519ddf\") " pod="openshift-console/console-5cccfc6d89-9j7zs" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.583504 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ac4de3b3-b463-4e5b-b935-ee6561519ddf-trusted-ca-bundle\") pod \"console-5cccfc6d89-9j7zs\" (UID: \"ac4de3b3-b463-4e5b-b935-ee6561519ddf\") " pod="openshift-console/console-5cccfc6d89-9j7zs" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.583526 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ac4de3b3-b463-4e5b-b935-ee6561519ddf-service-ca\") pod \"console-5cccfc6d89-9j7zs\" (UID: \"ac4de3b3-b463-4e5b-b935-ee6561519ddf\") " pod="openshift-console/console-5cccfc6d89-9j7zs" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.583552 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7bp9\" (UniqueName: \"kubernetes.io/projected/ac4de3b3-b463-4e5b-b935-ee6561519ddf-kube-api-access-t7bp9\") pod \"console-5cccfc6d89-9j7zs\" (UID: \"ac4de3b3-b463-4e5b-b935-ee6561519ddf\") " pod="openshift-console/console-5cccfc6d89-9j7zs" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.583676 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ac4de3b3-b463-4e5b-b935-ee6561519ddf-console-serving-cert\") pod \"console-5cccfc6d89-9j7zs\" (UID: \"ac4de3b3-b463-4e5b-b935-ee6561519ddf\") " pod="openshift-console/console-5cccfc6d89-9j7zs" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.583722 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ac4de3b3-b463-4e5b-b935-ee6561519ddf-oauth-serving-cert\") pod \"console-5cccfc6d89-9j7zs\" (UID: \"ac4de3b3-b463-4e5b-b935-ee6561519ddf\") " pod="openshift-console/console-5cccfc6d89-9j7zs" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.584891 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ac4de3b3-b463-4e5b-b935-ee6561519ddf-oauth-serving-cert\") pod \"console-5cccfc6d89-9j7zs\" (UID: \"ac4de3b3-b463-4e5b-b935-ee6561519ddf\") " pod="openshift-console/console-5cccfc6d89-9j7zs" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.585735 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ac4de3b3-b463-4e5b-b935-ee6561519ddf-console-config\") pod \"console-5cccfc6d89-9j7zs\" (UID: \"ac4de3b3-b463-4e5b-b935-ee6561519ddf\") " pod="openshift-console/console-5cccfc6d89-9j7zs" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.587483 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ac4de3b3-b463-4e5b-b935-ee6561519ddf-trusted-ca-bundle\") pod \"console-5cccfc6d89-9j7zs\" (UID: \"ac4de3b3-b463-4e5b-b935-ee6561519ddf\") " pod="openshift-console/console-5cccfc6d89-9j7zs" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.588144 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ac4de3b3-b463-4e5b-b935-ee6561519ddf-service-ca\") pod \"console-5cccfc6d89-9j7zs\" (UID: \"ac4de3b3-b463-4e5b-b935-ee6561519ddf\") " pod="openshift-console/console-5cccfc6d89-9j7zs" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.593928 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ac4de3b3-b463-4e5b-b935-ee6561519ddf-console-oauth-config\") pod \"console-5cccfc6d89-9j7zs\" (UID: \"ac4de3b3-b463-4e5b-b935-ee6561519ddf\") " pod="openshift-console/console-5cccfc6d89-9j7zs" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.601031 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ac4de3b3-b463-4e5b-b935-ee6561519ddf-console-serving-cert\") pod \"console-5cccfc6d89-9j7zs\" (UID: \"ac4de3b3-b463-4e5b-b935-ee6561519ddf\") " pod="openshift-console/console-5cccfc6d89-9j7zs" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.606965 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7bp9\" (UniqueName: \"kubernetes.io/projected/ac4de3b3-b463-4e5b-b935-ee6561519ddf-kube-api-access-t7bp9\") pod \"console-5cccfc6d89-9j7zs\" (UID: \"ac4de3b3-b463-4e5b-b935-ee6561519ddf\") " pod="openshift-console/console-5cccfc6d89-9j7zs" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.607831 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.612830 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.619381 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.619679 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.619830 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.620163 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-cx7x2" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.625338 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.637284 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.646794 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.685396 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6be1b144-2e51-42a1-9430-201ee08ce602-serving-cert\") pod \"observability-ui-dashboards-7d5fb4cbfb-sj6gs\" (UID: \"6be1b144-2e51-42a1-9430-201ee08ce602\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-sj6gs" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.685456 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f6536b25-1cc1-4f74-bf46-613965ddf185-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"f6536b25-1cc1-4f74-bf46-613965ddf185\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.686803 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f6536b25-1cc1-4f74-bf46-613965ddf185-config\") pod \"prometheus-metric-storage-0\" (UID: \"f6536b25-1cc1-4f74-bf46-613965ddf185\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.686858 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/f6536b25-1cc1-4f74-bf46-613965ddf185-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"f6536b25-1cc1-4f74-bf46-613965ddf185\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.686888 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"prometheus-metric-storage-0\" (UID: \"f6536b25-1cc1-4f74-bf46-613965ddf185\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.686907 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mz4q\" (UniqueName: \"kubernetes.io/projected/f6536b25-1cc1-4f74-bf46-613965ddf185-kube-api-access-8mz4q\") pod \"prometheus-metric-storage-0\" (UID: \"f6536b25-1cc1-4f74-bf46-613965ddf185\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.687018 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f6536b25-1cc1-4f74-bf46-613965ddf185-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"f6536b25-1cc1-4f74-bf46-613965ddf185\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.687062 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/f6536b25-1cc1-4f74-bf46-613965ddf185-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"f6536b25-1cc1-4f74-bf46-613965ddf185\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.687096 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f6536b25-1cc1-4f74-bf46-613965ddf185-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"f6536b25-1cc1-4f74-bf46-613965ddf185\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.695639 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6be1b144-2e51-42a1-9430-201ee08ce602-serving-cert\") pod \"observability-ui-dashboards-7d5fb4cbfb-sj6gs\" (UID: \"6be1b144-2e51-42a1-9430-201ee08ce602\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-sj6gs" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.726448 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5cccfc6d89-9j7zs" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.789376 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f6536b25-1cc1-4f74-bf46-613965ddf185-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"f6536b25-1cc1-4f74-bf46-613965ddf185\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.789436 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/f6536b25-1cc1-4f74-bf46-613965ddf185-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"f6536b25-1cc1-4f74-bf46-613965ddf185\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.790231 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f6536b25-1cc1-4f74-bf46-613965ddf185-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"f6536b25-1cc1-4f74-bf46-613965ddf185\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.790303 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f6536b25-1cc1-4f74-bf46-613965ddf185-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"f6536b25-1cc1-4f74-bf46-613965ddf185\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.790852 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f6536b25-1cc1-4f74-bf46-613965ddf185-config\") pod \"prometheus-metric-storage-0\" (UID: \"f6536b25-1cc1-4f74-bf46-613965ddf185\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.790920 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/f6536b25-1cc1-4f74-bf46-613965ddf185-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"f6536b25-1cc1-4f74-bf46-613965ddf185\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.790970 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"prometheus-metric-storage-0\" (UID: \"f6536b25-1cc1-4f74-bf46-613965ddf185\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.790991 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mz4q\" (UniqueName: \"kubernetes.io/projected/f6536b25-1cc1-4f74-bf46-613965ddf185-kube-api-access-8mz4q\") pod \"prometheus-metric-storage-0\" (UID: \"f6536b25-1cc1-4f74-bf46-613965ddf185\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.792305 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/f6536b25-1cc1-4f74-bf46-613965ddf185-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"f6536b25-1cc1-4f74-bf46-613965ddf185\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.792887 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f6536b25-1cc1-4f74-bf46-613965ddf185-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"f6536b25-1cc1-4f74-bf46-613965ddf185\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.793185 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f6536b25-1cc1-4f74-bf46-613965ddf185-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"f6536b25-1cc1-4f74-bf46-613965ddf185\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.793561 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/f6536b25-1cc1-4f74-bf46-613965ddf185-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"f6536b25-1cc1-4f74-bf46-613965ddf185\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.794198 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f6536b25-1cc1-4f74-bf46-613965ddf185-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"f6536b25-1cc1-4f74-bf46-613965ddf185\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.794366 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"prometheus-metric-storage-0\" (UID: \"f6536b25-1cc1-4f74-bf46-613965ddf185\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/prometheus-metric-storage-0" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.811753 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/f6536b25-1cc1-4f74-bf46-613965ddf185-config\") pod \"prometheus-metric-storage-0\" (UID: \"f6536b25-1cc1-4f74-bf46-613965ddf185\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.821622 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mz4q\" (UniqueName: \"kubernetes.io/projected/f6536b25-1cc1-4f74-bf46-613965ddf185-kube-api-access-8mz4q\") pod \"prometheus-metric-storage-0\" (UID: \"f6536b25-1cc1-4f74-bf46-613965ddf185\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.836643 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"prometheus-metric-storage-0\" (UID: \"f6536b25-1cc1-4f74-bf46-613965ddf185\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.969639 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-sj6gs" Dec 06 06:14:40 crc kubenswrapper[4809]: I1206 06:14:40.990743 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 06 06:14:42 crc kubenswrapper[4809]: I1206 06:14:42.452603 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1ceabc13-38c5-4425-8d85-c3afa8935243","Type":"ContainerStarted","Data":"d8209ae66e9765531af7b6a5fecfd56f9b7c3444a84bdbc8c222f43b3642712e"} Dec 06 06:14:42 crc kubenswrapper[4809]: I1206 06:14:42.952312 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-xhlzm"] Dec 06 06:14:42 crc kubenswrapper[4809]: I1206 06:14:42.954053 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xhlzm" Dec 06 06:14:42 crc kubenswrapper[4809]: I1206 06:14:42.960642 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-fdthh" Dec 06 06:14:42 crc kubenswrapper[4809]: I1206 06:14:42.961517 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 06 06:14:42 crc kubenswrapper[4809]: I1206 06:14:42.961657 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 06 06:14:42 crc kubenswrapper[4809]: I1206 06:14:42.976191 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-xhlzm"] Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.009203 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-jgqjw"] Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.011498 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-jgqjw" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.042021 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-jgqjw"] Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.056226 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/92ce9234-d708-449e-9874-97ac8a4b994c-var-run\") pod \"ovn-controller-ovs-jgqjw\" (UID: \"92ce9234-d708-449e-9874-97ac8a4b994c\") " pod="openstack/ovn-controller-ovs-jgqjw" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.056537 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6b64391f-74d7-4b69-a898-f7375057c6d5-var-log-ovn\") pod \"ovn-controller-xhlzm\" (UID: \"6b64391f-74d7-4b69-a898-f7375057c6d5\") " pod="openstack/ovn-controller-xhlzm" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.056692 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6b64391f-74d7-4b69-a898-f7375057c6d5-var-run\") pod \"ovn-controller-xhlzm\" (UID: \"6b64391f-74d7-4b69-a898-f7375057c6d5\") " pod="openstack/ovn-controller-xhlzm" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.056988 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/92ce9234-d708-449e-9874-97ac8a4b994c-var-lib\") pod \"ovn-controller-ovs-jgqjw\" (UID: \"92ce9234-d708-449e-9874-97ac8a4b994c\") " pod="openstack/ovn-controller-ovs-jgqjw" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.057141 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b64391f-74d7-4b69-a898-f7375057c6d5-ovn-controller-tls-certs\") pod \"ovn-controller-xhlzm\" (UID: \"6b64391f-74d7-4b69-a898-f7375057c6d5\") " pod="openstack/ovn-controller-xhlzm" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.057244 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/92ce9234-d708-449e-9874-97ac8a4b994c-var-log\") pod \"ovn-controller-ovs-jgqjw\" (UID: \"92ce9234-d708-449e-9874-97ac8a4b994c\") " pod="openstack/ovn-controller-ovs-jgqjw" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.057335 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lv72g\" (UniqueName: \"kubernetes.io/projected/6b64391f-74d7-4b69-a898-f7375057c6d5-kube-api-access-lv72g\") pod \"ovn-controller-xhlzm\" (UID: \"6b64391f-74d7-4b69-a898-f7375057c6d5\") " pod="openstack/ovn-controller-xhlzm" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.057413 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/92ce9234-d708-449e-9874-97ac8a4b994c-scripts\") pod \"ovn-controller-ovs-jgqjw\" (UID: \"92ce9234-d708-449e-9874-97ac8a4b994c\") " pod="openstack/ovn-controller-ovs-jgqjw" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.057604 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6b64391f-74d7-4b69-a898-f7375057c6d5-var-run-ovn\") pod \"ovn-controller-xhlzm\" (UID: \"6b64391f-74d7-4b69-a898-f7375057c6d5\") " pod="openstack/ovn-controller-xhlzm" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.057694 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xw9rv\" (UniqueName: \"kubernetes.io/projected/92ce9234-d708-449e-9874-97ac8a4b994c-kube-api-access-xw9rv\") pod \"ovn-controller-ovs-jgqjw\" (UID: \"92ce9234-d708-449e-9874-97ac8a4b994c\") " pod="openstack/ovn-controller-ovs-jgqjw" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.057858 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6b64391f-74d7-4b69-a898-f7375057c6d5-scripts\") pod \"ovn-controller-xhlzm\" (UID: \"6b64391f-74d7-4b69-a898-f7375057c6d5\") " pod="openstack/ovn-controller-xhlzm" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.057949 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/92ce9234-d708-449e-9874-97ac8a4b994c-etc-ovs\") pod \"ovn-controller-ovs-jgqjw\" (UID: \"92ce9234-d708-449e-9874-97ac8a4b994c\") " pod="openstack/ovn-controller-ovs-jgqjw" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.058028 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b64391f-74d7-4b69-a898-f7375057c6d5-combined-ca-bundle\") pod \"ovn-controller-xhlzm\" (UID: \"6b64391f-74d7-4b69-a898-f7375057c6d5\") " pod="openstack/ovn-controller-xhlzm" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.159709 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6b64391f-74d7-4b69-a898-f7375057c6d5-scripts\") pod \"ovn-controller-xhlzm\" (UID: \"6b64391f-74d7-4b69-a898-f7375057c6d5\") " pod="openstack/ovn-controller-xhlzm" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.159765 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b64391f-74d7-4b69-a898-f7375057c6d5-combined-ca-bundle\") pod \"ovn-controller-xhlzm\" (UID: \"6b64391f-74d7-4b69-a898-f7375057c6d5\") " pod="openstack/ovn-controller-xhlzm" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.159792 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/92ce9234-d708-449e-9874-97ac8a4b994c-etc-ovs\") pod \"ovn-controller-ovs-jgqjw\" (UID: \"92ce9234-d708-449e-9874-97ac8a4b994c\") " pod="openstack/ovn-controller-ovs-jgqjw" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.159825 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/92ce9234-d708-449e-9874-97ac8a4b994c-var-run\") pod \"ovn-controller-ovs-jgqjw\" (UID: \"92ce9234-d708-449e-9874-97ac8a4b994c\") " pod="openstack/ovn-controller-ovs-jgqjw" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.159850 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6b64391f-74d7-4b69-a898-f7375057c6d5-var-log-ovn\") pod \"ovn-controller-xhlzm\" (UID: \"6b64391f-74d7-4b69-a898-f7375057c6d5\") " pod="openstack/ovn-controller-xhlzm" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.159891 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6b64391f-74d7-4b69-a898-f7375057c6d5-var-run\") pod \"ovn-controller-xhlzm\" (UID: \"6b64391f-74d7-4b69-a898-f7375057c6d5\") " pod="openstack/ovn-controller-xhlzm" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.160016 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/92ce9234-d708-449e-9874-97ac8a4b994c-var-lib\") pod \"ovn-controller-ovs-jgqjw\" (UID: \"92ce9234-d708-449e-9874-97ac8a4b994c\") " pod="openstack/ovn-controller-ovs-jgqjw" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.160071 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b64391f-74d7-4b69-a898-f7375057c6d5-ovn-controller-tls-certs\") pod \"ovn-controller-xhlzm\" (UID: \"6b64391f-74d7-4b69-a898-f7375057c6d5\") " pod="openstack/ovn-controller-xhlzm" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.160108 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/92ce9234-d708-449e-9874-97ac8a4b994c-var-log\") pod \"ovn-controller-ovs-jgqjw\" (UID: \"92ce9234-d708-449e-9874-97ac8a4b994c\") " pod="openstack/ovn-controller-ovs-jgqjw" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.160136 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lv72g\" (UniqueName: \"kubernetes.io/projected/6b64391f-74d7-4b69-a898-f7375057c6d5-kube-api-access-lv72g\") pod \"ovn-controller-xhlzm\" (UID: \"6b64391f-74d7-4b69-a898-f7375057c6d5\") " pod="openstack/ovn-controller-xhlzm" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.160164 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/92ce9234-d708-449e-9874-97ac8a4b994c-scripts\") pod \"ovn-controller-ovs-jgqjw\" (UID: \"92ce9234-d708-449e-9874-97ac8a4b994c\") " pod="openstack/ovn-controller-ovs-jgqjw" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.160204 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6b64391f-74d7-4b69-a898-f7375057c6d5-var-run-ovn\") pod \"ovn-controller-xhlzm\" (UID: \"6b64391f-74d7-4b69-a898-f7375057c6d5\") " pod="openstack/ovn-controller-xhlzm" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.160227 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xw9rv\" (UniqueName: \"kubernetes.io/projected/92ce9234-d708-449e-9874-97ac8a4b994c-kube-api-access-xw9rv\") pod \"ovn-controller-ovs-jgqjw\" (UID: \"92ce9234-d708-449e-9874-97ac8a4b994c\") " pod="openstack/ovn-controller-ovs-jgqjw" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.161042 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/92ce9234-d708-449e-9874-97ac8a4b994c-var-lib\") pod \"ovn-controller-ovs-jgqjw\" (UID: \"92ce9234-d708-449e-9874-97ac8a4b994c\") " pod="openstack/ovn-controller-ovs-jgqjw" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.161209 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/92ce9234-d708-449e-9874-97ac8a4b994c-var-run\") pod \"ovn-controller-ovs-jgqjw\" (UID: \"92ce9234-d708-449e-9874-97ac8a4b994c\") " pod="openstack/ovn-controller-ovs-jgqjw" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.161342 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/92ce9234-d708-449e-9874-97ac8a4b994c-var-log\") pod \"ovn-controller-ovs-jgqjw\" (UID: \"92ce9234-d708-449e-9874-97ac8a4b994c\") " pod="openstack/ovn-controller-ovs-jgqjw" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.161407 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6b64391f-74d7-4b69-a898-f7375057c6d5-var-log-ovn\") pod \"ovn-controller-xhlzm\" (UID: \"6b64391f-74d7-4b69-a898-f7375057c6d5\") " pod="openstack/ovn-controller-xhlzm" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.161440 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/92ce9234-d708-449e-9874-97ac8a4b994c-etc-ovs\") pod \"ovn-controller-ovs-jgqjw\" (UID: \"92ce9234-d708-449e-9874-97ac8a4b994c\") " pod="openstack/ovn-controller-ovs-jgqjw" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.161460 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6b64391f-74d7-4b69-a898-f7375057c6d5-var-run-ovn\") pod \"ovn-controller-xhlzm\" (UID: \"6b64391f-74d7-4b69-a898-f7375057c6d5\") " pod="openstack/ovn-controller-xhlzm" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.161492 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6b64391f-74d7-4b69-a898-f7375057c6d5-var-run\") pod \"ovn-controller-xhlzm\" (UID: \"6b64391f-74d7-4b69-a898-f7375057c6d5\") " pod="openstack/ovn-controller-xhlzm" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.162097 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6b64391f-74d7-4b69-a898-f7375057c6d5-scripts\") pod \"ovn-controller-xhlzm\" (UID: \"6b64391f-74d7-4b69-a898-f7375057c6d5\") " pod="openstack/ovn-controller-xhlzm" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.163225 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/92ce9234-d708-449e-9874-97ac8a4b994c-scripts\") pod \"ovn-controller-ovs-jgqjw\" (UID: \"92ce9234-d708-449e-9874-97ac8a4b994c\") " pod="openstack/ovn-controller-ovs-jgqjw" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.176374 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b64391f-74d7-4b69-a898-f7375057c6d5-ovn-controller-tls-certs\") pod \"ovn-controller-xhlzm\" (UID: \"6b64391f-74d7-4b69-a898-f7375057c6d5\") " pod="openstack/ovn-controller-xhlzm" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.181041 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lv72g\" (UniqueName: \"kubernetes.io/projected/6b64391f-74d7-4b69-a898-f7375057c6d5-kube-api-access-lv72g\") pod \"ovn-controller-xhlzm\" (UID: \"6b64391f-74d7-4b69-a898-f7375057c6d5\") " pod="openstack/ovn-controller-xhlzm" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.181070 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xw9rv\" (UniqueName: \"kubernetes.io/projected/92ce9234-d708-449e-9874-97ac8a4b994c-kube-api-access-xw9rv\") pod \"ovn-controller-ovs-jgqjw\" (UID: \"92ce9234-d708-449e-9874-97ac8a4b994c\") " pod="openstack/ovn-controller-ovs-jgqjw" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.182665 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b64391f-74d7-4b69-a898-f7375057c6d5-combined-ca-bundle\") pod \"ovn-controller-xhlzm\" (UID: \"6b64391f-74d7-4b69-a898-f7375057c6d5\") " pod="openstack/ovn-controller-xhlzm" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.297537 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xhlzm" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.328353 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-jgqjw" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.805771 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.808062 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.836716 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.836795 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.836812 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.836723 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-56k6j" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.837142 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.859857 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.994387 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/932f2f72-a070-425a-b065-64e293954cc5-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"932f2f72-a070-425a-b065-64e293954cc5\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.994515 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/932f2f72-a070-425a-b065-64e293954cc5-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"932f2f72-a070-425a-b065-64e293954cc5\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.994567 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"932f2f72-a070-425a-b065-64e293954cc5\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.994588 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/932f2f72-a070-425a-b065-64e293954cc5-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"932f2f72-a070-425a-b065-64e293954cc5\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.994611 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75pwn\" (UniqueName: \"kubernetes.io/projected/932f2f72-a070-425a-b065-64e293954cc5-kube-api-access-75pwn\") pod \"ovsdbserver-nb-0\" (UID: \"932f2f72-a070-425a-b065-64e293954cc5\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.994650 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/932f2f72-a070-425a-b065-64e293954cc5-config\") pod \"ovsdbserver-nb-0\" (UID: \"932f2f72-a070-425a-b065-64e293954cc5\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.994698 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/932f2f72-a070-425a-b065-64e293954cc5-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"932f2f72-a070-425a-b065-64e293954cc5\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:14:43 crc kubenswrapper[4809]: I1206 06:14:43.994743 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/932f2f72-a070-425a-b065-64e293954cc5-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"932f2f72-a070-425a-b065-64e293954cc5\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:14:44 crc kubenswrapper[4809]: I1206 06:14:44.097246 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/932f2f72-a070-425a-b065-64e293954cc5-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"932f2f72-a070-425a-b065-64e293954cc5\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:14:44 crc kubenswrapper[4809]: I1206 06:14:44.097329 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"932f2f72-a070-425a-b065-64e293954cc5\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:14:44 crc kubenswrapper[4809]: I1206 06:14:44.097351 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/932f2f72-a070-425a-b065-64e293954cc5-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"932f2f72-a070-425a-b065-64e293954cc5\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:14:44 crc kubenswrapper[4809]: I1206 06:14:44.097372 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75pwn\" (UniqueName: \"kubernetes.io/projected/932f2f72-a070-425a-b065-64e293954cc5-kube-api-access-75pwn\") pod \"ovsdbserver-nb-0\" (UID: \"932f2f72-a070-425a-b065-64e293954cc5\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:14:44 crc kubenswrapper[4809]: I1206 06:14:44.097398 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/932f2f72-a070-425a-b065-64e293954cc5-config\") pod \"ovsdbserver-nb-0\" (UID: \"932f2f72-a070-425a-b065-64e293954cc5\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:14:44 crc kubenswrapper[4809]: I1206 06:14:44.097446 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/932f2f72-a070-425a-b065-64e293954cc5-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"932f2f72-a070-425a-b065-64e293954cc5\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:14:44 crc kubenswrapper[4809]: I1206 06:14:44.097476 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/932f2f72-a070-425a-b065-64e293954cc5-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"932f2f72-a070-425a-b065-64e293954cc5\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:14:44 crc kubenswrapper[4809]: I1206 06:14:44.098617 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/932f2f72-a070-425a-b065-64e293954cc5-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"932f2f72-a070-425a-b065-64e293954cc5\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:14:44 crc kubenswrapper[4809]: I1206 06:14:44.098701 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/932f2f72-a070-425a-b065-64e293954cc5-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"932f2f72-a070-425a-b065-64e293954cc5\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:14:44 crc kubenswrapper[4809]: I1206 06:14:44.098813 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/932f2f72-a070-425a-b065-64e293954cc5-config\") pod \"ovsdbserver-nb-0\" (UID: \"932f2f72-a070-425a-b065-64e293954cc5\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:14:44 crc kubenswrapper[4809]: I1206 06:14:44.099277 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/932f2f72-a070-425a-b065-64e293954cc5-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"932f2f72-a070-425a-b065-64e293954cc5\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:14:44 crc kubenswrapper[4809]: I1206 06:14:44.100467 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"932f2f72-a070-425a-b065-64e293954cc5\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/ovsdbserver-nb-0" Dec 06 06:14:44 crc kubenswrapper[4809]: I1206 06:14:44.105030 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/932f2f72-a070-425a-b065-64e293954cc5-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"932f2f72-a070-425a-b065-64e293954cc5\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:14:44 crc kubenswrapper[4809]: I1206 06:14:44.110761 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/932f2f72-a070-425a-b065-64e293954cc5-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"932f2f72-a070-425a-b065-64e293954cc5\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:14:44 crc kubenswrapper[4809]: I1206 06:14:44.111224 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/932f2f72-a070-425a-b065-64e293954cc5-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"932f2f72-a070-425a-b065-64e293954cc5\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:14:44 crc kubenswrapper[4809]: I1206 06:14:44.123569 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75pwn\" (UniqueName: \"kubernetes.io/projected/932f2f72-a070-425a-b065-64e293954cc5-kube-api-access-75pwn\") pod \"ovsdbserver-nb-0\" (UID: \"932f2f72-a070-425a-b065-64e293954cc5\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:14:44 crc kubenswrapper[4809]: I1206 06:14:44.145283 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"932f2f72-a070-425a-b065-64e293954cc5\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:14:44 crc kubenswrapper[4809]: I1206 06:14:44.158254 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 06 06:14:47 crc kubenswrapper[4809]: I1206 06:14:47.040476 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 06 06:14:47 crc kubenswrapper[4809]: I1206 06:14:47.043991 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 06 06:14:47 crc kubenswrapper[4809]: I1206 06:14:47.047438 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 06 06:14:47 crc kubenswrapper[4809]: I1206 06:14:47.047620 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 06 06:14:47 crc kubenswrapper[4809]: I1206 06:14:47.047678 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-f6cvf" Dec 06 06:14:47 crc kubenswrapper[4809]: I1206 06:14:47.047864 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 06 06:14:47 crc kubenswrapper[4809]: I1206 06:14:47.065311 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 06 06:14:47 crc kubenswrapper[4809]: I1206 06:14:47.105431 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cdxx\" (UniqueName: \"kubernetes.io/projected/1ffa5f54-0c40-45a6-9ebb-a5eca04810ee-kube-api-access-9cdxx\") pod \"ovsdbserver-sb-0\" (UID: \"1ffa5f54-0c40-45a6-9ebb-a5eca04810ee\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:14:47 crc kubenswrapper[4809]: I1206 06:14:47.105535 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ffa5f54-0c40-45a6-9ebb-a5eca04810ee-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1ffa5f54-0c40-45a6-9ebb-a5eca04810ee\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:14:47 crc kubenswrapper[4809]: I1206 06:14:47.105570 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ffa5f54-0c40-45a6-9ebb-a5eca04810ee-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"1ffa5f54-0c40-45a6-9ebb-a5eca04810ee\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:14:47 crc kubenswrapper[4809]: I1206 06:14:47.105603 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1ffa5f54-0c40-45a6-9ebb-a5eca04810ee-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"1ffa5f54-0c40-45a6-9ebb-a5eca04810ee\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:14:47 crc kubenswrapper[4809]: I1206 06:14:47.105906 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ffa5f54-0c40-45a6-9ebb-a5eca04810ee-config\") pod \"ovsdbserver-sb-0\" (UID: \"1ffa5f54-0c40-45a6-9ebb-a5eca04810ee\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:14:47 crc kubenswrapper[4809]: I1206 06:14:47.106129 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ffa5f54-0c40-45a6-9ebb-a5eca04810ee-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1ffa5f54-0c40-45a6-9ebb-a5eca04810ee\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:14:47 crc kubenswrapper[4809]: I1206 06:14:47.106228 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"1ffa5f54-0c40-45a6-9ebb-a5eca04810ee\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:14:47 crc kubenswrapper[4809]: I1206 06:14:47.106324 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ffa5f54-0c40-45a6-9ebb-a5eca04810ee-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"1ffa5f54-0c40-45a6-9ebb-a5eca04810ee\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:14:47 crc kubenswrapper[4809]: I1206 06:14:47.208966 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cdxx\" (UniqueName: \"kubernetes.io/projected/1ffa5f54-0c40-45a6-9ebb-a5eca04810ee-kube-api-access-9cdxx\") pod \"ovsdbserver-sb-0\" (UID: \"1ffa5f54-0c40-45a6-9ebb-a5eca04810ee\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:14:47 crc kubenswrapper[4809]: I1206 06:14:47.209036 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ffa5f54-0c40-45a6-9ebb-a5eca04810ee-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1ffa5f54-0c40-45a6-9ebb-a5eca04810ee\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:14:47 crc kubenswrapper[4809]: I1206 06:14:47.209062 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ffa5f54-0c40-45a6-9ebb-a5eca04810ee-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"1ffa5f54-0c40-45a6-9ebb-a5eca04810ee\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:14:47 crc kubenswrapper[4809]: I1206 06:14:47.209098 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1ffa5f54-0c40-45a6-9ebb-a5eca04810ee-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"1ffa5f54-0c40-45a6-9ebb-a5eca04810ee\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:14:47 crc kubenswrapper[4809]: I1206 06:14:47.209224 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ffa5f54-0c40-45a6-9ebb-a5eca04810ee-config\") pod \"ovsdbserver-sb-0\" (UID: \"1ffa5f54-0c40-45a6-9ebb-a5eca04810ee\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:14:47 crc kubenswrapper[4809]: I1206 06:14:47.209276 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ffa5f54-0c40-45a6-9ebb-a5eca04810ee-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1ffa5f54-0c40-45a6-9ebb-a5eca04810ee\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:14:47 crc kubenswrapper[4809]: I1206 06:14:47.209309 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"1ffa5f54-0c40-45a6-9ebb-a5eca04810ee\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:14:47 crc kubenswrapper[4809]: I1206 06:14:47.209352 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ffa5f54-0c40-45a6-9ebb-a5eca04810ee-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"1ffa5f54-0c40-45a6-9ebb-a5eca04810ee\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:14:47 crc kubenswrapper[4809]: I1206 06:14:47.210592 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ffa5f54-0c40-45a6-9ebb-a5eca04810ee-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"1ffa5f54-0c40-45a6-9ebb-a5eca04810ee\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:14:47 crc kubenswrapper[4809]: I1206 06:14:47.216172 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ffa5f54-0c40-45a6-9ebb-a5eca04810ee-config\") pod \"ovsdbserver-sb-0\" (UID: \"1ffa5f54-0c40-45a6-9ebb-a5eca04810ee\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:14:47 crc kubenswrapper[4809]: I1206 06:14:47.216521 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"1ffa5f54-0c40-45a6-9ebb-a5eca04810ee\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/ovsdbserver-sb-0" Dec 06 06:14:47 crc kubenswrapper[4809]: I1206 06:14:47.218506 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1ffa5f54-0c40-45a6-9ebb-a5eca04810ee-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"1ffa5f54-0c40-45a6-9ebb-a5eca04810ee\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:14:47 crc kubenswrapper[4809]: I1206 06:14:47.220903 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ffa5f54-0c40-45a6-9ebb-a5eca04810ee-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"1ffa5f54-0c40-45a6-9ebb-a5eca04810ee\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:14:47 crc kubenswrapper[4809]: I1206 06:14:47.221499 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ffa5f54-0c40-45a6-9ebb-a5eca04810ee-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1ffa5f54-0c40-45a6-9ebb-a5eca04810ee\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:14:47 crc kubenswrapper[4809]: I1206 06:14:47.224652 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ffa5f54-0c40-45a6-9ebb-a5eca04810ee-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1ffa5f54-0c40-45a6-9ebb-a5eca04810ee\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:14:47 crc kubenswrapper[4809]: I1206 06:14:47.234489 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cdxx\" (UniqueName: \"kubernetes.io/projected/1ffa5f54-0c40-45a6-9ebb-a5eca04810ee-kube-api-access-9cdxx\") pod \"ovsdbserver-sb-0\" (UID: \"1ffa5f54-0c40-45a6-9ebb-a5eca04810ee\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:14:47 crc kubenswrapper[4809]: I1206 06:14:47.271397 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"1ffa5f54-0c40-45a6-9ebb-a5eca04810ee\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:14:47 crc kubenswrapper[4809]: I1206 06:14:47.380815 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 06 06:14:47 crc kubenswrapper[4809]: I1206 06:14:47.593168 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 06 06:14:54 crc kubenswrapper[4809]: W1206 06:14:54.484946 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode7056938_6e36_4631_adfb_01f116eba651.slice/crio-90d5b7ca7fa1d0864e104a96e568666a53541441aab63b197254ed4a3fb9bac5 WatchSource:0}: Error finding container 90d5b7ca7fa1d0864e104a96e568666a53541441aab63b197254ed4a3fb9bac5: Status 404 returned error can't find the container with id 90d5b7ca7fa1d0864e104a96e568666a53541441aab63b197254ed4a3fb9bac5 Dec 06 06:14:54 crc kubenswrapper[4809]: I1206 06:14:54.601346 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e7056938-6e36-4631-adfb-01f116eba651","Type":"ContainerStarted","Data":"90d5b7ca7fa1d0864e104a96e568666a53541441aab63b197254ed4a3fb9bac5"} Dec 06 06:14:54 crc kubenswrapper[4809]: I1206 06:14:54.980260 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 06 06:14:56 crc kubenswrapper[4809]: E1206 06:14:56.299655 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 06 06:14:56 crc kubenswrapper[4809]: E1206 06:14:56.300429 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tr9lq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-xm4sd_openstack(68b27fd9-c5ea-4e7a-9639-afb6adecf768): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 06:14:56 crc kubenswrapper[4809]: E1206 06:14:56.302173 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-xm4sd" podUID="68b27fd9-c5ea-4e7a-9639-afb6adecf768" Dec 06 06:14:56 crc kubenswrapper[4809]: E1206 06:14:56.304402 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 06 06:14:56 crc kubenswrapper[4809]: E1206 06:14:56.305984 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b5lhx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-9bft5_openstack(78537fe2-3d4b-4659-82fc-1c8fa02d0f66): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 06:14:56 crc kubenswrapper[4809]: E1206 06:14:56.307131 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-9bft5" podUID="78537fe2-3d4b-4659-82fc-1c8fa02d0f66" Dec 06 06:14:56 crc kubenswrapper[4809]: I1206 06:14:56.673442 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"7c596e14-b96f-4e5f-a8ac-3dc66fa0276d","Type":"ContainerStarted","Data":"b010c5208c8e1889ef7cfce2031aaad74f565f3c1d108dd47054d87fdc2ad434"} Dec 06 06:14:56 crc kubenswrapper[4809]: I1206 06:14:56.866742 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 06 06:14:57 crc kubenswrapper[4809]: W1206 06:14:57.118598 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcdf93098_585a_4391_8419_c5bc7ab1c567.slice/crio-66f5db3f29fbb91d8689d12339457c220d225d60ff1721cc5985909e88e6d5d1 WatchSource:0}: Error finding container 66f5db3f29fbb91d8689d12339457c220d225d60ff1721cc5985909e88e6d5d1: Status 404 returned error can't find the container with id 66f5db3f29fbb91d8689d12339457c220d225d60ff1721cc5985909e88e6d5d1 Dec 06 06:14:57 crc kubenswrapper[4809]: I1206 06:14:57.127395 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 06:14:57 crc kubenswrapper[4809]: I1206 06:14:57.591427 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-xm4sd" Dec 06 06:14:57 crc kubenswrapper[4809]: I1206 06:14:57.687212 4809 generic.go:334] "Generic (PLEG): container finished" podID="de481cc6-bc8d-4a07-991b-72cad738f1d5" containerID="aafb8c934a2e51c587c3669678f2f7e847e52f557a0d98bf2c575830dbb87a35" exitCode=0 Dec 06 06:14:57 crc kubenswrapper[4809]: I1206 06:14:57.687282 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-k6s2d" event={"ID":"de481cc6-bc8d-4a07-991b-72cad738f1d5","Type":"ContainerDied","Data":"aafb8c934a2e51c587c3669678f2f7e847e52f557a0d98bf2c575830dbb87a35"} Dec 06 06:14:57 crc kubenswrapper[4809]: I1206 06:14:57.691009 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-xm4sd" Dec 06 06:14:57 crc kubenswrapper[4809]: I1206 06:14:57.691322 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-xm4sd" event={"ID":"68b27fd9-c5ea-4e7a-9639-afb6adecf768","Type":"ContainerDied","Data":"7a4833bc6acf76d02c3fae3b9da7758e13cd8e7e8048b3b234324c9fc01d7359"} Dec 06 06:14:57 crc kubenswrapper[4809]: I1206 06:14:57.692127 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tr9lq\" (UniqueName: \"kubernetes.io/projected/68b27fd9-c5ea-4e7a-9639-afb6adecf768-kube-api-access-tr9lq\") pod \"68b27fd9-c5ea-4e7a-9639-afb6adecf768\" (UID: \"68b27fd9-c5ea-4e7a-9639-afb6adecf768\") " Dec 06 06:14:57 crc kubenswrapper[4809]: I1206 06:14:57.692343 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68b27fd9-c5ea-4e7a-9639-afb6adecf768-config\") pod \"68b27fd9-c5ea-4e7a-9639-afb6adecf768\" (UID: \"68b27fd9-c5ea-4e7a-9639-afb6adecf768\") " Dec 06 06:14:57 crc kubenswrapper[4809]: I1206 06:14:57.693459 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68b27fd9-c5ea-4e7a-9639-afb6adecf768-config" (OuterVolumeSpecName: "config") pod "68b27fd9-c5ea-4e7a-9639-afb6adecf768" (UID: "68b27fd9-c5ea-4e7a-9639-afb6adecf768"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:14:57 crc kubenswrapper[4809]: I1206 06:14:57.707085 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"cdf93098-585a-4391-8419-c5bc7ab1c567","Type":"ContainerStarted","Data":"66f5db3f29fbb91d8689d12339457c220d225d60ff1721cc5985909e88e6d5d1"} Dec 06 06:14:57 crc kubenswrapper[4809]: I1206 06:14:57.707196 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68b27fd9-c5ea-4e7a-9639-afb6adecf768-kube-api-access-tr9lq" (OuterVolumeSpecName: "kube-api-access-tr9lq") pod "68b27fd9-c5ea-4e7a-9639-afb6adecf768" (UID: "68b27fd9-c5ea-4e7a-9639-afb6adecf768"). InnerVolumeSpecName "kube-api-access-tr9lq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:14:57 crc kubenswrapper[4809]: I1206 06:14:57.713199 4809 generic.go:334] "Generic (PLEG): container finished" podID="a8bfcbfb-3946-420c-b90f-f390887145f0" containerID="69cad62602dd98f3cd2dbd728a1ff4fcdce7c03ce1e2531ef7756a170e75fa5a" exitCode=0 Dec 06 06:14:57 crc kubenswrapper[4809]: I1206 06:14:57.713343 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-gr99p" event={"ID":"a8bfcbfb-3946-420c-b90f-f390887145f0","Type":"ContainerDied","Data":"69cad62602dd98f3cd2dbd728a1ff4fcdce7c03ce1e2531ef7756a170e75fa5a"} Dec 06 06:14:57 crc kubenswrapper[4809]: I1206 06:14:57.716011 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"9d6c40a5-e2ff-4983-b9d8-fc483703870c","Type":"ContainerStarted","Data":"5d67516d4802a93c2fb1b813c999681593890bd4bb0c92046a1af25b7e99a03d"} Dec 06 06:14:57 crc kubenswrapper[4809]: I1206 06:14:57.795792 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68b27fd9-c5ea-4e7a-9639-afb6adecf768-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:14:57 crc kubenswrapper[4809]: I1206 06:14:57.796217 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tr9lq\" (UniqueName: \"kubernetes.io/projected/68b27fd9-c5ea-4e7a-9639-afb6adecf768-kube-api-access-tr9lq\") on node \"crc\" DevicePath \"\"" Dec 06 06:14:58 crc kubenswrapper[4809]: I1206 06:14:58.047494 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-ui-dashboards-7d5fb4cbfb-sj6gs"] Dec 06 06:14:58 crc kubenswrapper[4809]: I1206 06:14:58.059985 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5cccfc6d89-9j7zs"] Dec 06 06:14:58 crc kubenswrapper[4809]: I1206 06:14:58.077055 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 06 06:14:58 crc kubenswrapper[4809]: I1206 06:14:58.164043 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-xm4sd"] Dec 06 06:14:58 crc kubenswrapper[4809]: I1206 06:14:58.175880 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-xm4sd"] Dec 06 06:14:58 crc kubenswrapper[4809]: I1206 06:14:58.371381 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-xhlzm"] Dec 06 06:14:58 crc kubenswrapper[4809]: I1206 06:14:58.421704 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-jgqjw"] Dec 06 06:14:58 crc kubenswrapper[4809]: W1206 06:14:58.474069 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6be1b144_2e51_42a1_9430_201ee08ce602.slice/crio-16d0b81c7785ae188c9c6b120394bec80173aeb7b9c10e720e99c5f6d1b4de19 WatchSource:0}: Error finding container 16d0b81c7785ae188c9c6b120394bec80173aeb7b9c10e720e99c5f6d1b4de19: Status 404 returned error can't find the container with id 16d0b81c7785ae188c9c6b120394bec80173aeb7b9c10e720e99c5f6d1b4de19 Dec 06 06:14:58 crc kubenswrapper[4809]: I1206 06:14:58.512097 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 06 06:14:58 crc kubenswrapper[4809]: I1206 06:14:58.758333 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-gr99p" event={"ID":"a8bfcbfb-3946-420c-b90f-f390887145f0","Type":"ContainerStarted","Data":"e4f1e32a64e611e44984494ef945ec0948d0c06be22b8dd98b0210df10ed1d23"} Dec 06 06:14:58 crc kubenswrapper[4809]: I1206 06:14:58.758721 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-gr99p" Dec 06 06:14:58 crc kubenswrapper[4809]: I1206 06:14:58.762802 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1ceabc13-38c5-4425-8d85-c3afa8935243","Type":"ContainerStarted","Data":"eb40e5f3076606ef761086938cb7e8b4aad3ddeaa2e7eb9a9934ed5936e19c64"} Dec 06 06:14:58 crc kubenswrapper[4809]: I1206 06:14:58.769355 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-sj6gs" event={"ID":"6be1b144-2e51-42a1-9430-201ee08ce602","Type":"ContainerStarted","Data":"16d0b81c7785ae188c9c6b120394bec80173aeb7b9c10e720e99c5f6d1b4de19"} Dec 06 06:14:58 crc kubenswrapper[4809]: I1206 06:14:58.772339 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"eb1bd4de-6456-4817-a85d-ab0176153f2c","Type":"ContainerStarted","Data":"cd1fea316ddc0d8b812b68b8ae2836932648f88ecde0d2a0cb057eb24d8002a9"} Dec 06 06:14:58 crc kubenswrapper[4809]: I1206 06:14:58.790226 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-gr99p" podStartSLOduration=3.181124407 podStartE2EDuration="25.790204286s" podCreationTimestamp="2025-12-06 06:14:33 +0000 UTC" firstStartedPulling="2025-12-06 06:14:33.972784901 +0000 UTC m=+1398.861767843" lastFinishedPulling="2025-12-06 06:14:56.58186478 +0000 UTC m=+1421.470847722" observedRunningTime="2025-12-06 06:14:58.782769415 +0000 UTC m=+1423.671752357" watchObservedRunningTime="2025-12-06 06:14:58.790204286 +0000 UTC m=+1423.679187228" Dec 06 06:14:59 crc kubenswrapper[4809]: I1206 06:14:59.019664 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-9bft5" Dec 06 06:14:59 crc kubenswrapper[4809]: I1206 06:14:59.150375 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78537fe2-3d4b-4659-82fc-1c8fa02d0f66-dns-svc\") pod \"78537fe2-3d4b-4659-82fc-1c8fa02d0f66\" (UID: \"78537fe2-3d4b-4659-82fc-1c8fa02d0f66\") " Dec 06 06:14:59 crc kubenswrapper[4809]: I1206 06:14:59.150605 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78537fe2-3d4b-4659-82fc-1c8fa02d0f66-config\") pod \"78537fe2-3d4b-4659-82fc-1c8fa02d0f66\" (UID: \"78537fe2-3d4b-4659-82fc-1c8fa02d0f66\") " Dec 06 06:14:59 crc kubenswrapper[4809]: I1206 06:14:59.151028 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5lhx\" (UniqueName: \"kubernetes.io/projected/78537fe2-3d4b-4659-82fc-1c8fa02d0f66-kube-api-access-b5lhx\") pod \"78537fe2-3d4b-4659-82fc-1c8fa02d0f66\" (UID: \"78537fe2-3d4b-4659-82fc-1c8fa02d0f66\") " Dec 06 06:14:59 crc kubenswrapper[4809]: I1206 06:14:59.151165 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78537fe2-3d4b-4659-82fc-1c8fa02d0f66-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "78537fe2-3d4b-4659-82fc-1c8fa02d0f66" (UID: "78537fe2-3d4b-4659-82fc-1c8fa02d0f66"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:14:59 crc kubenswrapper[4809]: I1206 06:14:59.151248 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78537fe2-3d4b-4659-82fc-1c8fa02d0f66-config" (OuterVolumeSpecName: "config") pod "78537fe2-3d4b-4659-82fc-1c8fa02d0f66" (UID: "78537fe2-3d4b-4659-82fc-1c8fa02d0f66"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:14:59 crc kubenswrapper[4809]: I1206 06:14:59.151716 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78537fe2-3d4b-4659-82fc-1c8fa02d0f66-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 06:14:59 crc kubenswrapper[4809]: I1206 06:14:59.151737 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78537fe2-3d4b-4659-82fc-1c8fa02d0f66-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:14:59 crc kubenswrapper[4809]: I1206 06:14:59.157655 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78537fe2-3d4b-4659-82fc-1c8fa02d0f66-kube-api-access-b5lhx" (OuterVolumeSpecName: "kube-api-access-b5lhx") pod "78537fe2-3d4b-4659-82fc-1c8fa02d0f66" (UID: "78537fe2-3d4b-4659-82fc-1c8fa02d0f66"). InnerVolumeSpecName "kube-api-access-b5lhx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:14:59 crc kubenswrapper[4809]: I1206 06:14:59.162335 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 06 06:14:59 crc kubenswrapper[4809]: W1206 06:14:59.179353 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1ffa5f54_0c40_45a6_9ebb_a5eca04810ee.slice/crio-383ab5daa992d3e7065396b73fdd9c04bbd9044f281e9342e2dd8be5e4fda8be WatchSource:0}: Error finding container 383ab5daa992d3e7065396b73fdd9c04bbd9044f281e9342e2dd8be5e4fda8be: Status 404 returned error can't find the container with id 383ab5daa992d3e7065396b73fdd9c04bbd9044f281e9342e2dd8be5e4fda8be Dec 06 06:14:59 crc kubenswrapper[4809]: E1206 06:14:59.207693 4809 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Dec 06 06:14:59 crc kubenswrapper[4809]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/de481cc6-bc8d-4a07-991b-72cad738f1d5/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Dec 06 06:14:59 crc kubenswrapper[4809]: > podSandboxID="2c8fde9c0e29b0ed8ca9b413a966026f7c882190f281e14ed5c48d80d16cf1f4" Dec 06 06:14:59 crc kubenswrapper[4809]: E1206 06:14:59.207957 4809 kuberuntime_manager.go:1274] "Unhandled Error" err=< Dec 06 06:14:59 crc kubenswrapper[4809]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sc5mn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-k6s2d_openstack(de481cc6-bc8d-4a07-991b-72cad738f1d5): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/de481cc6-bc8d-4a07-991b-72cad738f1d5/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Dec 06 06:14:59 crc kubenswrapper[4809]: > logger="UnhandledError" Dec 06 06:14:59 crc kubenswrapper[4809]: E1206 06:14:59.210006 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/de481cc6-bc8d-4a07-991b-72cad738f1d5/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-666b6646f7-k6s2d" podUID="de481cc6-bc8d-4a07-991b-72cad738f1d5" Dec 06 06:14:59 crc kubenswrapper[4809]: I1206 06:14:59.254463 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5lhx\" (UniqueName: \"kubernetes.io/projected/78537fe2-3d4b-4659-82fc-1c8fa02d0f66-kube-api-access-b5lhx\") on node \"crc\" DevicePath \"\"" Dec 06 06:14:59 crc kubenswrapper[4809]: I1206 06:14:59.419910 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68b27fd9-c5ea-4e7a-9639-afb6adecf768" path="/var/lib/kubelet/pods/68b27fd9-c5ea-4e7a-9639-afb6adecf768/volumes" Dec 06 06:14:59 crc kubenswrapper[4809]: I1206 06:14:59.804707 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-xhlzm" event={"ID":"6b64391f-74d7-4b69-a898-f7375057c6d5","Type":"ContainerStarted","Data":"b2ab9ebde5cc9a80f97a7498ee09c937e3295b932162bbd91cc5a3a6abf6abec"} Dec 06 06:14:59 crc kubenswrapper[4809]: I1206 06:14:59.806973 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-jgqjw" event={"ID":"92ce9234-d708-449e-9874-97ac8a4b994c","Type":"ContainerStarted","Data":"d987cb0657ad1abb5ea7779e3825a2a229a4c3009c56667d5efff2be3e4e3b81"} Dec 06 06:14:59 crc kubenswrapper[4809]: I1206 06:14:59.809481 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f6536b25-1cc1-4f74-bf46-613965ddf185","Type":"ContainerStarted","Data":"44015bc9fdb90f164a86a172d0c7a083c8cbce1f698d4de2a734ce7aa5f22bce"} Dec 06 06:14:59 crc kubenswrapper[4809]: I1206 06:14:59.811229 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5cccfc6d89-9j7zs" event={"ID":"ac4de3b3-b463-4e5b-b935-ee6561519ddf","Type":"ContainerStarted","Data":"822b38733fd9e145c7b8cfcb2dec8062809c96f9fcb0b594cfb2f762c29fd306"} Dec 06 06:14:59 crc kubenswrapper[4809]: I1206 06:14:59.812546 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"1ffa5f54-0c40-45a6-9ebb-a5eca04810ee","Type":"ContainerStarted","Data":"383ab5daa992d3e7065396b73fdd9c04bbd9044f281e9342e2dd8be5e4fda8be"} Dec 06 06:14:59 crc kubenswrapper[4809]: I1206 06:14:59.813583 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"932f2f72-a070-425a-b065-64e293954cc5","Type":"ContainerStarted","Data":"89bd336be0c2b06ee68c8096cb5583ccbac108db3ea9e0e5b5a58390e095c41a"} Dec 06 06:14:59 crc kubenswrapper[4809]: I1206 06:14:59.816096 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-9bft5" event={"ID":"78537fe2-3d4b-4659-82fc-1c8fa02d0f66","Type":"ContainerDied","Data":"07f0f5ce67102fa261a1d47b13ff3c8824183d2e6e07d1d15afbe5c667a1c656"} Dec 06 06:14:59 crc kubenswrapper[4809]: I1206 06:14:59.816431 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-9bft5" Dec 06 06:14:59 crc kubenswrapper[4809]: I1206 06:14:59.886608 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-9bft5"] Dec 06 06:14:59 crc kubenswrapper[4809]: I1206 06:14:59.895582 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-9bft5"] Dec 06 06:15:00 crc kubenswrapper[4809]: I1206 06:15:00.150586 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416695-sc6ls"] Dec 06 06:15:00 crc kubenswrapper[4809]: I1206 06:15:00.152513 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416695-sc6ls" Dec 06 06:15:00 crc kubenswrapper[4809]: I1206 06:15:00.154702 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 06:15:00 crc kubenswrapper[4809]: I1206 06:15:00.154742 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 06:15:00 crc kubenswrapper[4809]: I1206 06:15:00.164476 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416695-sc6ls"] Dec 06 06:15:00 crc kubenswrapper[4809]: I1206 06:15:00.289588 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6d92fc9c-754a-403e-b49f-f0ee023f60d6-config-volume\") pod \"collect-profiles-29416695-sc6ls\" (UID: \"6d92fc9c-754a-403e-b49f-f0ee023f60d6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416695-sc6ls" Dec 06 06:15:00 crc kubenswrapper[4809]: I1206 06:15:00.290214 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6d92fc9c-754a-403e-b49f-f0ee023f60d6-secret-volume\") pod \"collect-profiles-29416695-sc6ls\" (UID: \"6d92fc9c-754a-403e-b49f-f0ee023f60d6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416695-sc6ls" Dec 06 06:15:00 crc kubenswrapper[4809]: I1206 06:15:00.290305 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ptqnr\" (UniqueName: \"kubernetes.io/projected/6d92fc9c-754a-403e-b49f-f0ee023f60d6-kube-api-access-ptqnr\") pod \"collect-profiles-29416695-sc6ls\" (UID: \"6d92fc9c-754a-403e-b49f-f0ee023f60d6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416695-sc6ls" Dec 06 06:15:00 crc kubenswrapper[4809]: I1206 06:15:00.391662 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ptqnr\" (UniqueName: \"kubernetes.io/projected/6d92fc9c-754a-403e-b49f-f0ee023f60d6-kube-api-access-ptqnr\") pod \"collect-profiles-29416695-sc6ls\" (UID: \"6d92fc9c-754a-403e-b49f-f0ee023f60d6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416695-sc6ls" Dec 06 06:15:00 crc kubenswrapper[4809]: I1206 06:15:00.391790 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6d92fc9c-754a-403e-b49f-f0ee023f60d6-config-volume\") pod \"collect-profiles-29416695-sc6ls\" (UID: \"6d92fc9c-754a-403e-b49f-f0ee023f60d6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416695-sc6ls" Dec 06 06:15:00 crc kubenswrapper[4809]: I1206 06:15:00.391814 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6d92fc9c-754a-403e-b49f-f0ee023f60d6-secret-volume\") pod \"collect-profiles-29416695-sc6ls\" (UID: \"6d92fc9c-754a-403e-b49f-f0ee023f60d6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416695-sc6ls" Dec 06 06:15:00 crc kubenswrapper[4809]: I1206 06:15:00.400581 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6d92fc9c-754a-403e-b49f-f0ee023f60d6-config-volume\") pod \"collect-profiles-29416695-sc6ls\" (UID: \"6d92fc9c-754a-403e-b49f-f0ee023f60d6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416695-sc6ls" Dec 06 06:15:00 crc kubenswrapper[4809]: I1206 06:15:00.422770 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6d92fc9c-754a-403e-b49f-f0ee023f60d6-secret-volume\") pod \"collect-profiles-29416695-sc6ls\" (UID: \"6d92fc9c-754a-403e-b49f-f0ee023f60d6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416695-sc6ls" Dec 06 06:15:00 crc kubenswrapper[4809]: I1206 06:15:00.431774 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ptqnr\" (UniqueName: \"kubernetes.io/projected/6d92fc9c-754a-403e-b49f-f0ee023f60d6-kube-api-access-ptqnr\") pod \"collect-profiles-29416695-sc6ls\" (UID: \"6d92fc9c-754a-403e-b49f-f0ee023f60d6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416695-sc6ls" Dec 06 06:15:00 crc kubenswrapper[4809]: I1206 06:15:00.511589 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416695-sc6ls" Dec 06 06:15:01 crc kubenswrapper[4809]: I1206 06:15:01.402706 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78537fe2-3d4b-4659-82fc-1c8fa02d0f66" path="/var/lib/kubelet/pods/78537fe2-3d4b-4659-82fc-1c8fa02d0f66/volumes" Dec 06 06:15:03 crc kubenswrapper[4809]: I1206 06:15:03.407062 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-gr99p" Dec 06 06:15:03 crc kubenswrapper[4809]: I1206 06:15:03.499873 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-k6s2d"] Dec 06 06:15:04 crc kubenswrapper[4809]: I1206 06:15:04.495921 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:15:04 crc kubenswrapper[4809]: I1206 06:15:04.496033 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:15:04 crc kubenswrapper[4809]: I1206 06:15:04.496082 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-npms2" Dec 06 06:15:04 crc kubenswrapper[4809]: I1206 06:15:04.496762 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fac0817a1ba54b752b764ab02ba5dbd6db50ecfcf2de9f18d391e8a61207257c"} pod="openshift-machine-config-operator/machine-config-daemon-npms2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 06:15:04 crc kubenswrapper[4809]: I1206 06:15:04.496867 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" containerID="cri-o://fac0817a1ba54b752b764ab02ba5dbd6db50ecfcf2de9f18d391e8a61207257c" gracePeriod=600 Dec 06 06:15:04 crc kubenswrapper[4809]: I1206 06:15:04.891191 4809 generic.go:334] "Generic (PLEG): container finished" podID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerID="fac0817a1ba54b752b764ab02ba5dbd6db50ecfcf2de9f18d391e8a61207257c" exitCode=0 Dec 06 06:15:04 crc kubenswrapper[4809]: I1206 06:15:04.891538 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" event={"ID":"cbf4cf62-024e-4703-a8b8-9aecda9cd26a","Type":"ContainerDied","Data":"fac0817a1ba54b752b764ab02ba5dbd6db50ecfcf2de9f18d391e8a61207257c"} Dec 06 06:15:04 crc kubenswrapper[4809]: I1206 06:15:04.891576 4809 scope.go:117] "RemoveContainer" containerID="9b0b6808cda0fc2270940d0650cf854006725ea29b339e76c18a025e4f361d32" Dec 06 06:15:05 crc kubenswrapper[4809]: I1206 06:15:05.364357 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416695-sc6ls"] Dec 06 06:15:05 crc kubenswrapper[4809]: I1206 06:15:05.901355 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5cccfc6d89-9j7zs" event={"ID":"ac4de3b3-b463-4e5b-b935-ee6561519ddf","Type":"ContainerStarted","Data":"6d6ce77cdd04240fb4320bd0fb54dcf521422effd4a9d25f1cb2122d5e5f7008"} Dec 06 06:15:05 crc kubenswrapper[4809]: I1206 06:15:05.922179 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-5cccfc6d89-9j7zs" podStartSLOduration=25.92213877 podStartE2EDuration="25.92213877s" podCreationTimestamp="2025-12-06 06:14:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:15:05.92177665 +0000 UTC m=+1430.810759642" watchObservedRunningTime="2025-12-06 06:15:05.92213877 +0000 UTC m=+1430.811121722" Dec 06 06:15:06 crc kubenswrapper[4809]: I1206 06:15:06.911542 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-k6s2d" event={"ID":"de481cc6-bc8d-4a07-991b-72cad738f1d5","Type":"ContainerStarted","Data":"0465a19ffa3649a2d4a3c7f812fb5d1d57b19e3d4acbdf3678a7c4668d13ddb8"} Dec 06 06:15:06 crc kubenswrapper[4809]: I1206 06:15:06.911644 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-k6s2d" podUID="de481cc6-bc8d-4a07-991b-72cad738f1d5" containerName="dnsmasq-dns" containerID="cri-o://0465a19ffa3649a2d4a3c7f812fb5d1d57b19e3d4acbdf3678a7c4668d13ddb8" gracePeriod=10 Dec 06 06:15:06 crc kubenswrapper[4809]: I1206 06:15:06.912037 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-k6s2d" Dec 06 06:15:06 crc kubenswrapper[4809]: I1206 06:15:06.934368 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-k6s2d" podStartSLOduration=12.230050936 podStartE2EDuration="34.934348943s" podCreationTimestamp="2025-12-06 06:14:32 +0000 UTC" firstStartedPulling="2025-12-06 06:14:33.795780238 +0000 UTC m=+1398.684763180" lastFinishedPulling="2025-12-06 06:14:56.500078245 +0000 UTC m=+1421.389061187" observedRunningTime="2025-12-06 06:15:06.929975245 +0000 UTC m=+1431.818958187" watchObservedRunningTime="2025-12-06 06:15:06.934348943 +0000 UTC m=+1431.823331885" Dec 06 06:15:07 crc kubenswrapper[4809]: W1206 06:15:07.168037 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6d92fc9c_754a_403e_b49f_f0ee023f60d6.slice/crio-9f6df899c96abe9d3f575d00561fe5f60b4b5384224ce6ecc5b700b89902b010 WatchSource:0}: Error finding container 9f6df899c96abe9d3f575d00561fe5f60b4b5384224ce6ecc5b700b89902b010: Status 404 returned error can't find the container with id 9f6df899c96abe9d3f575d00561fe5f60b4b5384224ce6ecc5b700b89902b010 Dec 06 06:15:07 crc kubenswrapper[4809]: I1206 06:15:07.925296 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416695-sc6ls" event={"ID":"6d92fc9c-754a-403e-b49f-f0ee023f60d6","Type":"ContainerStarted","Data":"9f6df899c96abe9d3f575d00561fe5f60b4b5384224ce6ecc5b700b89902b010"} Dec 06 06:15:07 crc kubenswrapper[4809]: I1206 06:15:07.928652 4809 generic.go:334] "Generic (PLEG): container finished" podID="de481cc6-bc8d-4a07-991b-72cad738f1d5" containerID="0465a19ffa3649a2d4a3c7f812fb5d1d57b19e3d4acbdf3678a7c4668d13ddb8" exitCode=0 Dec 06 06:15:07 crc kubenswrapper[4809]: I1206 06:15:07.928720 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-k6s2d" event={"ID":"de481cc6-bc8d-4a07-991b-72cad738f1d5","Type":"ContainerDied","Data":"0465a19ffa3649a2d4a3c7f812fb5d1d57b19e3d4acbdf3678a7c4668d13ddb8"} Dec 06 06:15:08 crc kubenswrapper[4809]: I1206 06:15:08.076653 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-k6s2d" Dec 06 06:15:08 crc kubenswrapper[4809]: I1206 06:15:08.188394 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sc5mn\" (UniqueName: \"kubernetes.io/projected/de481cc6-bc8d-4a07-991b-72cad738f1d5-kube-api-access-sc5mn\") pod \"de481cc6-bc8d-4a07-991b-72cad738f1d5\" (UID: \"de481cc6-bc8d-4a07-991b-72cad738f1d5\") " Dec 06 06:15:08 crc kubenswrapper[4809]: I1206 06:15:08.188812 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de481cc6-bc8d-4a07-991b-72cad738f1d5-config\") pod \"de481cc6-bc8d-4a07-991b-72cad738f1d5\" (UID: \"de481cc6-bc8d-4a07-991b-72cad738f1d5\") " Dec 06 06:15:08 crc kubenswrapper[4809]: I1206 06:15:08.188896 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de481cc6-bc8d-4a07-991b-72cad738f1d5-dns-svc\") pod \"de481cc6-bc8d-4a07-991b-72cad738f1d5\" (UID: \"de481cc6-bc8d-4a07-991b-72cad738f1d5\") " Dec 06 06:15:08 crc kubenswrapper[4809]: I1206 06:15:08.194215 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de481cc6-bc8d-4a07-991b-72cad738f1d5-kube-api-access-sc5mn" (OuterVolumeSpecName: "kube-api-access-sc5mn") pod "de481cc6-bc8d-4a07-991b-72cad738f1d5" (UID: "de481cc6-bc8d-4a07-991b-72cad738f1d5"). InnerVolumeSpecName "kube-api-access-sc5mn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:15:08 crc kubenswrapper[4809]: I1206 06:15:08.238996 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de481cc6-bc8d-4a07-991b-72cad738f1d5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "de481cc6-bc8d-4a07-991b-72cad738f1d5" (UID: "de481cc6-bc8d-4a07-991b-72cad738f1d5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:15:08 crc kubenswrapper[4809]: I1206 06:15:08.254228 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de481cc6-bc8d-4a07-991b-72cad738f1d5-config" (OuterVolumeSpecName: "config") pod "de481cc6-bc8d-4a07-991b-72cad738f1d5" (UID: "de481cc6-bc8d-4a07-991b-72cad738f1d5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:15:08 crc kubenswrapper[4809]: I1206 06:15:08.291777 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sc5mn\" (UniqueName: \"kubernetes.io/projected/de481cc6-bc8d-4a07-991b-72cad738f1d5-kube-api-access-sc5mn\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:08 crc kubenswrapper[4809]: I1206 06:15:08.291831 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de481cc6-bc8d-4a07-991b-72cad738f1d5-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:08 crc kubenswrapper[4809]: I1206 06:15:08.291847 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de481cc6-bc8d-4a07-991b-72cad738f1d5-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:08 crc kubenswrapper[4809]: I1206 06:15:08.945300 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-k6s2d" event={"ID":"de481cc6-bc8d-4a07-991b-72cad738f1d5","Type":"ContainerDied","Data":"2c8fde9c0e29b0ed8ca9b413a966026f7c882190f281e14ed5c48d80d16cf1f4"} Dec 06 06:15:08 crc kubenswrapper[4809]: I1206 06:15:08.945354 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-k6s2d" Dec 06 06:15:08 crc kubenswrapper[4809]: I1206 06:15:08.945371 4809 scope.go:117] "RemoveContainer" containerID="0465a19ffa3649a2d4a3c7f812fb5d1d57b19e3d4acbdf3678a7c4668d13ddb8" Dec 06 06:15:09 crc kubenswrapper[4809]: I1206 06:15:09.118979 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 06:15:09 crc kubenswrapper[4809]: I1206 06:15:09.248108 4809 scope.go:117] "RemoveContainer" containerID="aafb8c934a2e51c587c3669678f2f7e847e52f557a0d98bf2c575830dbb87a35" Dec 06 06:15:09 crc kubenswrapper[4809]: I1206 06:15:09.281526 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-k6s2d"] Dec 06 06:15:09 crc kubenswrapper[4809]: I1206 06:15:09.294715 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-k6s2d"] Dec 06 06:15:09 crc kubenswrapper[4809]: I1206 06:15:09.404213 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de481cc6-bc8d-4a07-991b-72cad738f1d5" path="/var/lib/kubelet/pods/de481cc6-bc8d-4a07-991b-72cad738f1d5/volumes" Dec 06 06:15:09 crc kubenswrapper[4809]: I1206 06:15:09.961831 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e7056938-6e36-4631-adfb-01f116eba651","Type":"ContainerStarted","Data":"80b446b080b5af08c3f8e8f791d6a62e837e30096220e20519c490abddf5dc99"} Dec 06 06:15:09 crc kubenswrapper[4809]: I1206 06:15:09.967512 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" event={"ID":"cbf4cf62-024e-4703-a8b8-9aecda9cd26a","Type":"ContainerStarted","Data":"9bc2dc88b3b9f309ce0169aa510326dd77022dece4c69cab3141df00fe5c0106"} Dec 06 06:15:09 crc kubenswrapper[4809]: I1206 06:15:09.970374 4809 generic.go:334] "Generic (PLEG): container finished" podID="92ce9234-d708-449e-9874-97ac8a4b994c" containerID="78781dd57495d992802cbd4bf6ee10ac2016c11a7bf4321db9b6605385c3d4bf" exitCode=0 Dec 06 06:15:09 crc kubenswrapper[4809]: I1206 06:15:09.970449 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-jgqjw" event={"ID":"92ce9234-d708-449e-9874-97ac8a4b994c","Type":"ContainerDied","Data":"78781dd57495d992802cbd4bf6ee10ac2016c11a7bf4321db9b6605385c3d4bf"} Dec 06 06:15:09 crc kubenswrapper[4809]: I1206 06:15:09.972057 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"cdf93098-585a-4391-8419-c5bc7ab1c567","Type":"ContainerStarted","Data":"28674d3c480d379e5c1c3169207c00292adf403404c050cd9ae2bb9fcdc42dc9"} Dec 06 06:15:09 crc kubenswrapper[4809]: I1206 06:15:09.972598 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 06 06:15:09 crc kubenswrapper[4809]: I1206 06:15:09.973913 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"932f2f72-a070-425a-b065-64e293954cc5","Type":"ContainerStarted","Data":"5c7eab35f71731244de430f088ffa5dcb664971763735e0832411ba7e30added"} Dec 06 06:15:09 crc kubenswrapper[4809]: I1206 06:15:09.985249 4809 generic.go:334] "Generic (PLEG): container finished" podID="6d92fc9c-754a-403e-b49f-f0ee023f60d6" containerID="62d3284237ce1ce31a88a82b5a0128ef317c936516d5ccef3cbb0b1ba7ed5598" exitCode=0 Dec 06 06:15:09 crc kubenswrapper[4809]: I1206 06:15:09.985536 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416695-sc6ls" event={"ID":"6d92fc9c-754a-403e-b49f-f0ee023f60d6","Type":"ContainerDied","Data":"62d3284237ce1ce31a88a82b5a0128ef317c936516d5ccef3cbb0b1ba7ed5598"} Dec 06 06:15:09 crc kubenswrapper[4809]: I1206 06:15:09.996084 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"7c596e14-b96f-4e5f-a8ac-3dc66fa0276d","Type":"ContainerStarted","Data":"ab28f695346582d425f15312a5a3d1cb1012fba72f5ab5cf6105f00e0a069792"} Dec 06 06:15:10 crc kubenswrapper[4809]: I1206 06:15:10.005098 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"9d6c40a5-e2ff-4983-b9d8-fc483703870c","Type":"ContainerStarted","Data":"5156f98bbc3225408d967cb01dae970115bf4a580fb7f2d11995ba12cc58568c"} Dec 06 06:15:10 crc kubenswrapper[4809]: I1206 06:15:10.006781 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 06 06:15:10 crc kubenswrapper[4809]: I1206 06:15:10.013056 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-xhlzm" event={"ID":"6b64391f-74d7-4b69-a898-f7375057c6d5","Type":"ContainerStarted","Data":"bc068182b878e903c8186cc73f966e63c46d5ae106802c108058ef9ba685f71e"} Dec 06 06:15:10 crc kubenswrapper[4809]: I1206 06:15:10.014161 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-xhlzm" Dec 06 06:15:10 crc kubenswrapper[4809]: I1206 06:15:10.028869 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-sj6gs" event={"ID":"6be1b144-2e51-42a1-9430-201ee08ce602","Type":"ContainerStarted","Data":"579b9fa21a7380ab38ffd1e8c029c38f45f541f2c5cc4ef8d0a9fb4d7a1f131a"} Dec 06 06:15:10 crc kubenswrapper[4809]: I1206 06:15:10.040114 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"1ffa5f54-0c40-45a6-9ebb-a5eca04810ee","Type":"ContainerStarted","Data":"bd8b6cd793c929664b4ae4ab6f8b07cd231cdc82ed372c09e1dce0e83422734c"} Dec 06 06:15:10 crc kubenswrapper[4809]: I1206 06:15:10.074538 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-xhlzm" podStartSLOduration=18.852276667 podStartE2EDuration="28.074517444s" podCreationTimestamp="2025-12-06 06:14:42 +0000 UTC" firstStartedPulling="2025-12-06 06:14:58.915419702 +0000 UTC m=+1423.804402644" lastFinishedPulling="2025-12-06 06:15:08.137660479 +0000 UTC m=+1433.026643421" observedRunningTime="2025-12-06 06:15:10.06621266 +0000 UTC m=+1434.955195602" watchObservedRunningTime="2025-12-06 06:15:10.074517444 +0000 UTC m=+1434.963500376" Dec 06 06:15:10 crc kubenswrapper[4809]: I1206 06:15:10.193609 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=25.995440864 podStartE2EDuration="33.193590025s" podCreationTimestamp="2025-12-06 06:14:37 +0000 UTC" firstStartedPulling="2025-12-06 06:14:57.110247177 +0000 UTC m=+1421.999230119" lastFinishedPulling="2025-12-06 06:15:04.308396338 +0000 UTC m=+1429.197379280" observedRunningTime="2025-12-06 06:15:10.141311705 +0000 UTC m=+1435.030294647" watchObservedRunningTime="2025-12-06 06:15:10.193590025 +0000 UTC m=+1435.082572967" Dec 06 06:15:10 crc kubenswrapper[4809]: I1206 06:15:10.283055 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=20.172468503 podStartE2EDuration="31.283030037s" podCreationTimestamp="2025-12-06 06:14:39 +0000 UTC" firstStartedPulling="2025-12-06 06:14:57.129565428 +0000 UTC m=+1422.018548370" lastFinishedPulling="2025-12-06 06:15:08.240126962 +0000 UTC m=+1433.129109904" observedRunningTime="2025-12-06 06:15:10.174986973 +0000 UTC m=+1435.063969915" watchObservedRunningTime="2025-12-06 06:15:10.283030037 +0000 UTC m=+1435.172012979" Dec 06 06:15:10 crc kubenswrapper[4809]: I1206 06:15:10.388239 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-sj6gs" podStartSLOduration=23.434165271 podStartE2EDuration="31.388216303s" podCreationTimestamp="2025-12-06 06:14:39 +0000 UTC" firstStartedPulling="2025-12-06 06:14:58.482578681 +0000 UTC m=+1423.371561623" lastFinishedPulling="2025-12-06 06:15:06.436629713 +0000 UTC m=+1431.325612655" observedRunningTime="2025-12-06 06:15:10.379315283 +0000 UTC m=+1435.268298225" watchObservedRunningTime="2025-12-06 06:15:10.388216303 +0000 UTC m=+1435.277199245" Dec 06 06:15:10 crc kubenswrapper[4809]: I1206 06:15:10.726836 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-5cccfc6d89-9j7zs" Dec 06 06:15:10 crc kubenswrapper[4809]: I1206 06:15:10.726897 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-5cccfc6d89-9j7zs" Dec 06 06:15:10 crc kubenswrapper[4809]: I1206 06:15:10.732760 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-5cccfc6d89-9j7zs" Dec 06 06:15:11 crc kubenswrapper[4809]: I1206 06:15:11.053647 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-5cccfc6d89-9j7zs" Dec 06 06:15:11 crc kubenswrapper[4809]: I1206 06:15:11.126086 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-6499fbfdc8-929cv"] Dec 06 06:15:11 crc kubenswrapper[4809]: I1206 06:15:11.517416 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416695-sc6ls" Dec 06 06:15:11 crc kubenswrapper[4809]: I1206 06:15:11.687158 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6d92fc9c-754a-403e-b49f-f0ee023f60d6-secret-volume\") pod \"6d92fc9c-754a-403e-b49f-f0ee023f60d6\" (UID: \"6d92fc9c-754a-403e-b49f-f0ee023f60d6\") " Dec 06 06:15:11 crc kubenswrapper[4809]: I1206 06:15:11.687413 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ptqnr\" (UniqueName: \"kubernetes.io/projected/6d92fc9c-754a-403e-b49f-f0ee023f60d6-kube-api-access-ptqnr\") pod \"6d92fc9c-754a-403e-b49f-f0ee023f60d6\" (UID: \"6d92fc9c-754a-403e-b49f-f0ee023f60d6\") " Dec 06 06:15:11 crc kubenswrapper[4809]: I1206 06:15:11.687468 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6d92fc9c-754a-403e-b49f-f0ee023f60d6-config-volume\") pod \"6d92fc9c-754a-403e-b49f-f0ee023f60d6\" (UID: \"6d92fc9c-754a-403e-b49f-f0ee023f60d6\") " Dec 06 06:15:11 crc kubenswrapper[4809]: I1206 06:15:11.688660 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d92fc9c-754a-403e-b49f-f0ee023f60d6-config-volume" (OuterVolumeSpecName: "config-volume") pod "6d92fc9c-754a-403e-b49f-f0ee023f60d6" (UID: "6d92fc9c-754a-403e-b49f-f0ee023f60d6"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:15:11 crc kubenswrapper[4809]: I1206 06:15:11.689724 4809 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6d92fc9c-754a-403e-b49f-f0ee023f60d6-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:11 crc kubenswrapper[4809]: I1206 06:15:11.697635 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d92fc9c-754a-403e-b49f-f0ee023f60d6-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "6d92fc9c-754a-403e-b49f-f0ee023f60d6" (UID: "6d92fc9c-754a-403e-b49f-f0ee023f60d6"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:15:11 crc kubenswrapper[4809]: I1206 06:15:11.792267 4809 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6d92fc9c-754a-403e-b49f-f0ee023f60d6-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:11 crc kubenswrapper[4809]: I1206 06:15:11.877687 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d92fc9c-754a-403e-b49f-f0ee023f60d6-kube-api-access-ptqnr" (OuterVolumeSpecName: "kube-api-access-ptqnr") pod "6d92fc9c-754a-403e-b49f-f0ee023f60d6" (UID: "6d92fc9c-754a-403e-b49f-f0ee023f60d6"). InnerVolumeSpecName "kube-api-access-ptqnr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:15:11 crc kubenswrapper[4809]: I1206 06:15:11.894088 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ptqnr\" (UniqueName: \"kubernetes.io/projected/6d92fc9c-754a-403e-b49f-f0ee023f60d6-kube-api-access-ptqnr\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:12 crc kubenswrapper[4809]: I1206 06:15:12.061359 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-jgqjw" event={"ID":"92ce9234-d708-449e-9874-97ac8a4b994c","Type":"ContainerStarted","Data":"c010f417f454d40a412962389f5ce39c0a7fecb8a5a3e0694a76f4be993561ae"} Dec 06 06:15:12 crc kubenswrapper[4809]: I1206 06:15:12.063162 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416695-sc6ls" Dec 06 06:15:12 crc kubenswrapper[4809]: I1206 06:15:12.064212 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416695-sc6ls" event={"ID":"6d92fc9c-754a-403e-b49f-f0ee023f60d6","Type":"ContainerDied","Data":"9f6df899c96abe9d3f575d00561fe5f60b4b5384224ce6ecc5b700b89902b010"} Dec 06 06:15:12 crc kubenswrapper[4809]: I1206 06:15:12.064241 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f6df899c96abe9d3f575d00561fe5f60b4b5384224ce6ecc5b700b89902b010" Dec 06 06:15:13 crc kubenswrapper[4809]: I1206 06:15:13.075359 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-jgqjw" event={"ID":"92ce9234-d708-449e-9874-97ac8a4b994c","Type":"ContainerStarted","Data":"a0914b9f6737357240eb417370cb7d98dd4bee9a4e6d6011fcc3add789ec4448"} Dec 06 06:15:13 crc kubenswrapper[4809]: I1206 06:15:13.078138 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-jgqjw" Dec 06 06:15:13 crc kubenswrapper[4809]: I1206 06:15:13.078182 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-jgqjw" Dec 06 06:15:13 crc kubenswrapper[4809]: I1206 06:15:13.081153 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f6536b25-1cc1-4f74-bf46-613965ddf185","Type":"ContainerStarted","Data":"df03251430e15c56ceb99e1135791e64f339c2265989799fec5b3d7450799a7d"} Dec 06 06:15:13 crc kubenswrapper[4809]: I1206 06:15:13.102483 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-jgqjw" podStartSLOduration=21.940789945 podStartE2EDuration="31.102460969s" podCreationTimestamp="2025-12-06 06:14:42 +0000 UTC" firstStartedPulling="2025-12-06 06:14:58.946203362 +0000 UTC m=+1423.835186304" lastFinishedPulling="2025-12-06 06:15:08.107874386 +0000 UTC m=+1432.996857328" observedRunningTime="2025-12-06 06:15:13.097360682 +0000 UTC m=+1437.986343644" watchObservedRunningTime="2025-12-06 06:15:13.102460969 +0000 UTC m=+1437.991443911" Dec 06 06:15:16 crc kubenswrapper[4809]: I1206 06:15:16.123861 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"1ffa5f54-0c40-45a6-9ebb-a5eca04810ee","Type":"ContainerStarted","Data":"391a1ceca492139b1a168141745d8a443c5b4b3be4d0517945c0a780900a4e7f"} Dec 06 06:15:16 crc kubenswrapper[4809]: I1206 06:15:16.126453 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"932f2f72-a070-425a-b065-64e293954cc5","Type":"ContainerStarted","Data":"f67ad09ba5aa7ceb19d24a0aff9b98b9be7fbbdb052ea579ca859c1933e9e023"} Dec 06 06:15:16 crc kubenswrapper[4809]: I1206 06:15:16.149204 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=13.701588959 podStartE2EDuration="30.1491839s" podCreationTimestamp="2025-12-06 06:14:46 +0000 UTC" firstStartedPulling="2025-12-06 06:14:59.20760138 +0000 UTC m=+1424.096584322" lastFinishedPulling="2025-12-06 06:15:15.655196321 +0000 UTC m=+1440.544179263" observedRunningTime="2025-12-06 06:15:16.14287688 +0000 UTC m=+1441.031859822" watchObservedRunningTime="2025-12-06 06:15:16.1491839 +0000 UTC m=+1441.038166852" Dec 06 06:15:16 crc kubenswrapper[4809]: I1206 06:15:16.165884 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=17.454390114 podStartE2EDuration="34.1658572s" podCreationTimestamp="2025-12-06 06:14:42 +0000 UTC" firstStartedPulling="2025-12-06 06:14:58.946653784 +0000 UTC m=+1423.835636726" lastFinishedPulling="2025-12-06 06:15:15.65812087 +0000 UTC m=+1440.547103812" observedRunningTime="2025-12-06 06:15:16.160714542 +0000 UTC m=+1441.049697484" watchObservedRunningTime="2025-12-06 06:15:16.1658572 +0000 UTC m=+1441.054840142" Dec 06 06:15:17 crc kubenswrapper[4809]: I1206 06:15:17.158648 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 06 06:15:17 crc kubenswrapper[4809]: I1206 06:15:17.196517 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 06 06:15:17 crc kubenswrapper[4809]: I1206 06:15:17.381223 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 06 06:15:17 crc kubenswrapper[4809]: I1206 06:15:17.381268 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 06 06:15:17 crc kubenswrapper[4809]: I1206 06:15:17.442501 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 06 06:15:17 crc kubenswrapper[4809]: I1206 06:15:17.709237 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.147621 4809 generic.go:334] "Generic (PLEG): container finished" podID="7c596e14-b96f-4e5f-a8ac-3dc66fa0276d" containerID="ab28f695346582d425f15312a5a3d1cb1012fba72f5ab5cf6105f00e0a069792" exitCode=0 Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.147683 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"7c596e14-b96f-4e5f-a8ac-3dc66fa0276d","Type":"ContainerDied","Data":"ab28f695346582d425f15312a5a3d1cb1012fba72f5ab5cf6105f00e0a069792"} Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.166780 4809 generic.go:334] "Generic (PLEG): container finished" podID="e7056938-6e36-4631-adfb-01f116eba651" containerID="80b446b080b5af08c3f8e8f791d6a62e837e30096220e20519c490abddf5dc99" exitCode=0 Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.166892 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e7056938-6e36-4631-adfb-01f116eba651","Type":"ContainerDied","Data":"80b446b080b5af08c3f8e8f791d6a62e837e30096220e20519c490abddf5dc99"} Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.169366 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.238615 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.289712 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.454760 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-vnwqm"] Dec 06 06:15:18 crc kubenswrapper[4809]: E1206 06:15:18.455669 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d92fc9c-754a-403e-b49f-f0ee023f60d6" containerName="collect-profiles" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.455688 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d92fc9c-754a-403e-b49f-f0ee023f60d6" containerName="collect-profiles" Dec 06 06:15:18 crc kubenswrapper[4809]: E1206 06:15:18.455710 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de481cc6-bc8d-4a07-991b-72cad738f1d5" containerName="dnsmasq-dns" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.455718 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="de481cc6-bc8d-4a07-991b-72cad738f1d5" containerName="dnsmasq-dns" Dec 06 06:15:18 crc kubenswrapper[4809]: E1206 06:15:18.455750 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de481cc6-bc8d-4a07-991b-72cad738f1d5" containerName="init" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.455759 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="de481cc6-bc8d-4a07-991b-72cad738f1d5" containerName="init" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.456004 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d92fc9c-754a-403e-b49f-f0ee023f60d6" containerName="collect-profiles" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.456023 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="de481cc6-bc8d-4a07-991b-72cad738f1d5" containerName="dnsmasq-dns" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.457329 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-vnwqm" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.464211 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.472850 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-vnwqm"] Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.543337 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxf5t\" (UniqueName: \"kubernetes.io/projected/10497d10-a5e4-42e5-817c-1c41a2ef29e2-kube-api-access-kxf5t\") pod \"dnsmasq-dns-7f896c8c65-vnwqm\" (UID: \"10497d10-a5e4-42e5-817c-1c41a2ef29e2\") " pod="openstack/dnsmasq-dns-7f896c8c65-vnwqm" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.543394 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/10497d10-a5e4-42e5-817c-1c41a2ef29e2-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-vnwqm\" (UID: \"10497d10-a5e4-42e5-817c-1c41a2ef29e2\") " pod="openstack/dnsmasq-dns-7f896c8c65-vnwqm" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.543431 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10497d10-a5e4-42e5-817c-1c41a2ef29e2-config\") pod \"dnsmasq-dns-7f896c8c65-vnwqm\" (UID: \"10497d10-a5e4-42e5-817c-1c41a2ef29e2\") " pod="openstack/dnsmasq-dns-7f896c8c65-vnwqm" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.543585 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/10497d10-a5e4-42e5-817c-1c41a2ef29e2-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-vnwqm\" (UID: \"10497d10-a5e4-42e5-817c-1c41a2ef29e2\") " pod="openstack/dnsmasq-dns-7f896c8c65-vnwqm" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.567244 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-9rd86"] Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.571856 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-9rd86" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.576732 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.597250 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-9rd86"] Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.645538 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/10497d10-a5e4-42e5-817c-1c41a2ef29e2-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-vnwqm\" (UID: \"10497d10-a5e4-42e5-817c-1c41a2ef29e2\") " pod="openstack/dnsmasq-dns-7f896c8c65-vnwqm" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.645593 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9077c720-7ad1-4f5c-a525-e28acd3dccb8-combined-ca-bundle\") pod \"ovn-controller-metrics-9rd86\" (UID: \"9077c720-7ad1-4f5c-a525-e28acd3dccb8\") " pod="openstack/ovn-controller-metrics-9rd86" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.645663 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9077c720-7ad1-4f5c-a525-e28acd3dccb8-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-9rd86\" (UID: \"9077c720-7ad1-4f5c-a525-e28acd3dccb8\") " pod="openstack/ovn-controller-metrics-9rd86" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.645690 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/9077c720-7ad1-4f5c-a525-e28acd3dccb8-ovs-rundir\") pod \"ovn-controller-metrics-9rd86\" (UID: \"9077c720-7ad1-4f5c-a525-e28acd3dccb8\") " pod="openstack/ovn-controller-metrics-9rd86" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.645707 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46qq9\" (UniqueName: \"kubernetes.io/projected/9077c720-7ad1-4f5c-a525-e28acd3dccb8-kube-api-access-46qq9\") pod \"ovn-controller-metrics-9rd86\" (UID: \"9077c720-7ad1-4f5c-a525-e28acd3dccb8\") " pod="openstack/ovn-controller-metrics-9rd86" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.645752 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxf5t\" (UniqueName: \"kubernetes.io/projected/10497d10-a5e4-42e5-817c-1c41a2ef29e2-kube-api-access-kxf5t\") pod \"dnsmasq-dns-7f896c8c65-vnwqm\" (UID: \"10497d10-a5e4-42e5-817c-1c41a2ef29e2\") " pod="openstack/dnsmasq-dns-7f896c8c65-vnwqm" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.645779 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/10497d10-a5e4-42e5-817c-1c41a2ef29e2-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-vnwqm\" (UID: \"10497d10-a5e4-42e5-817c-1c41a2ef29e2\") " pod="openstack/dnsmasq-dns-7f896c8c65-vnwqm" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.645797 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/9077c720-7ad1-4f5c-a525-e28acd3dccb8-ovn-rundir\") pod \"ovn-controller-metrics-9rd86\" (UID: \"9077c720-7ad1-4f5c-a525-e28acd3dccb8\") " pod="openstack/ovn-controller-metrics-9rd86" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.645823 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9077c720-7ad1-4f5c-a525-e28acd3dccb8-config\") pod \"ovn-controller-metrics-9rd86\" (UID: \"9077c720-7ad1-4f5c-a525-e28acd3dccb8\") " pod="openstack/ovn-controller-metrics-9rd86" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.645848 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10497d10-a5e4-42e5-817c-1c41a2ef29e2-config\") pod \"dnsmasq-dns-7f896c8c65-vnwqm\" (UID: \"10497d10-a5e4-42e5-817c-1c41a2ef29e2\") " pod="openstack/dnsmasq-dns-7f896c8c65-vnwqm" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.646729 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10497d10-a5e4-42e5-817c-1c41a2ef29e2-config\") pod \"dnsmasq-dns-7f896c8c65-vnwqm\" (UID: \"10497d10-a5e4-42e5-817c-1c41a2ef29e2\") " pod="openstack/dnsmasq-dns-7f896c8c65-vnwqm" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.647246 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/10497d10-a5e4-42e5-817c-1c41a2ef29e2-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-vnwqm\" (UID: \"10497d10-a5e4-42e5-817c-1c41a2ef29e2\") " pod="openstack/dnsmasq-dns-7f896c8c65-vnwqm" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.648040 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/10497d10-a5e4-42e5-817c-1c41a2ef29e2-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-vnwqm\" (UID: \"10497d10-a5e4-42e5-817c-1c41a2ef29e2\") " pod="openstack/dnsmasq-dns-7f896c8c65-vnwqm" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.679076 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxf5t\" (UniqueName: \"kubernetes.io/projected/10497d10-a5e4-42e5-817c-1c41a2ef29e2-kube-api-access-kxf5t\") pod \"dnsmasq-dns-7f896c8c65-vnwqm\" (UID: \"10497d10-a5e4-42e5-817c-1c41a2ef29e2\") " pod="openstack/dnsmasq-dns-7f896c8c65-vnwqm" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.747401 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9077c720-7ad1-4f5c-a525-e28acd3dccb8-combined-ca-bundle\") pod \"ovn-controller-metrics-9rd86\" (UID: \"9077c720-7ad1-4f5c-a525-e28acd3dccb8\") " pod="openstack/ovn-controller-metrics-9rd86" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.747520 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9077c720-7ad1-4f5c-a525-e28acd3dccb8-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-9rd86\" (UID: \"9077c720-7ad1-4f5c-a525-e28acd3dccb8\") " pod="openstack/ovn-controller-metrics-9rd86" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.747562 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/9077c720-7ad1-4f5c-a525-e28acd3dccb8-ovs-rundir\") pod \"ovn-controller-metrics-9rd86\" (UID: \"9077c720-7ad1-4f5c-a525-e28acd3dccb8\") " pod="openstack/ovn-controller-metrics-9rd86" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.747586 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46qq9\" (UniqueName: \"kubernetes.io/projected/9077c720-7ad1-4f5c-a525-e28acd3dccb8-kube-api-access-46qq9\") pod \"ovn-controller-metrics-9rd86\" (UID: \"9077c720-7ad1-4f5c-a525-e28acd3dccb8\") " pod="openstack/ovn-controller-metrics-9rd86" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.747658 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/9077c720-7ad1-4f5c-a525-e28acd3dccb8-ovn-rundir\") pod \"ovn-controller-metrics-9rd86\" (UID: \"9077c720-7ad1-4f5c-a525-e28acd3dccb8\") " pod="openstack/ovn-controller-metrics-9rd86" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.747689 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9077c720-7ad1-4f5c-a525-e28acd3dccb8-config\") pod \"ovn-controller-metrics-9rd86\" (UID: \"9077c720-7ad1-4f5c-a525-e28acd3dccb8\") " pod="openstack/ovn-controller-metrics-9rd86" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.748528 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9077c720-7ad1-4f5c-a525-e28acd3dccb8-config\") pod \"ovn-controller-metrics-9rd86\" (UID: \"9077c720-7ad1-4f5c-a525-e28acd3dccb8\") " pod="openstack/ovn-controller-metrics-9rd86" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.749055 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/9077c720-7ad1-4f5c-a525-e28acd3dccb8-ovs-rundir\") pod \"ovn-controller-metrics-9rd86\" (UID: \"9077c720-7ad1-4f5c-a525-e28acd3dccb8\") " pod="openstack/ovn-controller-metrics-9rd86" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.749776 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/9077c720-7ad1-4f5c-a525-e28acd3dccb8-ovn-rundir\") pod \"ovn-controller-metrics-9rd86\" (UID: \"9077c720-7ad1-4f5c-a525-e28acd3dccb8\") " pod="openstack/ovn-controller-metrics-9rd86" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.753743 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9077c720-7ad1-4f5c-a525-e28acd3dccb8-combined-ca-bundle\") pod \"ovn-controller-metrics-9rd86\" (UID: \"9077c720-7ad1-4f5c-a525-e28acd3dccb8\") " pod="openstack/ovn-controller-metrics-9rd86" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.755400 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9077c720-7ad1-4f5c-a525-e28acd3dccb8-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-9rd86\" (UID: \"9077c720-7ad1-4f5c-a525-e28acd3dccb8\") " pod="openstack/ovn-controller-metrics-9rd86" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.776676 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46qq9\" (UniqueName: \"kubernetes.io/projected/9077c720-7ad1-4f5c-a525-e28acd3dccb8-kube-api-access-46qq9\") pod \"ovn-controller-metrics-9rd86\" (UID: \"9077c720-7ad1-4f5c-a525-e28acd3dccb8\") " pod="openstack/ovn-controller-metrics-9rd86" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.777276 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-vnwqm" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.832672 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-vnwqm"] Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.877095 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-fwqkk"] Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.889203 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-fwqkk" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.892430 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-9rd86" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.909526 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.924215 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-fwqkk"] Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.952702 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.955167 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.961221 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.962474 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.962627 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.962757 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-dg5sz" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.965545 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6f7349dd-e05d-4c2f-ab31-5d328994e551-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-fwqkk\" (UID: \"6f7349dd-e05d-4c2f-ab31-5d328994e551\") " pod="openstack/dnsmasq-dns-86db49b7ff-fwqkk" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.965715 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6f7349dd-e05d-4c2f-ab31-5d328994e551-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-fwqkk\" (UID: \"6f7349dd-e05d-4c2f-ab31-5d328994e551\") " pod="openstack/dnsmasq-dns-86db49b7ff-fwqkk" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.965801 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7hzx\" (UniqueName: \"kubernetes.io/projected/6f7349dd-e05d-4c2f-ab31-5d328994e551-kube-api-access-d7hzx\") pod \"dnsmasq-dns-86db49b7ff-fwqkk\" (UID: \"6f7349dd-e05d-4c2f-ab31-5d328994e551\") " pod="openstack/dnsmasq-dns-86db49b7ff-fwqkk" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.965959 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f7349dd-e05d-4c2f-ab31-5d328994e551-config\") pod \"dnsmasq-dns-86db49b7ff-fwqkk\" (UID: \"6f7349dd-e05d-4c2f-ab31-5d328994e551\") " pod="openstack/dnsmasq-dns-86db49b7ff-fwqkk" Dec 06 06:15:18 crc kubenswrapper[4809]: I1206 06:15:18.965991 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f7349dd-e05d-4c2f-ab31-5d328994e551-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-fwqkk\" (UID: \"6f7349dd-e05d-4c2f-ab31-5d328994e551\") " pod="openstack/dnsmasq-dns-86db49b7ff-fwqkk" Dec 06 06:15:19 crc kubenswrapper[4809]: I1206 06:15:19.000021 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 06 06:15:19 crc kubenswrapper[4809]: I1206 06:15:19.071296 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1360c733-da74-4d37-8842-75050a356334-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"1360c733-da74-4d37-8842-75050a356334\") " pod="openstack/ovn-northd-0" Dec 06 06:15:19 crc kubenswrapper[4809]: I1206 06:15:19.071374 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7hzx\" (UniqueName: \"kubernetes.io/projected/6f7349dd-e05d-4c2f-ab31-5d328994e551-kube-api-access-d7hzx\") pod \"dnsmasq-dns-86db49b7ff-fwqkk\" (UID: \"6f7349dd-e05d-4c2f-ab31-5d328994e551\") " pod="openstack/dnsmasq-dns-86db49b7ff-fwqkk" Dec 06 06:15:19 crc kubenswrapper[4809]: I1206 06:15:19.071416 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/1360c733-da74-4d37-8842-75050a356334-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"1360c733-da74-4d37-8842-75050a356334\") " pod="openstack/ovn-northd-0" Dec 06 06:15:19 crc kubenswrapper[4809]: I1206 06:15:19.071455 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1360c733-da74-4d37-8842-75050a356334-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"1360c733-da74-4d37-8842-75050a356334\") " pod="openstack/ovn-northd-0" Dec 06 06:15:19 crc kubenswrapper[4809]: I1206 06:15:19.071525 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f7349dd-e05d-4c2f-ab31-5d328994e551-config\") pod \"dnsmasq-dns-86db49b7ff-fwqkk\" (UID: \"6f7349dd-e05d-4c2f-ab31-5d328994e551\") " pod="openstack/dnsmasq-dns-86db49b7ff-fwqkk" Dec 06 06:15:19 crc kubenswrapper[4809]: I1206 06:15:19.071563 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f7349dd-e05d-4c2f-ab31-5d328994e551-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-fwqkk\" (UID: \"6f7349dd-e05d-4c2f-ab31-5d328994e551\") " pod="openstack/dnsmasq-dns-86db49b7ff-fwqkk" Dec 06 06:15:19 crc kubenswrapper[4809]: I1206 06:15:19.071593 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1360c733-da74-4d37-8842-75050a356334-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"1360c733-da74-4d37-8842-75050a356334\") " pod="openstack/ovn-northd-0" Dec 06 06:15:19 crc kubenswrapper[4809]: I1206 06:15:19.071695 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6f7349dd-e05d-4c2f-ab31-5d328994e551-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-fwqkk\" (UID: \"6f7349dd-e05d-4c2f-ab31-5d328994e551\") " pod="openstack/dnsmasq-dns-86db49b7ff-fwqkk" Dec 06 06:15:19 crc kubenswrapper[4809]: I1206 06:15:19.071761 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1360c733-da74-4d37-8842-75050a356334-scripts\") pod \"ovn-northd-0\" (UID: \"1360c733-da74-4d37-8842-75050a356334\") " pod="openstack/ovn-northd-0" Dec 06 06:15:19 crc kubenswrapper[4809]: I1206 06:15:19.071794 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1360c733-da74-4d37-8842-75050a356334-config\") pod \"ovn-northd-0\" (UID: \"1360c733-da74-4d37-8842-75050a356334\") " pod="openstack/ovn-northd-0" Dec 06 06:15:19 crc kubenswrapper[4809]: I1206 06:15:19.071859 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pblvk\" (UniqueName: \"kubernetes.io/projected/1360c733-da74-4d37-8842-75050a356334-kube-api-access-pblvk\") pod \"ovn-northd-0\" (UID: \"1360c733-da74-4d37-8842-75050a356334\") " pod="openstack/ovn-northd-0" Dec 06 06:15:19 crc kubenswrapper[4809]: I1206 06:15:19.071896 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6f7349dd-e05d-4c2f-ab31-5d328994e551-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-fwqkk\" (UID: \"6f7349dd-e05d-4c2f-ab31-5d328994e551\") " pod="openstack/dnsmasq-dns-86db49b7ff-fwqkk" Dec 06 06:15:19 crc kubenswrapper[4809]: I1206 06:15:19.073040 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6f7349dd-e05d-4c2f-ab31-5d328994e551-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-fwqkk\" (UID: \"6f7349dd-e05d-4c2f-ab31-5d328994e551\") " pod="openstack/dnsmasq-dns-86db49b7ff-fwqkk" Dec 06 06:15:19 crc kubenswrapper[4809]: I1206 06:15:19.073052 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6f7349dd-e05d-4c2f-ab31-5d328994e551-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-fwqkk\" (UID: \"6f7349dd-e05d-4c2f-ab31-5d328994e551\") " pod="openstack/dnsmasq-dns-86db49b7ff-fwqkk" Dec 06 06:15:19 crc kubenswrapper[4809]: I1206 06:15:19.074771 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f7349dd-e05d-4c2f-ab31-5d328994e551-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-fwqkk\" (UID: \"6f7349dd-e05d-4c2f-ab31-5d328994e551\") " pod="openstack/dnsmasq-dns-86db49b7ff-fwqkk" Dec 06 06:15:19 crc kubenswrapper[4809]: I1206 06:15:19.075164 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f7349dd-e05d-4c2f-ab31-5d328994e551-config\") pod \"dnsmasq-dns-86db49b7ff-fwqkk\" (UID: \"6f7349dd-e05d-4c2f-ab31-5d328994e551\") " pod="openstack/dnsmasq-dns-86db49b7ff-fwqkk" Dec 06 06:15:19 crc kubenswrapper[4809]: I1206 06:15:19.097058 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7hzx\" (UniqueName: \"kubernetes.io/projected/6f7349dd-e05d-4c2f-ab31-5d328994e551-kube-api-access-d7hzx\") pod \"dnsmasq-dns-86db49b7ff-fwqkk\" (UID: \"6f7349dd-e05d-4c2f-ab31-5d328994e551\") " pod="openstack/dnsmasq-dns-86db49b7ff-fwqkk" Dec 06 06:15:19 crc kubenswrapper[4809]: I1206 06:15:19.175057 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1360c733-da74-4d37-8842-75050a356334-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"1360c733-da74-4d37-8842-75050a356334\") " pod="openstack/ovn-northd-0" Dec 06 06:15:19 crc kubenswrapper[4809]: I1206 06:15:19.175339 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1360c733-da74-4d37-8842-75050a356334-scripts\") pod \"ovn-northd-0\" (UID: \"1360c733-da74-4d37-8842-75050a356334\") " pod="openstack/ovn-northd-0" Dec 06 06:15:19 crc kubenswrapper[4809]: I1206 06:15:19.175366 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1360c733-da74-4d37-8842-75050a356334-config\") pod \"ovn-northd-0\" (UID: \"1360c733-da74-4d37-8842-75050a356334\") " pod="openstack/ovn-northd-0" Dec 06 06:15:19 crc kubenswrapper[4809]: I1206 06:15:19.175398 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pblvk\" (UniqueName: \"kubernetes.io/projected/1360c733-da74-4d37-8842-75050a356334-kube-api-access-pblvk\") pod \"ovn-northd-0\" (UID: \"1360c733-da74-4d37-8842-75050a356334\") " pod="openstack/ovn-northd-0" Dec 06 06:15:19 crc kubenswrapper[4809]: I1206 06:15:19.175432 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1360c733-da74-4d37-8842-75050a356334-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"1360c733-da74-4d37-8842-75050a356334\") " pod="openstack/ovn-northd-0" Dec 06 06:15:19 crc kubenswrapper[4809]: I1206 06:15:19.175462 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/1360c733-da74-4d37-8842-75050a356334-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"1360c733-da74-4d37-8842-75050a356334\") " pod="openstack/ovn-northd-0" Dec 06 06:15:19 crc kubenswrapper[4809]: I1206 06:15:19.175487 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1360c733-da74-4d37-8842-75050a356334-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"1360c733-da74-4d37-8842-75050a356334\") " pod="openstack/ovn-northd-0" Dec 06 06:15:19 crc kubenswrapper[4809]: I1206 06:15:19.175972 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1360c733-da74-4d37-8842-75050a356334-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"1360c733-da74-4d37-8842-75050a356334\") " pod="openstack/ovn-northd-0" Dec 06 06:15:19 crc kubenswrapper[4809]: I1206 06:15:19.176678 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1360c733-da74-4d37-8842-75050a356334-scripts\") pod \"ovn-northd-0\" (UID: \"1360c733-da74-4d37-8842-75050a356334\") " pod="openstack/ovn-northd-0" Dec 06 06:15:19 crc kubenswrapper[4809]: I1206 06:15:19.177220 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1360c733-da74-4d37-8842-75050a356334-config\") pod \"ovn-northd-0\" (UID: \"1360c733-da74-4d37-8842-75050a356334\") " pod="openstack/ovn-northd-0" Dec 06 06:15:19 crc kubenswrapper[4809]: I1206 06:15:19.190503 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1360c733-da74-4d37-8842-75050a356334-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"1360c733-da74-4d37-8842-75050a356334\") " pod="openstack/ovn-northd-0" Dec 06 06:15:19 crc kubenswrapper[4809]: I1206 06:15:19.192713 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e7056938-6e36-4631-adfb-01f116eba651","Type":"ContainerStarted","Data":"c5e7932399fd9c59936d31395b31c6f5cc98b66aad28f9a1faffee5aa4d0028c"} Dec 06 06:15:19 crc kubenswrapper[4809]: I1206 06:15:19.193773 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/1360c733-da74-4d37-8842-75050a356334-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"1360c733-da74-4d37-8842-75050a356334\") " pod="openstack/ovn-northd-0" Dec 06 06:15:19 crc kubenswrapper[4809]: I1206 06:15:19.196612 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1360c733-da74-4d37-8842-75050a356334-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"1360c733-da74-4d37-8842-75050a356334\") " pod="openstack/ovn-northd-0" Dec 06 06:15:19 crc kubenswrapper[4809]: I1206 06:15:19.208697 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"7c596e14-b96f-4e5f-a8ac-3dc66fa0276d","Type":"ContainerStarted","Data":"a33b2b043d42437efe270d1e3ec31c3d840ad9b765f7e171518bdd9f6ff5ad80"} Dec 06 06:15:19 crc kubenswrapper[4809]: I1206 06:15:19.213013 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pblvk\" (UniqueName: \"kubernetes.io/projected/1360c733-da74-4d37-8842-75050a356334-kube-api-access-pblvk\") pod \"ovn-northd-0\" (UID: \"1360c733-da74-4d37-8842-75050a356334\") " pod="openstack/ovn-northd-0" Dec 06 06:15:19 crc kubenswrapper[4809]: I1206 06:15:19.235847 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=33.963190528 podStartE2EDuration="44.235822498s" podCreationTimestamp="2025-12-06 06:14:35 +0000 UTC" firstStartedPulling="2025-12-06 06:14:54.491004642 +0000 UTC m=+1419.379987584" lastFinishedPulling="2025-12-06 06:15:04.763636612 +0000 UTC m=+1429.652619554" observedRunningTime="2025-12-06 06:15:19.223404713 +0000 UTC m=+1444.112387655" watchObservedRunningTime="2025-12-06 06:15:19.235822498 +0000 UTC m=+1444.124805440" Dec 06 06:15:19 crc kubenswrapper[4809]: I1206 06:15:19.253567 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-fwqkk" Dec 06 06:15:19 crc kubenswrapper[4809]: I1206 06:15:19.268099 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=35.125520236 podStartE2EDuration="45.268073498s" podCreationTimestamp="2025-12-06 06:14:34 +0000 UTC" firstStartedPulling="2025-12-06 06:14:56.295464498 +0000 UTC m=+1421.184447440" lastFinishedPulling="2025-12-06 06:15:06.43801776 +0000 UTC m=+1431.327000702" observedRunningTime="2025-12-06 06:15:19.252533619 +0000 UTC m=+1444.141516561" watchObservedRunningTime="2025-12-06 06:15:19.268073498 +0000 UTC m=+1444.157056440" Dec 06 06:15:19 crc kubenswrapper[4809]: I1206 06:15:19.328132 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 06 06:15:19 crc kubenswrapper[4809]: I1206 06:15:19.671329 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 06 06:15:19 crc kubenswrapper[4809]: I1206 06:15:19.697870 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-fwqkk"] Dec 06 06:15:19 crc kubenswrapper[4809]: I1206 06:15:19.722383 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-vnwqm"] Dec 06 06:15:19 crc kubenswrapper[4809]: I1206 06:15:19.737100 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-r89dg"] Dec 06 06:15:19 crc kubenswrapper[4809]: I1206 06:15:19.738960 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-r89dg" Dec 06 06:15:19 crc kubenswrapper[4809]: I1206 06:15:19.754142 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-r89dg"] Dec 06 06:15:19 crc kubenswrapper[4809]: I1206 06:15:19.778152 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-9rd86"] Dec 06 06:15:20 crc kubenswrapper[4809]: I1206 06:15:19.907767 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c24a4f1-4cfb-4aae-ac2e-fa7680751c26-dns-svc\") pod \"dnsmasq-dns-698758b865-r89dg\" (UID: \"0c24a4f1-4cfb-4aae-ac2e-fa7680751c26\") " pod="openstack/dnsmasq-dns-698758b865-r89dg" Dec 06 06:15:20 crc kubenswrapper[4809]: I1206 06:15:19.908189 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c24a4f1-4cfb-4aae-ac2e-fa7680751c26-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-r89dg\" (UID: \"0c24a4f1-4cfb-4aae-ac2e-fa7680751c26\") " pod="openstack/dnsmasq-dns-698758b865-r89dg" Dec 06 06:15:20 crc kubenswrapper[4809]: I1206 06:15:19.908266 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c24a4f1-4cfb-4aae-ac2e-fa7680751c26-config\") pod \"dnsmasq-dns-698758b865-r89dg\" (UID: \"0c24a4f1-4cfb-4aae-ac2e-fa7680751c26\") " pod="openstack/dnsmasq-dns-698758b865-r89dg" Dec 06 06:15:20 crc kubenswrapper[4809]: I1206 06:15:19.908394 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5c2k\" (UniqueName: \"kubernetes.io/projected/0c24a4f1-4cfb-4aae-ac2e-fa7680751c26-kube-api-access-v5c2k\") pod \"dnsmasq-dns-698758b865-r89dg\" (UID: \"0c24a4f1-4cfb-4aae-ac2e-fa7680751c26\") " pod="openstack/dnsmasq-dns-698758b865-r89dg" Dec 06 06:15:20 crc kubenswrapper[4809]: I1206 06:15:19.908427 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c24a4f1-4cfb-4aae-ac2e-fa7680751c26-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-r89dg\" (UID: \"0c24a4f1-4cfb-4aae-ac2e-fa7680751c26\") " pod="openstack/dnsmasq-dns-698758b865-r89dg" Dec 06 06:15:20 crc kubenswrapper[4809]: I1206 06:15:20.010347 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5c2k\" (UniqueName: \"kubernetes.io/projected/0c24a4f1-4cfb-4aae-ac2e-fa7680751c26-kube-api-access-v5c2k\") pod \"dnsmasq-dns-698758b865-r89dg\" (UID: \"0c24a4f1-4cfb-4aae-ac2e-fa7680751c26\") " pod="openstack/dnsmasq-dns-698758b865-r89dg" Dec 06 06:15:20 crc kubenswrapper[4809]: I1206 06:15:20.010388 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c24a4f1-4cfb-4aae-ac2e-fa7680751c26-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-r89dg\" (UID: \"0c24a4f1-4cfb-4aae-ac2e-fa7680751c26\") " pod="openstack/dnsmasq-dns-698758b865-r89dg" Dec 06 06:15:20 crc kubenswrapper[4809]: I1206 06:15:20.010470 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c24a4f1-4cfb-4aae-ac2e-fa7680751c26-dns-svc\") pod \"dnsmasq-dns-698758b865-r89dg\" (UID: \"0c24a4f1-4cfb-4aae-ac2e-fa7680751c26\") " pod="openstack/dnsmasq-dns-698758b865-r89dg" Dec 06 06:15:20 crc kubenswrapper[4809]: I1206 06:15:20.010498 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c24a4f1-4cfb-4aae-ac2e-fa7680751c26-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-r89dg\" (UID: \"0c24a4f1-4cfb-4aae-ac2e-fa7680751c26\") " pod="openstack/dnsmasq-dns-698758b865-r89dg" Dec 06 06:15:20 crc kubenswrapper[4809]: I1206 06:15:20.010550 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c24a4f1-4cfb-4aae-ac2e-fa7680751c26-config\") pod \"dnsmasq-dns-698758b865-r89dg\" (UID: \"0c24a4f1-4cfb-4aae-ac2e-fa7680751c26\") " pod="openstack/dnsmasq-dns-698758b865-r89dg" Dec 06 06:15:20 crc kubenswrapper[4809]: I1206 06:15:20.012089 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c24a4f1-4cfb-4aae-ac2e-fa7680751c26-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-r89dg\" (UID: \"0c24a4f1-4cfb-4aae-ac2e-fa7680751c26\") " pod="openstack/dnsmasq-dns-698758b865-r89dg" Dec 06 06:15:20 crc kubenswrapper[4809]: I1206 06:15:20.012278 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c24a4f1-4cfb-4aae-ac2e-fa7680751c26-dns-svc\") pod \"dnsmasq-dns-698758b865-r89dg\" (UID: \"0c24a4f1-4cfb-4aae-ac2e-fa7680751c26\") " pod="openstack/dnsmasq-dns-698758b865-r89dg" Dec 06 06:15:20 crc kubenswrapper[4809]: I1206 06:15:20.012440 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c24a4f1-4cfb-4aae-ac2e-fa7680751c26-config\") pod \"dnsmasq-dns-698758b865-r89dg\" (UID: \"0c24a4f1-4cfb-4aae-ac2e-fa7680751c26\") " pod="openstack/dnsmasq-dns-698758b865-r89dg" Dec 06 06:15:20 crc kubenswrapper[4809]: I1206 06:15:20.013596 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c24a4f1-4cfb-4aae-ac2e-fa7680751c26-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-r89dg\" (UID: \"0c24a4f1-4cfb-4aae-ac2e-fa7680751c26\") " pod="openstack/dnsmasq-dns-698758b865-r89dg" Dec 06 06:15:20 crc kubenswrapper[4809]: I1206 06:15:20.032544 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5c2k\" (UniqueName: \"kubernetes.io/projected/0c24a4f1-4cfb-4aae-ac2e-fa7680751c26-kube-api-access-v5c2k\") pod \"dnsmasq-dns-698758b865-r89dg\" (UID: \"0c24a4f1-4cfb-4aae-ac2e-fa7680751c26\") " pod="openstack/dnsmasq-dns-698758b865-r89dg" Dec 06 06:15:20 crc kubenswrapper[4809]: I1206 06:15:20.062031 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-fwqkk"] Dec 06 06:15:20 crc kubenswrapper[4809]: I1206 06:15:20.093115 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-r89dg" Dec 06 06:15:20 crc kubenswrapper[4809]: I1206 06:15:20.237306 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 06 06:15:20 crc kubenswrapper[4809]: W1206 06:15:20.259194 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1360c733_da74_4d37_8842_75050a356334.slice/crio-a6425dc7a09ecc3132a8aedd93cfc34d70c8d1d2e9d5a92cf43355ec7da30136 WatchSource:0}: Error finding container a6425dc7a09ecc3132a8aedd93cfc34d70c8d1d2e9d5a92cf43355ec7da30136: Status 404 returned error can't find the container with id a6425dc7a09ecc3132a8aedd93cfc34d70c8d1d2e9d5a92cf43355ec7da30136 Dec 06 06:15:20 crc kubenswrapper[4809]: I1206 06:15:20.269365 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-fwqkk" event={"ID":"6f7349dd-e05d-4c2f-ab31-5d328994e551","Type":"ContainerStarted","Data":"2bc4b6b62bc2123abff4cbfc3a52da6288fe65bd7969edb93eae017fa6693ee1"} Dec 06 06:15:20 crc kubenswrapper[4809]: I1206 06:15:20.286748 4809 generic.go:334] "Generic (PLEG): container finished" podID="10497d10-a5e4-42e5-817c-1c41a2ef29e2" containerID="283c6759135798c91fc633b457cc32279f83b57496114cf795b2afb02fdc6053" exitCode=0 Dec 06 06:15:20 crc kubenswrapper[4809]: I1206 06:15:20.286821 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-vnwqm" event={"ID":"10497d10-a5e4-42e5-817c-1c41a2ef29e2","Type":"ContainerDied","Data":"283c6759135798c91fc633b457cc32279f83b57496114cf795b2afb02fdc6053"} Dec 06 06:15:20 crc kubenswrapper[4809]: I1206 06:15:20.286849 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-vnwqm" event={"ID":"10497d10-a5e4-42e5-817c-1c41a2ef29e2","Type":"ContainerStarted","Data":"949f6491cca00a9a0b232b20810f164d9b46d313b6e707f43cda0905ddf14a5e"} Dec 06 06:15:20 crc kubenswrapper[4809]: I1206 06:15:20.289257 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-9rd86" event={"ID":"9077c720-7ad1-4f5c-a525-e28acd3dccb8","Type":"ContainerStarted","Data":"9747c897c110db7a8a06810e73a26c26d7b39f5b44f010a2bef4a39f2ae23edc"} Dec 06 06:15:20 crc kubenswrapper[4809]: I1206 06:15:20.289617 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-9rd86" event={"ID":"9077c720-7ad1-4f5c-a525-e28acd3dccb8","Type":"ContainerStarted","Data":"ddb99ee3ae57321e30829439b3a12e5f942882ad66ad063b99e4457e7d57cbc4"} Dec 06 06:15:20 crc kubenswrapper[4809]: I1206 06:15:20.299359 4809 generic.go:334] "Generic (PLEG): container finished" podID="f6536b25-1cc1-4f74-bf46-613965ddf185" containerID="df03251430e15c56ceb99e1135791e64f339c2265989799fec5b3d7450799a7d" exitCode=0 Dec 06 06:15:20 crc kubenswrapper[4809]: I1206 06:15:20.299660 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f6536b25-1cc1-4f74-bf46-613965ddf185","Type":"ContainerDied","Data":"df03251430e15c56ceb99e1135791e64f339c2265989799fec5b3d7450799a7d"} Dec 06 06:15:20 crc kubenswrapper[4809]: I1206 06:15:20.365677 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-9rd86" podStartSLOduration=2.365653543 podStartE2EDuration="2.365653543s" podCreationTimestamp="2025-12-06 06:15:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:15:20.359543388 +0000 UTC m=+1445.248526350" watchObservedRunningTime="2025-12-06 06:15:20.365653543 +0000 UTC m=+1445.254636485" Dec 06 06:15:20 crc kubenswrapper[4809]: I1206 06:15:20.759082 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 06 06:15:20 crc kubenswrapper[4809]: I1206 06:15:20.776970 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 06 06:15:20 crc kubenswrapper[4809]: I1206 06:15:20.781573 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 06 06:15:20 crc kubenswrapper[4809]: I1206 06:15:20.781596 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 06 06:15:20 crc kubenswrapper[4809]: I1206 06:15:20.781630 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 06 06:15:20 crc kubenswrapper[4809]: I1206 06:15:20.781793 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-7fl4s" Dec 06 06:15:20 crc kubenswrapper[4809]: I1206 06:15:20.789005 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 06 06:15:20 crc kubenswrapper[4809]: I1206 06:15:20.952271 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/cfd13e5d-ca28-4061-8ca5-2c4566332bf0-lock\") pod \"swift-storage-0\" (UID: \"cfd13e5d-ca28-4061-8ca5-2c4566332bf0\") " pod="openstack/swift-storage-0" Dec 06 06:15:20 crc kubenswrapper[4809]: I1206 06:15:20.952448 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"cfd13e5d-ca28-4061-8ca5-2c4566332bf0\") " pod="openstack/swift-storage-0" Dec 06 06:15:20 crc kubenswrapper[4809]: I1206 06:15:20.952514 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cfd13e5d-ca28-4061-8ca5-2c4566332bf0-etc-swift\") pod \"swift-storage-0\" (UID: \"cfd13e5d-ca28-4061-8ca5-2c4566332bf0\") " pod="openstack/swift-storage-0" Dec 06 06:15:20 crc kubenswrapper[4809]: I1206 06:15:20.952545 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56sqz\" (UniqueName: \"kubernetes.io/projected/cfd13e5d-ca28-4061-8ca5-2c4566332bf0-kube-api-access-56sqz\") pod \"swift-storage-0\" (UID: \"cfd13e5d-ca28-4061-8ca5-2c4566332bf0\") " pod="openstack/swift-storage-0" Dec 06 06:15:20 crc kubenswrapper[4809]: I1206 06:15:20.952578 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/cfd13e5d-ca28-4061-8ca5-2c4566332bf0-cache\") pod \"swift-storage-0\" (UID: \"cfd13e5d-ca28-4061-8ca5-2c4566332bf0\") " pod="openstack/swift-storage-0" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.058098 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cfd13e5d-ca28-4061-8ca5-2c4566332bf0-etc-swift\") pod \"swift-storage-0\" (UID: \"cfd13e5d-ca28-4061-8ca5-2c4566332bf0\") " pod="openstack/swift-storage-0" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.058175 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56sqz\" (UniqueName: \"kubernetes.io/projected/cfd13e5d-ca28-4061-8ca5-2c4566332bf0-kube-api-access-56sqz\") pod \"swift-storage-0\" (UID: \"cfd13e5d-ca28-4061-8ca5-2c4566332bf0\") " pod="openstack/swift-storage-0" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.058237 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/cfd13e5d-ca28-4061-8ca5-2c4566332bf0-cache\") pod \"swift-storage-0\" (UID: \"cfd13e5d-ca28-4061-8ca5-2c4566332bf0\") " pod="openstack/swift-storage-0" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.058427 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/cfd13e5d-ca28-4061-8ca5-2c4566332bf0-lock\") pod \"swift-storage-0\" (UID: \"cfd13e5d-ca28-4061-8ca5-2c4566332bf0\") " pod="openstack/swift-storage-0" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.058728 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"cfd13e5d-ca28-4061-8ca5-2c4566332bf0\") " pod="openstack/swift-storage-0" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.060461 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"cfd13e5d-ca28-4061-8ca5-2c4566332bf0\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/swift-storage-0" Dec 06 06:15:21 crc kubenswrapper[4809]: E1206 06:15:21.065191 4809 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 06 06:15:21 crc kubenswrapper[4809]: E1206 06:15:21.065220 4809 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 06 06:15:21 crc kubenswrapper[4809]: E1206 06:15:21.065260 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/cfd13e5d-ca28-4061-8ca5-2c4566332bf0-etc-swift podName:cfd13e5d-ca28-4061-8ca5-2c4566332bf0 nodeName:}" failed. No retries permitted until 2025-12-06 06:15:21.565245017 +0000 UTC m=+1446.454227959 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/cfd13e5d-ca28-4061-8ca5-2c4566332bf0-etc-swift") pod "swift-storage-0" (UID: "cfd13e5d-ca28-4061-8ca5-2c4566332bf0") : configmap "swift-ring-files" not found Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.070601 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/cfd13e5d-ca28-4061-8ca5-2c4566332bf0-lock\") pod \"swift-storage-0\" (UID: \"cfd13e5d-ca28-4061-8ca5-2c4566332bf0\") " pod="openstack/swift-storage-0" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.076225 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/cfd13e5d-ca28-4061-8ca5-2c4566332bf0-cache\") pod \"swift-storage-0\" (UID: \"cfd13e5d-ca28-4061-8ca5-2c4566332bf0\") " pod="openstack/swift-storage-0" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.092290 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56sqz\" (UniqueName: \"kubernetes.io/projected/cfd13e5d-ca28-4061-8ca5-2c4566332bf0-kube-api-access-56sqz\") pod \"swift-storage-0\" (UID: \"cfd13e5d-ca28-4061-8ca5-2c4566332bf0\") " pod="openstack/swift-storage-0" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.103422 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"cfd13e5d-ca28-4061-8ca5-2c4566332bf0\") " pod="openstack/swift-storage-0" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.267994 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-vnwqm" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.269897 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-6xhch"] Dec 06 06:15:21 crc kubenswrapper[4809]: E1206 06:15:21.270874 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10497d10-a5e4-42e5-817c-1c41a2ef29e2" containerName="init" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.270897 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="10497d10-a5e4-42e5-817c-1c41a2ef29e2" containerName="init" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.271308 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="10497d10-a5e4-42e5-817c-1c41a2ef29e2" containerName="init" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.272496 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-6xhch" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.277420 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.277727 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.282380 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.310629 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-6xhch"] Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.327597 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-vnwqm" event={"ID":"10497d10-a5e4-42e5-817c-1c41a2ef29e2","Type":"ContainerDied","Data":"949f6491cca00a9a0b232b20810f164d9b46d313b6e707f43cda0905ddf14a5e"} Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.327649 4809 scope.go:117] "RemoveContainer" containerID="283c6759135798c91fc633b457cc32279f83b57496114cf795b2afb02fdc6053" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.327747 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-vnwqm" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.331448 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-6xhch"] Dec 06 06:15:21 crc kubenswrapper[4809]: E1206 06:15:21.332032 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-xx9r7 ring-data-devices scripts swiftconf], unattached volumes=[], failed to process volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-xx9r7 ring-data-devices scripts swiftconf]: context canceled" pod="openstack/swift-ring-rebalance-6xhch" podUID="50214a1e-8cea-4c86-b61b-162e286f6b64" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.338914 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"1360c733-da74-4d37-8842-75050a356334","Type":"ContainerStarted","Data":"a6425dc7a09ecc3132a8aedd93cfc34d70c8d1d2e9d5a92cf43355ec7da30136"} Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.340422 4809 generic.go:334] "Generic (PLEG): container finished" podID="6f7349dd-e05d-4c2f-ab31-5d328994e551" containerID="c37154aecc0e7bca137b5be30758f485495f63810d501a337a570aea85ec5849" exitCode=0 Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.341622 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-fwqkk" event={"ID":"6f7349dd-e05d-4c2f-ab31-5d328994e551","Type":"ContainerDied","Data":"c37154aecc0e7bca137b5be30758f485495f63810d501a337a570aea85ec5849"} Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.342296 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-wgmjm"] Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.343764 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-wgmjm" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.358812 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-wgmjm"] Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.369692 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kxf5t\" (UniqueName: \"kubernetes.io/projected/10497d10-a5e4-42e5-817c-1c41a2ef29e2-kube-api-access-kxf5t\") pod \"10497d10-a5e4-42e5-817c-1c41a2ef29e2\" (UID: \"10497d10-a5e4-42e5-817c-1c41a2ef29e2\") " Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.369836 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/10497d10-a5e4-42e5-817c-1c41a2ef29e2-dns-svc\") pod \"10497d10-a5e4-42e5-817c-1c41a2ef29e2\" (UID: \"10497d10-a5e4-42e5-817c-1c41a2ef29e2\") " Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.369971 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10497d10-a5e4-42e5-817c-1c41a2ef29e2-config\") pod \"10497d10-a5e4-42e5-817c-1c41a2ef29e2\" (UID: \"10497d10-a5e4-42e5-817c-1c41a2ef29e2\") " Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.370037 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/10497d10-a5e4-42e5-817c-1c41a2ef29e2-ovsdbserver-sb\") pod \"10497d10-a5e4-42e5-817c-1c41a2ef29e2\" (UID: \"10497d10-a5e4-42e5-817c-1c41a2ef29e2\") " Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.377291 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10497d10-a5e4-42e5-817c-1c41a2ef29e2-kube-api-access-kxf5t" (OuterVolumeSpecName: "kube-api-access-kxf5t") pod "10497d10-a5e4-42e5-817c-1c41a2ef29e2" (UID: "10497d10-a5e4-42e5-817c-1c41a2ef29e2"). InnerVolumeSpecName "kube-api-access-kxf5t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.377468 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-r89dg"] Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.402012 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10497d10-a5e4-42e5-817c-1c41a2ef29e2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "10497d10-a5e4-42e5-817c-1c41a2ef29e2" (UID: "10497d10-a5e4-42e5-817c-1c41a2ef29e2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.413751 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10497d10-a5e4-42e5-817c-1c41a2ef29e2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "10497d10-a5e4-42e5-817c-1c41a2ef29e2" (UID: "10497d10-a5e4-42e5-817c-1c41a2ef29e2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.419152 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10497d10-a5e4-42e5-817c-1c41a2ef29e2-config" (OuterVolumeSpecName: "config") pod "10497d10-a5e4-42e5-817c-1c41a2ef29e2" (UID: "10497d10-a5e4-42e5-817c-1c41a2ef29e2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.473133 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xx9r7\" (UniqueName: \"kubernetes.io/projected/50214a1e-8cea-4c86-b61b-162e286f6b64-kube-api-access-xx9r7\") pod \"swift-ring-rebalance-6xhch\" (UID: \"50214a1e-8cea-4c86-b61b-162e286f6b64\") " pod="openstack/swift-ring-rebalance-6xhch" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.473182 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jr2rt\" (UniqueName: \"kubernetes.io/projected/5f4f2323-f99a-4d85-8e0f-a0eb968de27b-kube-api-access-jr2rt\") pod \"swift-ring-rebalance-wgmjm\" (UID: \"5f4f2323-f99a-4d85-8e0f-a0eb968de27b\") " pod="openstack/swift-ring-rebalance-wgmjm" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.473267 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50214a1e-8cea-4c86-b61b-162e286f6b64-combined-ca-bundle\") pod \"swift-ring-rebalance-6xhch\" (UID: \"50214a1e-8cea-4c86-b61b-162e286f6b64\") " pod="openstack/swift-ring-rebalance-6xhch" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.473283 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f4f2323-f99a-4d85-8e0f-a0eb968de27b-combined-ca-bundle\") pod \"swift-ring-rebalance-wgmjm\" (UID: \"5f4f2323-f99a-4d85-8e0f-a0eb968de27b\") " pod="openstack/swift-ring-rebalance-wgmjm" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.473335 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/50214a1e-8cea-4c86-b61b-162e286f6b64-etc-swift\") pod \"swift-ring-rebalance-6xhch\" (UID: \"50214a1e-8cea-4c86-b61b-162e286f6b64\") " pod="openstack/swift-ring-rebalance-6xhch" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.473368 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5f4f2323-f99a-4d85-8e0f-a0eb968de27b-dispersionconf\") pod \"swift-ring-rebalance-wgmjm\" (UID: \"5f4f2323-f99a-4d85-8e0f-a0eb968de27b\") " pod="openstack/swift-ring-rebalance-wgmjm" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.473411 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/50214a1e-8cea-4c86-b61b-162e286f6b64-dispersionconf\") pod \"swift-ring-rebalance-6xhch\" (UID: \"50214a1e-8cea-4c86-b61b-162e286f6b64\") " pod="openstack/swift-ring-rebalance-6xhch" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.473426 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5f4f2323-f99a-4d85-8e0f-a0eb968de27b-ring-data-devices\") pod \"swift-ring-rebalance-wgmjm\" (UID: \"5f4f2323-f99a-4d85-8e0f-a0eb968de27b\") " pod="openstack/swift-ring-rebalance-wgmjm" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.473446 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/50214a1e-8cea-4c86-b61b-162e286f6b64-swiftconf\") pod \"swift-ring-rebalance-6xhch\" (UID: \"50214a1e-8cea-4c86-b61b-162e286f6b64\") " pod="openstack/swift-ring-rebalance-6xhch" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.473485 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/50214a1e-8cea-4c86-b61b-162e286f6b64-ring-data-devices\") pod \"swift-ring-rebalance-6xhch\" (UID: \"50214a1e-8cea-4c86-b61b-162e286f6b64\") " pod="openstack/swift-ring-rebalance-6xhch" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.474281 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/50214a1e-8cea-4c86-b61b-162e286f6b64-scripts\") pod \"swift-ring-rebalance-6xhch\" (UID: \"50214a1e-8cea-4c86-b61b-162e286f6b64\") " pod="openstack/swift-ring-rebalance-6xhch" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.474330 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5f4f2323-f99a-4d85-8e0f-a0eb968de27b-etc-swift\") pod \"swift-ring-rebalance-wgmjm\" (UID: \"5f4f2323-f99a-4d85-8e0f-a0eb968de27b\") " pod="openstack/swift-ring-rebalance-wgmjm" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.474355 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5f4f2323-f99a-4d85-8e0f-a0eb968de27b-swiftconf\") pod \"swift-ring-rebalance-wgmjm\" (UID: \"5f4f2323-f99a-4d85-8e0f-a0eb968de27b\") " pod="openstack/swift-ring-rebalance-wgmjm" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.474396 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5f4f2323-f99a-4d85-8e0f-a0eb968de27b-scripts\") pod \"swift-ring-rebalance-wgmjm\" (UID: \"5f4f2323-f99a-4d85-8e0f-a0eb968de27b\") " pod="openstack/swift-ring-rebalance-wgmjm" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.474478 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/10497d10-a5e4-42e5-817c-1c41a2ef29e2-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.474490 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10497d10-a5e4-42e5-817c-1c41a2ef29e2-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.474499 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/10497d10-a5e4-42e5-817c-1c41a2ef29e2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.474510 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kxf5t\" (UniqueName: \"kubernetes.io/projected/10497d10-a5e4-42e5-817c-1c41a2ef29e2-kube-api-access-kxf5t\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.576391 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xx9r7\" (UniqueName: \"kubernetes.io/projected/50214a1e-8cea-4c86-b61b-162e286f6b64-kube-api-access-xx9r7\") pod \"swift-ring-rebalance-6xhch\" (UID: \"50214a1e-8cea-4c86-b61b-162e286f6b64\") " pod="openstack/swift-ring-rebalance-6xhch" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.576445 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jr2rt\" (UniqueName: \"kubernetes.io/projected/5f4f2323-f99a-4d85-8e0f-a0eb968de27b-kube-api-access-jr2rt\") pod \"swift-ring-rebalance-wgmjm\" (UID: \"5f4f2323-f99a-4d85-8e0f-a0eb968de27b\") " pod="openstack/swift-ring-rebalance-wgmjm" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.576503 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50214a1e-8cea-4c86-b61b-162e286f6b64-combined-ca-bundle\") pod \"swift-ring-rebalance-6xhch\" (UID: \"50214a1e-8cea-4c86-b61b-162e286f6b64\") " pod="openstack/swift-ring-rebalance-6xhch" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.576526 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f4f2323-f99a-4d85-8e0f-a0eb968de27b-combined-ca-bundle\") pod \"swift-ring-rebalance-wgmjm\" (UID: \"5f4f2323-f99a-4d85-8e0f-a0eb968de27b\") " pod="openstack/swift-ring-rebalance-wgmjm" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.576588 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/50214a1e-8cea-4c86-b61b-162e286f6b64-etc-swift\") pod \"swift-ring-rebalance-6xhch\" (UID: \"50214a1e-8cea-4c86-b61b-162e286f6b64\") " pod="openstack/swift-ring-rebalance-6xhch" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.576636 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5f4f2323-f99a-4d85-8e0f-a0eb968de27b-dispersionconf\") pod \"swift-ring-rebalance-wgmjm\" (UID: \"5f4f2323-f99a-4d85-8e0f-a0eb968de27b\") " pod="openstack/swift-ring-rebalance-wgmjm" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.576678 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/50214a1e-8cea-4c86-b61b-162e286f6b64-dispersionconf\") pod \"swift-ring-rebalance-6xhch\" (UID: \"50214a1e-8cea-4c86-b61b-162e286f6b64\") " pod="openstack/swift-ring-rebalance-6xhch" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.576702 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5f4f2323-f99a-4d85-8e0f-a0eb968de27b-ring-data-devices\") pod \"swift-ring-rebalance-wgmjm\" (UID: \"5f4f2323-f99a-4d85-8e0f-a0eb968de27b\") " pod="openstack/swift-ring-rebalance-wgmjm" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.576729 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/50214a1e-8cea-4c86-b61b-162e286f6b64-swiftconf\") pod \"swift-ring-rebalance-6xhch\" (UID: \"50214a1e-8cea-4c86-b61b-162e286f6b64\") " pod="openstack/swift-ring-rebalance-6xhch" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.576791 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/50214a1e-8cea-4c86-b61b-162e286f6b64-ring-data-devices\") pod \"swift-ring-rebalance-6xhch\" (UID: \"50214a1e-8cea-4c86-b61b-162e286f6b64\") " pod="openstack/swift-ring-rebalance-6xhch" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.576841 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/50214a1e-8cea-4c86-b61b-162e286f6b64-scripts\") pod \"swift-ring-rebalance-6xhch\" (UID: \"50214a1e-8cea-4c86-b61b-162e286f6b64\") " pod="openstack/swift-ring-rebalance-6xhch" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.576904 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5f4f2323-f99a-4d85-8e0f-a0eb968de27b-etc-swift\") pod \"swift-ring-rebalance-wgmjm\" (UID: \"5f4f2323-f99a-4d85-8e0f-a0eb968de27b\") " pod="openstack/swift-ring-rebalance-wgmjm" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.576971 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5f4f2323-f99a-4d85-8e0f-a0eb968de27b-swiftconf\") pod \"swift-ring-rebalance-wgmjm\" (UID: \"5f4f2323-f99a-4d85-8e0f-a0eb968de27b\") " pod="openstack/swift-ring-rebalance-wgmjm" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.577192 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/50214a1e-8cea-4c86-b61b-162e286f6b64-etc-swift\") pod \"swift-ring-rebalance-6xhch\" (UID: \"50214a1e-8cea-4c86-b61b-162e286f6b64\") " pod="openstack/swift-ring-rebalance-6xhch" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.578916 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cfd13e5d-ca28-4061-8ca5-2c4566332bf0-etc-swift\") pod \"swift-storage-0\" (UID: \"cfd13e5d-ca28-4061-8ca5-2c4566332bf0\") " pod="openstack/swift-storage-0" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.579064 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5f4f2323-f99a-4d85-8e0f-a0eb968de27b-scripts\") pod \"swift-ring-rebalance-wgmjm\" (UID: \"5f4f2323-f99a-4d85-8e0f-a0eb968de27b\") " pod="openstack/swift-ring-rebalance-wgmjm" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.579540 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5f4f2323-f99a-4d85-8e0f-a0eb968de27b-etc-swift\") pod \"swift-ring-rebalance-wgmjm\" (UID: \"5f4f2323-f99a-4d85-8e0f-a0eb968de27b\") " pod="openstack/swift-ring-rebalance-wgmjm" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.580064 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5f4f2323-f99a-4d85-8e0f-a0eb968de27b-scripts\") pod \"swift-ring-rebalance-wgmjm\" (UID: \"5f4f2323-f99a-4d85-8e0f-a0eb968de27b\") " pod="openstack/swift-ring-rebalance-wgmjm" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.580124 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/50214a1e-8cea-4c86-b61b-162e286f6b64-scripts\") pod \"swift-ring-rebalance-6xhch\" (UID: \"50214a1e-8cea-4c86-b61b-162e286f6b64\") " pod="openstack/swift-ring-rebalance-6xhch" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.580676 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/50214a1e-8cea-4c86-b61b-162e286f6b64-ring-data-devices\") pod \"swift-ring-rebalance-6xhch\" (UID: \"50214a1e-8cea-4c86-b61b-162e286f6b64\") " pod="openstack/swift-ring-rebalance-6xhch" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.583135 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5f4f2323-f99a-4d85-8e0f-a0eb968de27b-ring-data-devices\") pod \"swift-ring-rebalance-wgmjm\" (UID: \"5f4f2323-f99a-4d85-8e0f-a0eb968de27b\") " pod="openstack/swift-ring-rebalance-wgmjm" Dec 06 06:15:21 crc kubenswrapper[4809]: E1206 06:15:21.583155 4809 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 06 06:15:21 crc kubenswrapper[4809]: E1206 06:15:21.583176 4809 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 06 06:15:21 crc kubenswrapper[4809]: E1206 06:15:21.583323 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/cfd13e5d-ca28-4061-8ca5-2c4566332bf0-etc-swift podName:cfd13e5d-ca28-4061-8ca5-2c4566332bf0 nodeName:}" failed. No retries permitted until 2025-12-06 06:15:22.583305936 +0000 UTC m=+1447.472288878 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/cfd13e5d-ca28-4061-8ca5-2c4566332bf0-etc-swift") pod "swift-storage-0" (UID: "cfd13e5d-ca28-4061-8ca5-2c4566332bf0") : configmap "swift-ring-files" not found Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.584297 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/50214a1e-8cea-4c86-b61b-162e286f6b64-swiftconf\") pod \"swift-ring-rebalance-6xhch\" (UID: \"50214a1e-8cea-4c86-b61b-162e286f6b64\") " pod="openstack/swift-ring-rebalance-6xhch" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.585794 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5f4f2323-f99a-4d85-8e0f-a0eb968de27b-dispersionconf\") pod \"swift-ring-rebalance-wgmjm\" (UID: \"5f4f2323-f99a-4d85-8e0f-a0eb968de27b\") " pod="openstack/swift-ring-rebalance-wgmjm" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.588156 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f4f2323-f99a-4d85-8e0f-a0eb968de27b-combined-ca-bundle\") pod \"swift-ring-rebalance-wgmjm\" (UID: \"5f4f2323-f99a-4d85-8e0f-a0eb968de27b\") " pod="openstack/swift-ring-rebalance-wgmjm" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.590678 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50214a1e-8cea-4c86-b61b-162e286f6b64-combined-ca-bundle\") pod \"swift-ring-rebalance-6xhch\" (UID: \"50214a1e-8cea-4c86-b61b-162e286f6b64\") " pod="openstack/swift-ring-rebalance-6xhch" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.591484 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5f4f2323-f99a-4d85-8e0f-a0eb968de27b-swiftconf\") pod \"swift-ring-rebalance-wgmjm\" (UID: \"5f4f2323-f99a-4d85-8e0f-a0eb968de27b\") " pod="openstack/swift-ring-rebalance-wgmjm" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.601664 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jr2rt\" (UniqueName: \"kubernetes.io/projected/5f4f2323-f99a-4d85-8e0f-a0eb968de27b-kube-api-access-jr2rt\") pod \"swift-ring-rebalance-wgmjm\" (UID: \"5f4f2323-f99a-4d85-8e0f-a0eb968de27b\") " pod="openstack/swift-ring-rebalance-wgmjm" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.601749 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xx9r7\" (UniqueName: \"kubernetes.io/projected/50214a1e-8cea-4c86-b61b-162e286f6b64-kube-api-access-xx9r7\") pod \"swift-ring-rebalance-6xhch\" (UID: \"50214a1e-8cea-4c86-b61b-162e286f6b64\") " pod="openstack/swift-ring-rebalance-6xhch" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.602520 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/50214a1e-8cea-4c86-b61b-162e286f6b64-dispersionconf\") pod \"swift-ring-rebalance-6xhch\" (UID: \"50214a1e-8cea-4c86-b61b-162e286f6b64\") " pod="openstack/swift-ring-rebalance-6xhch" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.660622 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-wgmjm" Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.759836 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-vnwqm"] Dec 06 06:15:21 crc kubenswrapper[4809]: I1206 06:15:21.770277 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-vnwqm"] Dec 06 06:15:22 crc kubenswrapper[4809]: I1206 06:15:22.349465 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-6xhch" Dec 06 06:15:22 crc kubenswrapper[4809]: I1206 06:15:22.359546 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-6xhch" Dec 06 06:15:22 crc kubenswrapper[4809]: I1206 06:15:22.501726 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/50214a1e-8cea-4c86-b61b-162e286f6b64-dispersionconf\") pod \"50214a1e-8cea-4c86-b61b-162e286f6b64\" (UID: \"50214a1e-8cea-4c86-b61b-162e286f6b64\") " Dec 06 06:15:22 crc kubenswrapper[4809]: I1206 06:15:22.501778 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/50214a1e-8cea-4c86-b61b-162e286f6b64-swiftconf\") pod \"50214a1e-8cea-4c86-b61b-162e286f6b64\" (UID: \"50214a1e-8cea-4c86-b61b-162e286f6b64\") " Dec 06 06:15:22 crc kubenswrapper[4809]: I1206 06:15:22.501839 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/50214a1e-8cea-4c86-b61b-162e286f6b64-ring-data-devices\") pod \"50214a1e-8cea-4c86-b61b-162e286f6b64\" (UID: \"50214a1e-8cea-4c86-b61b-162e286f6b64\") " Dec 06 06:15:22 crc kubenswrapper[4809]: I1206 06:15:22.501858 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/50214a1e-8cea-4c86-b61b-162e286f6b64-scripts\") pod \"50214a1e-8cea-4c86-b61b-162e286f6b64\" (UID: \"50214a1e-8cea-4c86-b61b-162e286f6b64\") " Dec 06 06:15:22 crc kubenswrapper[4809]: I1206 06:15:22.501927 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xx9r7\" (UniqueName: \"kubernetes.io/projected/50214a1e-8cea-4c86-b61b-162e286f6b64-kube-api-access-xx9r7\") pod \"50214a1e-8cea-4c86-b61b-162e286f6b64\" (UID: \"50214a1e-8cea-4c86-b61b-162e286f6b64\") " Dec 06 06:15:22 crc kubenswrapper[4809]: I1206 06:15:22.502014 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/50214a1e-8cea-4c86-b61b-162e286f6b64-etc-swift\") pod \"50214a1e-8cea-4c86-b61b-162e286f6b64\" (UID: \"50214a1e-8cea-4c86-b61b-162e286f6b64\") " Dec 06 06:15:22 crc kubenswrapper[4809]: I1206 06:15:22.502125 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50214a1e-8cea-4c86-b61b-162e286f6b64-combined-ca-bundle\") pod \"50214a1e-8cea-4c86-b61b-162e286f6b64\" (UID: \"50214a1e-8cea-4c86-b61b-162e286f6b64\") " Dec 06 06:15:22 crc kubenswrapper[4809]: I1206 06:15:22.502433 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50214a1e-8cea-4c86-b61b-162e286f6b64-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "50214a1e-8cea-4c86-b61b-162e286f6b64" (UID: "50214a1e-8cea-4c86-b61b-162e286f6b64"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:15:22 crc kubenswrapper[4809]: I1206 06:15:22.502734 4809 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/50214a1e-8cea-4c86-b61b-162e286f6b64-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:22 crc kubenswrapper[4809]: I1206 06:15:22.507908 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50214a1e-8cea-4c86-b61b-162e286f6b64-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "50214a1e-8cea-4c86-b61b-162e286f6b64" (UID: "50214a1e-8cea-4c86-b61b-162e286f6b64"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:15:22 crc kubenswrapper[4809]: I1206 06:15:22.508026 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50214a1e-8cea-4c86-b61b-162e286f6b64-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "50214a1e-8cea-4c86-b61b-162e286f6b64" (UID: "50214a1e-8cea-4c86-b61b-162e286f6b64"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:15:22 crc kubenswrapper[4809]: I1206 06:15:22.508092 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50214a1e-8cea-4c86-b61b-162e286f6b64-kube-api-access-xx9r7" (OuterVolumeSpecName: "kube-api-access-xx9r7") pod "50214a1e-8cea-4c86-b61b-162e286f6b64" (UID: "50214a1e-8cea-4c86-b61b-162e286f6b64"). InnerVolumeSpecName "kube-api-access-xx9r7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:15:22 crc kubenswrapper[4809]: I1206 06:15:22.508126 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50214a1e-8cea-4c86-b61b-162e286f6b64-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "50214a1e-8cea-4c86-b61b-162e286f6b64" (UID: "50214a1e-8cea-4c86-b61b-162e286f6b64"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:15:22 crc kubenswrapper[4809]: I1206 06:15:22.549442 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/50214a1e-8cea-4c86-b61b-162e286f6b64-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "50214a1e-8cea-4c86-b61b-162e286f6b64" (UID: "50214a1e-8cea-4c86-b61b-162e286f6b64"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:15:22 crc kubenswrapper[4809]: I1206 06:15:22.549492 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/50214a1e-8cea-4c86-b61b-162e286f6b64-scripts" (OuterVolumeSpecName: "scripts") pod "50214a1e-8cea-4c86-b61b-162e286f6b64" (UID: "50214a1e-8cea-4c86-b61b-162e286f6b64"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:15:22 crc kubenswrapper[4809]: I1206 06:15:22.604849 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cfd13e5d-ca28-4061-8ca5-2c4566332bf0-etc-swift\") pod \"swift-storage-0\" (UID: \"cfd13e5d-ca28-4061-8ca5-2c4566332bf0\") " pod="openstack/swift-storage-0" Dec 06 06:15:22 crc kubenswrapper[4809]: E1206 06:15:22.605139 4809 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 06 06:15:22 crc kubenswrapper[4809]: I1206 06:15:22.605169 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50214a1e-8cea-4c86-b61b-162e286f6b64-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:22 crc kubenswrapper[4809]: I1206 06:15:22.605198 4809 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/50214a1e-8cea-4c86-b61b-162e286f6b64-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:22 crc kubenswrapper[4809]: I1206 06:15:22.605217 4809 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/50214a1e-8cea-4c86-b61b-162e286f6b64-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:22 crc kubenswrapper[4809]: I1206 06:15:22.605230 4809 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/50214a1e-8cea-4c86-b61b-162e286f6b64-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:22 crc kubenswrapper[4809]: I1206 06:15:22.605243 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/50214a1e-8cea-4c86-b61b-162e286f6b64-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:22 crc kubenswrapper[4809]: I1206 06:15:22.605255 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xx9r7\" (UniqueName: \"kubernetes.io/projected/50214a1e-8cea-4c86-b61b-162e286f6b64-kube-api-access-xx9r7\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:22 crc kubenswrapper[4809]: E1206 06:15:22.605175 4809 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 06 06:15:22 crc kubenswrapper[4809]: E1206 06:15:22.605349 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/cfd13e5d-ca28-4061-8ca5-2c4566332bf0-etc-swift podName:cfd13e5d-ca28-4061-8ca5-2c4566332bf0 nodeName:}" failed. No retries permitted until 2025-12-06 06:15:24.605327743 +0000 UTC m=+1449.494310685 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/cfd13e5d-ca28-4061-8ca5-2c4566332bf0-etc-swift") pod "swift-storage-0" (UID: "cfd13e5d-ca28-4061-8ca5-2c4566332bf0") : configmap "swift-ring-files" not found Dec 06 06:15:23 crc kubenswrapper[4809]: I1206 06:15:23.358137 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-6xhch" Dec 06 06:15:23 crc kubenswrapper[4809]: I1206 06:15:23.454502 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10497d10-a5e4-42e5-817c-1c41a2ef29e2" path="/var/lib/kubelet/pods/10497d10-a5e4-42e5-817c-1c41a2ef29e2/volumes" Dec 06 06:15:23 crc kubenswrapper[4809]: I1206 06:15:23.455219 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-6xhch"] Dec 06 06:15:23 crc kubenswrapper[4809]: I1206 06:15:23.477751 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-6xhch"] Dec 06 06:15:24 crc kubenswrapper[4809]: I1206 06:15:24.679192 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cfd13e5d-ca28-4061-8ca5-2c4566332bf0-etc-swift\") pod \"swift-storage-0\" (UID: \"cfd13e5d-ca28-4061-8ca5-2c4566332bf0\") " pod="openstack/swift-storage-0" Dec 06 06:15:24 crc kubenswrapper[4809]: E1206 06:15:24.679415 4809 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 06 06:15:24 crc kubenswrapper[4809]: E1206 06:15:24.679589 4809 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 06 06:15:24 crc kubenswrapper[4809]: E1206 06:15:24.679637 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/cfd13e5d-ca28-4061-8ca5-2c4566332bf0-etc-swift podName:cfd13e5d-ca28-4061-8ca5-2c4566332bf0 nodeName:}" failed. No retries permitted until 2025-12-06 06:15:28.679623034 +0000 UTC m=+1453.568605976 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/cfd13e5d-ca28-4061-8ca5-2c4566332bf0-etc-swift") pod "swift-storage-0" (UID: "cfd13e5d-ca28-4061-8ca5-2c4566332bf0") : configmap "swift-ring-files" not found Dec 06 06:15:25 crc kubenswrapper[4809]: I1206 06:15:25.400835 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50214a1e-8cea-4c86-b61b-162e286f6b64" path="/var/lib/kubelet/pods/50214a1e-8cea-4c86-b61b-162e286f6b64/volumes" Dec 06 06:15:25 crc kubenswrapper[4809]: I1206 06:15:25.935365 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 06 06:15:25 crc kubenswrapper[4809]: I1206 06:15:25.935634 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 06 06:15:27 crc kubenswrapper[4809]: I1206 06:15:27.147495 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-fwqkk" Dec 06 06:15:27 crc kubenswrapper[4809]: I1206 06:15:27.239926 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f7349dd-e05d-4c2f-ab31-5d328994e551-dns-svc\") pod \"6f7349dd-e05d-4c2f-ab31-5d328994e551\" (UID: \"6f7349dd-e05d-4c2f-ab31-5d328994e551\") " Dec 06 06:15:27 crc kubenswrapper[4809]: I1206 06:15:27.240684 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d7hzx\" (UniqueName: \"kubernetes.io/projected/6f7349dd-e05d-4c2f-ab31-5d328994e551-kube-api-access-d7hzx\") pod \"6f7349dd-e05d-4c2f-ab31-5d328994e551\" (UID: \"6f7349dd-e05d-4c2f-ab31-5d328994e551\") " Dec 06 06:15:27 crc kubenswrapper[4809]: I1206 06:15:27.240914 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6f7349dd-e05d-4c2f-ab31-5d328994e551-ovsdbserver-nb\") pod \"6f7349dd-e05d-4c2f-ab31-5d328994e551\" (UID: \"6f7349dd-e05d-4c2f-ab31-5d328994e551\") " Dec 06 06:15:27 crc kubenswrapper[4809]: I1206 06:15:27.240992 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f7349dd-e05d-4c2f-ab31-5d328994e551-config\") pod \"6f7349dd-e05d-4c2f-ab31-5d328994e551\" (UID: \"6f7349dd-e05d-4c2f-ab31-5d328994e551\") " Dec 06 06:15:27 crc kubenswrapper[4809]: I1206 06:15:27.241059 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6f7349dd-e05d-4c2f-ab31-5d328994e551-ovsdbserver-sb\") pod \"6f7349dd-e05d-4c2f-ab31-5d328994e551\" (UID: \"6f7349dd-e05d-4c2f-ab31-5d328994e551\") " Dec 06 06:15:27 crc kubenswrapper[4809]: I1206 06:15:27.247604 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f7349dd-e05d-4c2f-ab31-5d328994e551-kube-api-access-d7hzx" (OuterVolumeSpecName: "kube-api-access-d7hzx") pod "6f7349dd-e05d-4c2f-ab31-5d328994e551" (UID: "6f7349dd-e05d-4c2f-ab31-5d328994e551"). InnerVolumeSpecName "kube-api-access-d7hzx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:15:27 crc kubenswrapper[4809]: I1206 06:15:27.275189 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f7349dd-e05d-4c2f-ab31-5d328994e551-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6f7349dd-e05d-4c2f-ab31-5d328994e551" (UID: "6f7349dd-e05d-4c2f-ab31-5d328994e551"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:15:27 crc kubenswrapper[4809]: I1206 06:15:27.276278 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f7349dd-e05d-4c2f-ab31-5d328994e551-config" (OuterVolumeSpecName: "config") pod "6f7349dd-e05d-4c2f-ab31-5d328994e551" (UID: "6f7349dd-e05d-4c2f-ab31-5d328994e551"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:15:27 crc kubenswrapper[4809]: I1206 06:15:27.278906 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f7349dd-e05d-4c2f-ab31-5d328994e551-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6f7349dd-e05d-4c2f-ab31-5d328994e551" (UID: "6f7349dd-e05d-4c2f-ab31-5d328994e551"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:15:27 crc kubenswrapper[4809]: I1206 06:15:27.287743 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f7349dd-e05d-4c2f-ab31-5d328994e551-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6f7349dd-e05d-4c2f-ab31-5d328994e551" (UID: "6f7349dd-e05d-4c2f-ab31-5d328994e551"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:15:27 crc kubenswrapper[4809]: I1206 06:15:27.327660 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 06 06:15:27 crc kubenswrapper[4809]: I1206 06:15:27.327716 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 06 06:15:27 crc kubenswrapper[4809]: I1206 06:15:27.349810 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6f7349dd-e05d-4c2f-ab31-5d328994e551-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:27 crc kubenswrapper[4809]: I1206 06:15:27.349851 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f7349dd-e05d-4c2f-ab31-5d328994e551-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:27 crc kubenswrapper[4809]: I1206 06:15:27.349866 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6f7349dd-e05d-4c2f-ab31-5d328994e551-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:27 crc kubenswrapper[4809]: I1206 06:15:27.349876 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f7349dd-e05d-4c2f-ab31-5d328994e551-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:27 crc kubenswrapper[4809]: I1206 06:15:27.349887 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d7hzx\" (UniqueName: \"kubernetes.io/projected/6f7349dd-e05d-4c2f-ab31-5d328994e551-kube-api-access-d7hzx\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:27 crc kubenswrapper[4809]: I1206 06:15:27.372158 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-wgmjm"] Dec 06 06:15:27 crc kubenswrapper[4809]: I1206 06:15:27.421704 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-wgmjm" event={"ID":"5f4f2323-f99a-4d85-8e0f-a0eb968de27b","Type":"ContainerStarted","Data":"629f7d646a99a3a7368147301a35395ab73534330c4109dbfa174e596452420c"} Dec 06 06:15:27 crc kubenswrapper[4809]: I1206 06:15:27.423350 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"1360c733-da74-4d37-8842-75050a356334","Type":"ContainerStarted","Data":"f3698b808f0f86837c731f207d09b30b74cbddc559ce0fa0ce086a9952e758ab"} Dec 06 06:15:27 crc kubenswrapper[4809]: I1206 06:15:27.425049 4809 generic.go:334] "Generic (PLEG): container finished" podID="0c24a4f1-4cfb-4aae-ac2e-fa7680751c26" containerID="66bc7af1b8f7b0ea275598143ea77ba9570c3706b7d8f18d1b05c2d5b7f4b7a5" exitCode=0 Dec 06 06:15:27 crc kubenswrapper[4809]: I1206 06:15:27.425097 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-r89dg" event={"ID":"0c24a4f1-4cfb-4aae-ac2e-fa7680751c26","Type":"ContainerDied","Data":"66bc7af1b8f7b0ea275598143ea77ba9570c3706b7d8f18d1b05c2d5b7f4b7a5"} Dec 06 06:15:27 crc kubenswrapper[4809]: I1206 06:15:27.425114 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-r89dg" event={"ID":"0c24a4f1-4cfb-4aae-ac2e-fa7680751c26","Type":"ContainerStarted","Data":"c6d176c838ee7982f8dba1d9a555740fc1ea025201e0fd30b602de6d5b97fcb9"} Dec 06 06:15:27 crc kubenswrapper[4809]: I1206 06:15:27.428246 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-fwqkk" event={"ID":"6f7349dd-e05d-4c2f-ab31-5d328994e551","Type":"ContainerDied","Data":"2bc4b6b62bc2123abff4cbfc3a52da6288fe65bd7969edb93eae017fa6693ee1"} Dec 06 06:15:27 crc kubenswrapper[4809]: I1206 06:15:27.428294 4809 scope.go:117] "RemoveContainer" containerID="c37154aecc0e7bca137b5be30758f485495f63810d501a337a570aea85ec5849" Dec 06 06:15:27 crc kubenswrapper[4809]: I1206 06:15:27.428411 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-fwqkk" Dec 06 06:15:27 crc kubenswrapper[4809]: I1206 06:15:27.556407 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-fwqkk"] Dec 06 06:15:27 crc kubenswrapper[4809]: I1206 06:15:27.567439 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-fwqkk"] Dec 06 06:15:28 crc kubenswrapper[4809]: I1206 06:15:28.453186 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"1360c733-da74-4d37-8842-75050a356334","Type":"ContainerStarted","Data":"ea4c4e8048a08da4501f84cbc055d02746fd187d3aa85ac01153ddcbfab9e1b2"} Dec 06 06:15:28 crc kubenswrapper[4809]: I1206 06:15:28.453556 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 06 06:15:28 crc kubenswrapper[4809]: I1206 06:15:28.462010 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-r89dg" event={"ID":"0c24a4f1-4cfb-4aae-ac2e-fa7680751c26","Type":"ContainerStarted","Data":"43136fcc5a12ae304cd1d4d1dc504b5c5485dce32274385a16b0fa0c26d6de99"} Dec 06 06:15:28 crc kubenswrapper[4809]: I1206 06:15:28.462997 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-r89dg" Dec 06 06:15:28 crc kubenswrapper[4809]: I1206 06:15:28.477685 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=3.856723589 podStartE2EDuration="10.477668724s" podCreationTimestamp="2025-12-06 06:15:18 +0000 UTC" firstStartedPulling="2025-12-06 06:15:20.270033235 +0000 UTC m=+1445.159016177" lastFinishedPulling="2025-12-06 06:15:26.89097837 +0000 UTC m=+1451.779961312" observedRunningTime="2025-12-06 06:15:28.475755402 +0000 UTC m=+1453.364738364" watchObservedRunningTime="2025-12-06 06:15:28.477668724 +0000 UTC m=+1453.366651666" Dec 06 06:15:28 crc kubenswrapper[4809]: I1206 06:15:28.505916 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-r89dg" podStartSLOduration=9.505891745 podStartE2EDuration="9.505891745s" podCreationTimestamp="2025-12-06 06:15:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:15:28.498369021 +0000 UTC m=+1453.387351963" watchObservedRunningTime="2025-12-06 06:15:28.505891745 +0000 UTC m=+1453.394874687" Dec 06 06:15:28 crc kubenswrapper[4809]: I1206 06:15:28.687733 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cfd13e5d-ca28-4061-8ca5-2c4566332bf0-etc-swift\") pod \"swift-storage-0\" (UID: \"cfd13e5d-ca28-4061-8ca5-2c4566332bf0\") " pod="openstack/swift-storage-0" Dec 06 06:15:28 crc kubenswrapper[4809]: E1206 06:15:28.687984 4809 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 06 06:15:28 crc kubenswrapper[4809]: E1206 06:15:28.688023 4809 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 06 06:15:28 crc kubenswrapper[4809]: E1206 06:15:28.688093 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/cfd13e5d-ca28-4061-8ca5-2c4566332bf0-etc-swift podName:cfd13e5d-ca28-4061-8ca5-2c4566332bf0 nodeName:}" failed. No retries permitted until 2025-12-06 06:15:36.688073497 +0000 UTC m=+1461.577056439 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/cfd13e5d-ca28-4061-8ca5-2c4566332bf0-etc-swift") pod "swift-storage-0" (UID: "cfd13e5d-ca28-4061-8ca5-2c4566332bf0") : configmap "swift-ring-files" not found Dec 06 06:15:28 crc kubenswrapper[4809]: I1206 06:15:28.861345 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 06 06:15:28 crc kubenswrapper[4809]: I1206 06:15:28.948505 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 06 06:15:29 crc kubenswrapper[4809]: I1206 06:15:29.404980 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f7349dd-e05d-4c2f-ab31-5d328994e551" path="/var/lib/kubelet/pods/6f7349dd-e05d-4c2f-ab31-5d328994e551/volumes" Dec 06 06:15:29 crc kubenswrapper[4809]: I1206 06:15:29.553375 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-rksl7"] Dec 06 06:15:29 crc kubenswrapper[4809]: E1206 06:15:29.553899 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f7349dd-e05d-4c2f-ab31-5d328994e551" containerName="init" Dec 06 06:15:29 crc kubenswrapper[4809]: I1206 06:15:29.553916 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f7349dd-e05d-4c2f-ab31-5d328994e551" containerName="init" Dec 06 06:15:29 crc kubenswrapper[4809]: I1206 06:15:29.554211 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f7349dd-e05d-4c2f-ab31-5d328994e551" containerName="init" Dec 06 06:15:29 crc kubenswrapper[4809]: I1206 06:15:29.555150 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-rksl7" Dec 06 06:15:29 crc kubenswrapper[4809]: I1206 06:15:29.562966 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-rksl7"] Dec 06 06:15:29 crc kubenswrapper[4809]: I1206 06:15:29.679460 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-c863-account-create-update-29vzq"] Dec 06 06:15:29 crc kubenswrapper[4809]: I1206 06:15:29.682415 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-c863-account-create-update-29vzq" Dec 06 06:15:29 crc kubenswrapper[4809]: I1206 06:15:29.685963 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-openstack-db-secret" Dec 06 06:15:29 crc kubenswrapper[4809]: I1206 06:15:29.690879 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-c863-account-create-update-29vzq"] Dec 06 06:15:29 crc kubenswrapper[4809]: I1206 06:15:29.746665 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc6d16d3-5a2f-4d15-ba47-a25e4608703b-operator-scripts\") pod \"mysqld-exporter-openstack-db-create-rksl7\" (UID: \"cc6d16d3-5a2f-4d15-ba47-a25e4608703b\") " pod="openstack/mysqld-exporter-openstack-db-create-rksl7" Dec 06 06:15:29 crc kubenswrapper[4809]: I1206 06:15:29.746758 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9njqg\" (UniqueName: \"kubernetes.io/projected/cc6d16d3-5a2f-4d15-ba47-a25e4608703b-kube-api-access-9njqg\") pod \"mysqld-exporter-openstack-db-create-rksl7\" (UID: \"cc6d16d3-5a2f-4d15-ba47-a25e4608703b\") " pod="openstack/mysqld-exporter-openstack-db-create-rksl7" Dec 06 06:15:29 crc kubenswrapper[4809]: I1206 06:15:29.849965 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ljkj\" (UniqueName: \"kubernetes.io/projected/2afacc78-69cc-4fd2-b857-60a53a29c650-kube-api-access-7ljkj\") pod \"mysqld-exporter-c863-account-create-update-29vzq\" (UID: \"2afacc78-69cc-4fd2-b857-60a53a29c650\") " pod="openstack/mysqld-exporter-c863-account-create-update-29vzq" Dec 06 06:15:29 crc kubenswrapper[4809]: I1206 06:15:29.850256 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc6d16d3-5a2f-4d15-ba47-a25e4608703b-operator-scripts\") pod \"mysqld-exporter-openstack-db-create-rksl7\" (UID: \"cc6d16d3-5a2f-4d15-ba47-a25e4608703b\") " pod="openstack/mysqld-exporter-openstack-db-create-rksl7" Dec 06 06:15:29 crc kubenswrapper[4809]: I1206 06:15:29.850450 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9njqg\" (UniqueName: \"kubernetes.io/projected/cc6d16d3-5a2f-4d15-ba47-a25e4608703b-kube-api-access-9njqg\") pod \"mysqld-exporter-openstack-db-create-rksl7\" (UID: \"cc6d16d3-5a2f-4d15-ba47-a25e4608703b\") " pod="openstack/mysqld-exporter-openstack-db-create-rksl7" Dec 06 06:15:29 crc kubenswrapper[4809]: I1206 06:15:29.850602 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2afacc78-69cc-4fd2-b857-60a53a29c650-operator-scripts\") pod \"mysqld-exporter-c863-account-create-update-29vzq\" (UID: \"2afacc78-69cc-4fd2-b857-60a53a29c650\") " pod="openstack/mysqld-exporter-c863-account-create-update-29vzq" Dec 06 06:15:29 crc kubenswrapper[4809]: I1206 06:15:29.851634 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc6d16d3-5a2f-4d15-ba47-a25e4608703b-operator-scripts\") pod \"mysqld-exporter-openstack-db-create-rksl7\" (UID: \"cc6d16d3-5a2f-4d15-ba47-a25e4608703b\") " pod="openstack/mysqld-exporter-openstack-db-create-rksl7" Dec 06 06:15:29 crc kubenswrapper[4809]: I1206 06:15:29.873061 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9njqg\" (UniqueName: \"kubernetes.io/projected/cc6d16d3-5a2f-4d15-ba47-a25e4608703b-kube-api-access-9njqg\") pod \"mysqld-exporter-openstack-db-create-rksl7\" (UID: \"cc6d16d3-5a2f-4d15-ba47-a25e4608703b\") " pod="openstack/mysqld-exporter-openstack-db-create-rksl7" Dec 06 06:15:29 crc kubenswrapper[4809]: I1206 06:15:29.875422 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-rksl7" Dec 06 06:15:29 crc kubenswrapper[4809]: I1206 06:15:29.954537 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2afacc78-69cc-4fd2-b857-60a53a29c650-operator-scripts\") pod \"mysqld-exporter-c863-account-create-update-29vzq\" (UID: \"2afacc78-69cc-4fd2-b857-60a53a29c650\") " pod="openstack/mysqld-exporter-c863-account-create-update-29vzq" Dec 06 06:15:29 crc kubenswrapper[4809]: I1206 06:15:29.954776 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ljkj\" (UniqueName: \"kubernetes.io/projected/2afacc78-69cc-4fd2-b857-60a53a29c650-kube-api-access-7ljkj\") pod \"mysqld-exporter-c863-account-create-update-29vzq\" (UID: \"2afacc78-69cc-4fd2-b857-60a53a29c650\") " pod="openstack/mysqld-exporter-c863-account-create-update-29vzq" Dec 06 06:15:29 crc kubenswrapper[4809]: I1206 06:15:29.956794 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2afacc78-69cc-4fd2-b857-60a53a29c650-operator-scripts\") pod \"mysqld-exporter-c863-account-create-update-29vzq\" (UID: \"2afacc78-69cc-4fd2-b857-60a53a29c650\") " pod="openstack/mysqld-exporter-c863-account-create-update-29vzq" Dec 06 06:15:29 crc kubenswrapper[4809]: I1206 06:15:29.958101 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 06 06:15:29 crc kubenswrapper[4809]: I1206 06:15:29.979841 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ljkj\" (UniqueName: \"kubernetes.io/projected/2afacc78-69cc-4fd2-b857-60a53a29c650-kube-api-access-7ljkj\") pod \"mysqld-exporter-c863-account-create-update-29vzq\" (UID: \"2afacc78-69cc-4fd2-b857-60a53a29c650\") " pod="openstack/mysqld-exporter-c863-account-create-update-29vzq" Dec 06 06:15:30 crc kubenswrapper[4809]: I1206 06:15:30.039916 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-c863-account-create-update-29vzq" Dec 06 06:15:30 crc kubenswrapper[4809]: I1206 06:15:30.078762 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 06 06:15:30 crc kubenswrapper[4809]: I1206 06:15:30.500582 4809 generic.go:334] "Generic (PLEG): container finished" podID="eb1bd4de-6456-4817-a85d-ab0176153f2c" containerID="cd1fea316ddc0d8b812b68b8ae2836932648f88ecde0d2a0cb057eb24d8002a9" exitCode=0 Dec 06 06:15:30 crc kubenswrapper[4809]: I1206 06:15:30.500903 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"eb1bd4de-6456-4817-a85d-ab0176153f2c","Type":"ContainerDied","Data":"cd1fea316ddc0d8b812b68b8ae2836932648f88ecde0d2a0cb057eb24d8002a9"} Dec 06 06:15:30 crc kubenswrapper[4809]: I1206 06:15:30.504718 4809 generic.go:334] "Generic (PLEG): container finished" podID="1ceabc13-38c5-4425-8d85-c3afa8935243" containerID="eb40e5f3076606ef761086938cb7e8b4aad3ddeaa2e7eb9a9934ed5936e19c64" exitCode=0 Dec 06 06:15:30 crc kubenswrapper[4809]: I1206 06:15:30.505774 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1ceabc13-38c5-4425-8d85-c3afa8935243","Type":"ContainerDied","Data":"eb40e5f3076606ef761086938cb7e8b4aad3ddeaa2e7eb9a9934ed5936e19c64"} Dec 06 06:15:35 crc kubenswrapper[4809]: I1206 06:15:35.095355 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-r89dg" Dec 06 06:15:35 crc kubenswrapper[4809]: I1206 06:15:35.178598 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-gr99p"] Dec 06 06:15:35 crc kubenswrapper[4809]: I1206 06:15:35.179526 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-gr99p" podUID="a8bfcbfb-3946-420c-b90f-f390887145f0" containerName="dnsmasq-dns" containerID="cri-o://e4f1e32a64e611e44984494ef945ec0948d0c06be22b8dd98b0210df10ed1d23" gracePeriod=10 Dec 06 06:15:36 crc kubenswrapper[4809]: I1206 06:15:36.202673 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-6499fbfdc8-929cv" podUID="ba14ef06-3f68-45a3-8fde-36980cd2d194" containerName="console" containerID="cri-o://783e760a1bf33d4a66aaf5565d4bab5a8309f7adb668a85a0e93f84348ca28a4" gracePeriod=15 Dec 06 06:15:36 crc kubenswrapper[4809]: I1206 06:15:36.575806 4809 generic.go:334] "Generic (PLEG): container finished" podID="a8bfcbfb-3946-420c-b90f-f390887145f0" containerID="e4f1e32a64e611e44984494ef945ec0948d0c06be22b8dd98b0210df10ed1d23" exitCode=0 Dec 06 06:15:36 crc kubenswrapper[4809]: I1206 06:15:36.576039 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-gr99p" event={"ID":"a8bfcbfb-3946-420c-b90f-f390887145f0","Type":"ContainerDied","Data":"e4f1e32a64e611e44984494ef945ec0948d0c06be22b8dd98b0210df10ed1d23"} Dec 06 06:15:36 crc kubenswrapper[4809]: I1206 06:15:36.581353 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-6499fbfdc8-929cv_ba14ef06-3f68-45a3-8fde-36980cd2d194/console/0.log" Dec 06 06:15:36 crc kubenswrapper[4809]: I1206 06:15:36.581412 4809 generic.go:334] "Generic (PLEG): container finished" podID="ba14ef06-3f68-45a3-8fde-36980cd2d194" containerID="783e760a1bf33d4a66aaf5565d4bab5a8309f7adb668a85a0e93f84348ca28a4" exitCode=2 Dec 06 06:15:36 crc kubenswrapper[4809]: I1206 06:15:36.581466 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6499fbfdc8-929cv" event={"ID":"ba14ef06-3f68-45a3-8fde-36980cd2d194","Type":"ContainerDied","Data":"783e760a1bf33d4a66aaf5565d4bab5a8309f7adb668a85a0e93f84348ca28a4"} Dec 06 06:15:36 crc kubenswrapper[4809]: I1206 06:15:36.728162 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cfd13e5d-ca28-4061-8ca5-2c4566332bf0-etc-swift\") pod \"swift-storage-0\" (UID: \"cfd13e5d-ca28-4061-8ca5-2c4566332bf0\") " pod="openstack/swift-storage-0" Dec 06 06:15:36 crc kubenswrapper[4809]: E1206 06:15:36.728395 4809 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 06 06:15:36 crc kubenswrapper[4809]: E1206 06:15:36.728425 4809 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 06 06:15:36 crc kubenswrapper[4809]: E1206 06:15:36.728491 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/cfd13e5d-ca28-4061-8ca5-2c4566332bf0-etc-swift podName:cfd13e5d-ca28-4061-8ca5-2c4566332bf0 nodeName:}" failed. No retries permitted until 2025-12-06 06:15:52.728468387 +0000 UTC m=+1477.617451329 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/cfd13e5d-ca28-4061-8ca5-2c4566332bf0-etc-swift") pod "swift-storage-0" (UID: "cfd13e5d-ca28-4061-8ca5-2c4566332bf0") : configmap "swift-ring-files" not found Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.051579 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-gr99p" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.144899 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a8bfcbfb-3946-420c-b90f-f390887145f0-dns-svc\") pod \"a8bfcbfb-3946-420c-b90f-f390887145f0\" (UID: \"a8bfcbfb-3946-420c-b90f-f390887145f0\") " Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.145094 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8bfcbfb-3946-420c-b90f-f390887145f0-config\") pod \"a8bfcbfb-3946-420c-b90f-f390887145f0\" (UID: \"a8bfcbfb-3946-420c-b90f-f390887145f0\") " Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.145153 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zt2b\" (UniqueName: \"kubernetes.io/projected/a8bfcbfb-3946-420c-b90f-f390887145f0-kube-api-access-4zt2b\") pod \"a8bfcbfb-3946-420c-b90f-f390887145f0\" (UID: \"a8bfcbfb-3946-420c-b90f-f390887145f0\") " Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.171827 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8bfcbfb-3946-420c-b90f-f390887145f0-kube-api-access-4zt2b" (OuterVolumeSpecName: "kube-api-access-4zt2b") pod "a8bfcbfb-3946-420c-b90f-f390887145f0" (UID: "a8bfcbfb-3946-420c-b90f-f390887145f0"). InnerVolumeSpecName "kube-api-access-4zt2b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.237573 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-6499fbfdc8-929cv_ba14ef06-3f68-45a3-8fde-36980cd2d194/console/0.log" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.237639 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6499fbfdc8-929cv" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.248869 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zt2b\" (UniqueName: \"kubernetes.io/projected/a8bfcbfb-3946-420c-b90f-f390887145f0-kube-api-access-4zt2b\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.319102 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-qz5fx"] Dec 06 06:15:37 crc kubenswrapper[4809]: E1206 06:15:37.329554 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba14ef06-3f68-45a3-8fde-36980cd2d194" containerName="console" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.329590 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba14ef06-3f68-45a3-8fde-36980cd2d194" containerName="console" Dec 06 06:15:37 crc kubenswrapper[4809]: E1206 06:15:37.329613 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8bfcbfb-3946-420c-b90f-f390887145f0" containerName="dnsmasq-dns" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.329619 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8bfcbfb-3946-420c-b90f-f390887145f0" containerName="dnsmasq-dns" Dec 06 06:15:37 crc kubenswrapper[4809]: E1206 06:15:37.329636 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8bfcbfb-3946-420c-b90f-f390887145f0" containerName="init" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.329644 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8bfcbfb-3946-420c-b90f-f390887145f0" containerName="init" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.329913 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba14ef06-3f68-45a3-8fde-36980cd2d194" containerName="console" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.329950 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8bfcbfb-3946-420c-b90f-f390887145f0" containerName="dnsmasq-dns" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.331161 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-qz5fx" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.351191 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ba14ef06-3f68-45a3-8fde-36980cd2d194-trusted-ca-bundle\") pod \"ba14ef06-3f68-45a3-8fde-36980cd2d194\" (UID: \"ba14ef06-3f68-45a3-8fde-36980cd2d194\") " Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.351289 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ba14ef06-3f68-45a3-8fde-36980cd2d194-service-ca\") pod \"ba14ef06-3f68-45a3-8fde-36980cd2d194\" (UID: \"ba14ef06-3f68-45a3-8fde-36980cd2d194\") " Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.351673 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ba14ef06-3f68-45a3-8fde-36980cd2d194-console-config\") pod \"ba14ef06-3f68-45a3-8fde-36980cd2d194\" (UID: \"ba14ef06-3f68-45a3-8fde-36980cd2d194\") " Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.351797 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ba14ef06-3f68-45a3-8fde-36980cd2d194-console-serving-cert\") pod \"ba14ef06-3f68-45a3-8fde-36980cd2d194\" (UID: \"ba14ef06-3f68-45a3-8fde-36980cd2d194\") " Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.351830 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mh2mm\" (UniqueName: \"kubernetes.io/projected/ba14ef06-3f68-45a3-8fde-36980cd2d194-kube-api-access-mh2mm\") pod \"ba14ef06-3f68-45a3-8fde-36980cd2d194\" (UID: \"ba14ef06-3f68-45a3-8fde-36980cd2d194\") " Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.351947 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ba14ef06-3f68-45a3-8fde-36980cd2d194-oauth-serving-cert\") pod \"ba14ef06-3f68-45a3-8fde-36980cd2d194\" (UID: \"ba14ef06-3f68-45a3-8fde-36980cd2d194\") " Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.351989 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ba14ef06-3f68-45a3-8fde-36980cd2d194-console-oauth-config\") pod \"ba14ef06-3f68-45a3-8fde-36980cd2d194\" (UID: \"ba14ef06-3f68-45a3-8fde-36980cd2d194\") " Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.356721 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba14ef06-3f68-45a3-8fde-36980cd2d194-console-config" (OuterVolumeSpecName: "console-config") pod "ba14ef06-3f68-45a3-8fde-36980cd2d194" (UID: "ba14ef06-3f68-45a3-8fde-36980cd2d194"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.357131 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba14ef06-3f68-45a3-8fde-36980cd2d194-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "ba14ef06-3f68-45a3-8fde-36980cd2d194" (UID: "ba14ef06-3f68-45a3-8fde-36980cd2d194"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.357466 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba14ef06-3f68-45a3-8fde-36980cd2d194-service-ca" (OuterVolumeSpecName: "service-ca") pod "ba14ef06-3f68-45a3-8fde-36980cd2d194" (UID: "ba14ef06-3f68-45a3-8fde-36980cd2d194"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.357813 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba14ef06-3f68-45a3-8fde-36980cd2d194-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "ba14ef06-3f68-45a3-8fde-36980cd2d194" (UID: "ba14ef06-3f68-45a3-8fde-36980cd2d194"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.362317 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba14ef06-3f68-45a3-8fde-36980cd2d194-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "ba14ef06-3f68-45a3-8fde-36980cd2d194" (UID: "ba14ef06-3f68-45a3-8fde-36980cd2d194"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.365275 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-qz5fx"] Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.405147 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba14ef06-3f68-45a3-8fde-36980cd2d194-kube-api-access-mh2mm" (OuterVolumeSpecName: "kube-api-access-mh2mm") pod "ba14ef06-3f68-45a3-8fde-36980cd2d194" (UID: "ba14ef06-3f68-45a3-8fde-36980cd2d194"). InnerVolumeSpecName "kube-api-access-mh2mm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.454697 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zstq8\" (UniqueName: \"kubernetes.io/projected/69e67ebc-dcc4-4f25-9366-7d8a23256d58-kube-api-access-zstq8\") pod \"keystone-db-create-qz5fx\" (UID: \"69e67ebc-dcc4-4f25-9366-7d8a23256d58\") " pod="openstack/keystone-db-create-qz5fx" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.454806 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/69e67ebc-dcc4-4f25-9366-7d8a23256d58-operator-scripts\") pod \"keystone-db-create-qz5fx\" (UID: \"69e67ebc-dcc4-4f25-9366-7d8a23256d58\") " pod="openstack/keystone-db-create-qz5fx" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.455045 4809 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ba14ef06-3f68-45a3-8fde-36980cd2d194-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.455130 4809 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ba14ef06-3f68-45a3-8fde-36980cd2d194-console-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.455211 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mh2mm\" (UniqueName: \"kubernetes.io/projected/ba14ef06-3f68-45a3-8fde-36980cd2d194-kube-api-access-mh2mm\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.455222 4809 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ba14ef06-3f68-45a3-8fde-36980cd2d194-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.455231 4809 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ba14ef06-3f68-45a3-8fde-36980cd2d194-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.455239 4809 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ba14ef06-3f68-45a3-8fde-36980cd2d194-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.465323 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-e6aa-account-create-update-pz4lp"] Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.466779 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-e6aa-account-create-update-pz4lp"] Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.466863 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-e6aa-account-create-update-pz4lp" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.469903 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.557810 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zstq8\" (UniqueName: \"kubernetes.io/projected/69e67ebc-dcc4-4f25-9366-7d8a23256d58-kube-api-access-zstq8\") pod \"keystone-db-create-qz5fx\" (UID: \"69e67ebc-dcc4-4f25-9366-7d8a23256d58\") " pod="openstack/keystone-db-create-qz5fx" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.557895 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42kts\" (UniqueName: \"kubernetes.io/projected/fd020248-5e26-4df3-8a57-420613ffc902-kube-api-access-42kts\") pod \"keystone-e6aa-account-create-update-pz4lp\" (UID: \"fd020248-5e26-4df3-8a57-420613ffc902\") " pod="openstack/keystone-e6aa-account-create-update-pz4lp" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.558012 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/69e67ebc-dcc4-4f25-9366-7d8a23256d58-operator-scripts\") pod \"keystone-db-create-qz5fx\" (UID: \"69e67ebc-dcc4-4f25-9366-7d8a23256d58\") " pod="openstack/keystone-db-create-qz5fx" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.558039 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fd020248-5e26-4df3-8a57-420613ffc902-operator-scripts\") pod \"keystone-e6aa-account-create-update-pz4lp\" (UID: \"fd020248-5e26-4df3-8a57-420613ffc902\") " pod="openstack/keystone-e6aa-account-create-update-pz4lp" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.561203 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/69e67ebc-dcc4-4f25-9366-7d8a23256d58-operator-scripts\") pod \"keystone-db-create-qz5fx\" (UID: \"69e67ebc-dcc4-4f25-9366-7d8a23256d58\") " pod="openstack/keystone-db-create-qz5fx" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.568159 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8bfcbfb-3946-420c-b90f-f390887145f0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a8bfcbfb-3946-420c-b90f-f390887145f0" (UID: "a8bfcbfb-3946-420c-b90f-f390887145f0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.579017 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba14ef06-3f68-45a3-8fde-36980cd2d194-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "ba14ef06-3f68-45a3-8fde-36980cd2d194" (UID: "ba14ef06-3f68-45a3-8fde-36980cd2d194"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.600130 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zstq8\" (UniqueName: \"kubernetes.io/projected/69e67ebc-dcc4-4f25-9366-7d8a23256d58-kube-api-access-zstq8\") pod \"keystone-db-create-qz5fx\" (UID: \"69e67ebc-dcc4-4f25-9366-7d8a23256d58\") " pod="openstack/keystone-db-create-qz5fx" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.608613 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8bfcbfb-3946-420c-b90f-f390887145f0-config" (OuterVolumeSpecName: "config") pod "a8bfcbfb-3946-420c-b90f-f390887145f0" (UID: "a8bfcbfb-3946-420c-b90f-f390887145f0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.609342 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f6536b25-1cc1-4f74-bf46-613965ddf185","Type":"ContainerStarted","Data":"b586d911296037f57743670c3c8e87969fa4c7f025ce88b226bef7b8edc0a9ee"} Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.617041 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-6499fbfdc8-929cv_ba14ef06-3f68-45a3-8fde-36980cd2d194/console/0.log" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.617142 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6499fbfdc8-929cv" event={"ID":"ba14ef06-3f68-45a3-8fde-36980cd2d194","Type":"ContainerDied","Data":"9daa8bc4fa7f8980924840bbf78cd6023ac2490b195c4022042729a2a1dcc06d"} Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.617189 4809 scope.go:117] "RemoveContainer" containerID="783e760a1bf33d4a66aaf5565d4bab5a8309f7adb668a85a0e93f84348ca28a4" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.617346 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6499fbfdc8-929cv" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.633773 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"eb1bd4de-6456-4817-a85d-ab0176153f2c","Type":"ContainerStarted","Data":"a90c2aa3e3fe73387e9b264aedea8b1f549bfaa7885df6302588dd069fb8db47"} Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.634024 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.635571 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-gr99p" event={"ID":"a8bfcbfb-3946-420c-b90f-f390887145f0","Type":"ContainerDied","Data":"ffd1005a8f86fc583dc6e3396dc2eb2e0c68d026f81484a8edef9ff0ac96ce7f"} Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.635634 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-gr99p" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.639004 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-8cd6q"] Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.640811 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-8cd6q" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.647873 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-8cd6q"] Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.650006 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1ceabc13-38c5-4425-8d85-c3afa8935243","Type":"ContainerStarted","Data":"162e28f89b66909f6341a5b02785bb02066b04eb4d2c67d23537c8ea8505ae27"} Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.650986 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.659307 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fd020248-5e26-4df3-8a57-420613ffc902-operator-scripts\") pod \"keystone-e6aa-account-create-update-pz4lp\" (UID: \"fd020248-5e26-4df3-8a57-420613ffc902\") " pod="openstack/keystone-e6aa-account-create-update-pz4lp" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.659859 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42kts\" (UniqueName: \"kubernetes.io/projected/fd020248-5e26-4df3-8a57-420613ffc902-kube-api-access-42kts\") pod \"keystone-e6aa-account-create-update-pz4lp\" (UID: \"fd020248-5e26-4df3-8a57-420613ffc902\") " pod="openstack/keystone-e6aa-account-create-update-pz4lp" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.662216 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fd020248-5e26-4df3-8a57-420613ffc902-operator-scripts\") pod \"keystone-e6aa-account-create-update-pz4lp\" (UID: \"fd020248-5e26-4df3-8a57-420613ffc902\") " pod="openstack/keystone-e6aa-account-create-update-pz4lp" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.662427 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8bfcbfb-3946-420c-b90f-f390887145f0-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.662475 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a8bfcbfb-3946-420c-b90f-f390887145f0-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.662491 4809 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ba14ef06-3f68-45a3-8fde-36980cd2d194-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.676793 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-d99d-account-create-update-l7vkv"] Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.678517 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-d99d-account-create-update-l7vkv" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.681299 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.685399 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42kts\" (UniqueName: \"kubernetes.io/projected/fd020248-5e26-4df3-8a57-420613ffc902-kube-api-access-42kts\") pod \"keystone-e6aa-account-create-update-pz4lp\" (UID: \"fd020248-5e26-4df3-8a57-420613ffc902\") " pod="openstack/keystone-e6aa-account-create-update-pz4lp" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.689681 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-qz5fx" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.714421 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-d99d-account-create-update-l7vkv"] Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.731407 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=44.064732624 podStartE2EDuration="1m5.731387211s" podCreationTimestamp="2025-12-06 06:14:32 +0000 UTC" firstStartedPulling="2025-12-06 06:14:34.831456124 +0000 UTC m=+1399.720439066" lastFinishedPulling="2025-12-06 06:14:56.498110711 +0000 UTC m=+1421.387093653" observedRunningTime="2025-12-06 06:15:37.697683532 +0000 UTC m=+1462.586666484" watchObservedRunningTime="2025-12-06 06:15:37.731387211 +0000 UTC m=+1462.620370153" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.746735 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-rksl7"] Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.763960 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwfvt\" (UniqueName: \"kubernetes.io/projected/6b5b8648-fb4d-4b44-9e6d-0fc714e55add-kube-api-access-cwfvt\") pod \"placement-db-create-8cd6q\" (UID: \"6b5b8648-fb4d-4b44-9e6d-0fc714e55add\") " pod="openstack/placement-db-create-8cd6q" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.764357 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53bf9e4c-33f6-4069-b6f1-8f99be5dbe67-operator-scripts\") pod \"placement-d99d-account-create-update-l7vkv\" (UID: \"53bf9e4c-33f6-4069-b6f1-8f99be5dbe67\") " pod="openstack/placement-d99d-account-create-update-l7vkv" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.764417 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ctck\" (UniqueName: \"kubernetes.io/projected/53bf9e4c-33f6-4069-b6f1-8f99be5dbe67-kube-api-access-2ctck\") pod \"placement-d99d-account-create-update-l7vkv\" (UID: \"53bf9e4c-33f6-4069-b6f1-8f99be5dbe67\") " pod="openstack/placement-d99d-account-create-update-l7vkv" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.764567 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6b5b8648-fb4d-4b44-9e6d-0fc714e55add-operator-scripts\") pod \"placement-db-create-8cd6q\" (UID: \"6b5b8648-fb4d-4b44-9e6d-0fc714e55add\") " pod="openstack/placement-db-create-8cd6q" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.789362 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-e6aa-account-create-update-pz4lp" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.797092 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=50.801373443 podStartE2EDuration="1m4.797066742s" podCreationTimestamp="2025-12-06 06:14:33 +0000 UTC" firstStartedPulling="2025-12-06 06:14:42.42519924 +0000 UTC m=+1407.314182172" lastFinishedPulling="2025-12-06 06:14:56.420892529 +0000 UTC m=+1421.309875471" observedRunningTime="2025-12-06 06:15:37.788170101 +0000 UTC m=+1462.677153033" watchObservedRunningTime="2025-12-06 06:15:37.797066742 +0000 UTC m=+1462.686049684" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.853857 4809 scope.go:117] "RemoveContainer" containerID="e4f1e32a64e611e44984494ef945ec0948d0c06be22b8dd98b0210df10ed1d23" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.859427 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-c863-account-create-update-29vzq"] Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.870175 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53bf9e4c-33f6-4069-b6f1-8f99be5dbe67-operator-scripts\") pod \"placement-d99d-account-create-update-l7vkv\" (UID: \"53bf9e4c-33f6-4069-b6f1-8f99be5dbe67\") " pod="openstack/placement-d99d-account-create-update-l7vkv" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.870218 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ctck\" (UniqueName: \"kubernetes.io/projected/53bf9e4c-33f6-4069-b6f1-8f99be5dbe67-kube-api-access-2ctck\") pod \"placement-d99d-account-create-update-l7vkv\" (UID: \"53bf9e4c-33f6-4069-b6f1-8f99be5dbe67\") " pod="openstack/placement-d99d-account-create-update-l7vkv" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.870284 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6b5b8648-fb4d-4b44-9e6d-0fc714e55add-operator-scripts\") pod \"placement-db-create-8cd6q\" (UID: \"6b5b8648-fb4d-4b44-9e6d-0fc714e55add\") " pod="openstack/placement-db-create-8cd6q" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.870536 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwfvt\" (UniqueName: \"kubernetes.io/projected/6b5b8648-fb4d-4b44-9e6d-0fc714e55add-kube-api-access-cwfvt\") pod \"placement-db-create-8cd6q\" (UID: \"6b5b8648-fb4d-4b44-9e6d-0fc714e55add\") " pod="openstack/placement-db-create-8cd6q" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.871897 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6b5b8648-fb4d-4b44-9e6d-0fc714e55add-operator-scripts\") pod \"placement-db-create-8cd6q\" (UID: \"6b5b8648-fb4d-4b44-9e6d-0fc714e55add\") " pod="openstack/placement-db-create-8cd6q" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.872606 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53bf9e4c-33f6-4069-b6f1-8f99be5dbe67-operator-scripts\") pod \"placement-d99d-account-create-update-l7vkv\" (UID: \"53bf9e4c-33f6-4069-b6f1-8f99be5dbe67\") " pod="openstack/placement-d99d-account-create-update-l7vkv" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.896159 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwfvt\" (UniqueName: \"kubernetes.io/projected/6b5b8648-fb4d-4b44-9e6d-0fc714e55add-kube-api-access-cwfvt\") pod \"placement-db-create-8cd6q\" (UID: \"6b5b8648-fb4d-4b44-9e6d-0fc714e55add\") " pod="openstack/placement-db-create-8cd6q" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.898751 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-6499fbfdc8-929cv"] Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.898750 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ctck\" (UniqueName: \"kubernetes.io/projected/53bf9e4c-33f6-4069-b6f1-8f99be5dbe67-kube-api-access-2ctck\") pod \"placement-d99d-account-create-update-l7vkv\" (UID: \"53bf9e4c-33f6-4069-b6f1-8f99be5dbe67\") " pod="openstack/placement-d99d-account-create-update-l7vkv" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.934053 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-6499fbfdc8-929cv"] Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.934068 4809 scope.go:117] "RemoveContainer" containerID="69cad62602dd98f3cd2dbd728a1ff4fcdce7c03ce1e2531ef7756a170e75fa5a" Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.959338 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-gr99p"] Dec 06 06:15:37 crc kubenswrapper[4809]: I1206 06:15:37.977627 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-gr99p"] Dec 06 06:15:38 crc kubenswrapper[4809]: I1206 06:15:38.020049 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-xrxv5"] Dec 06 06:15:38 crc kubenswrapper[4809]: I1206 06:15:38.028242 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-xrxv5" Dec 06 06:15:38 crc kubenswrapper[4809]: I1206 06:15:38.070331 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-xrxv5"] Dec 06 06:15:38 crc kubenswrapper[4809]: I1206 06:15:38.087152 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qswm8\" (UniqueName: \"kubernetes.io/projected/aa09ddab-61ad-4c63-af1c-f974502055f1-kube-api-access-qswm8\") pod \"glance-db-create-xrxv5\" (UID: \"aa09ddab-61ad-4c63-af1c-f974502055f1\") " pod="openstack/glance-db-create-xrxv5" Dec 06 06:15:38 crc kubenswrapper[4809]: I1206 06:15:38.087292 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aa09ddab-61ad-4c63-af1c-f974502055f1-operator-scripts\") pod \"glance-db-create-xrxv5\" (UID: \"aa09ddab-61ad-4c63-af1c-f974502055f1\") " pod="openstack/glance-db-create-xrxv5" Dec 06 06:15:38 crc kubenswrapper[4809]: I1206 06:15:38.095182 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-8cd6q" Dec 06 06:15:38 crc kubenswrapper[4809]: I1206 06:15:38.104121 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-d99d-account-create-update-l7vkv" Dec 06 06:15:38 crc kubenswrapper[4809]: I1206 06:15:38.121117 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-99dc-account-create-update-qf55d"] Dec 06 06:15:38 crc kubenswrapper[4809]: I1206 06:15:38.123686 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-99dc-account-create-update-qf55d" Dec 06 06:15:38 crc kubenswrapper[4809]: I1206 06:15:38.148871 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 06 06:15:38 crc kubenswrapper[4809]: I1206 06:15:38.164336 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-99dc-account-create-update-qf55d"] Dec 06 06:15:38 crc kubenswrapper[4809]: I1206 06:15:38.194573 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmwbc\" (UniqueName: \"kubernetes.io/projected/1c342dde-c9c4-4ead-a4a5-6db383219a32-kube-api-access-pmwbc\") pod \"glance-99dc-account-create-update-qf55d\" (UID: \"1c342dde-c9c4-4ead-a4a5-6db383219a32\") " pod="openstack/glance-99dc-account-create-update-qf55d" Dec 06 06:15:38 crc kubenswrapper[4809]: I1206 06:15:38.194641 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qswm8\" (UniqueName: \"kubernetes.io/projected/aa09ddab-61ad-4c63-af1c-f974502055f1-kube-api-access-qswm8\") pod \"glance-db-create-xrxv5\" (UID: \"aa09ddab-61ad-4c63-af1c-f974502055f1\") " pod="openstack/glance-db-create-xrxv5" Dec 06 06:15:38 crc kubenswrapper[4809]: I1206 06:15:38.194685 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aa09ddab-61ad-4c63-af1c-f974502055f1-operator-scripts\") pod \"glance-db-create-xrxv5\" (UID: \"aa09ddab-61ad-4c63-af1c-f974502055f1\") " pod="openstack/glance-db-create-xrxv5" Dec 06 06:15:38 crc kubenswrapper[4809]: I1206 06:15:38.194712 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c342dde-c9c4-4ead-a4a5-6db383219a32-operator-scripts\") pod \"glance-99dc-account-create-update-qf55d\" (UID: \"1c342dde-c9c4-4ead-a4a5-6db383219a32\") " pod="openstack/glance-99dc-account-create-update-qf55d" Dec 06 06:15:38 crc kubenswrapper[4809]: I1206 06:15:38.201677 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aa09ddab-61ad-4c63-af1c-f974502055f1-operator-scripts\") pod \"glance-db-create-xrxv5\" (UID: \"aa09ddab-61ad-4c63-af1c-f974502055f1\") " pod="openstack/glance-db-create-xrxv5" Dec 06 06:15:38 crc kubenswrapper[4809]: I1206 06:15:38.244722 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qswm8\" (UniqueName: \"kubernetes.io/projected/aa09ddab-61ad-4c63-af1c-f974502055f1-kube-api-access-qswm8\") pod \"glance-db-create-xrxv5\" (UID: \"aa09ddab-61ad-4c63-af1c-f974502055f1\") " pod="openstack/glance-db-create-xrxv5" Dec 06 06:15:38 crc kubenswrapper[4809]: I1206 06:15:38.298052 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmwbc\" (UniqueName: \"kubernetes.io/projected/1c342dde-c9c4-4ead-a4a5-6db383219a32-kube-api-access-pmwbc\") pod \"glance-99dc-account-create-update-qf55d\" (UID: \"1c342dde-c9c4-4ead-a4a5-6db383219a32\") " pod="openstack/glance-99dc-account-create-update-qf55d" Dec 06 06:15:38 crc kubenswrapper[4809]: I1206 06:15:38.298138 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c342dde-c9c4-4ead-a4a5-6db383219a32-operator-scripts\") pod \"glance-99dc-account-create-update-qf55d\" (UID: \"1c342dde-c9c4-4ead-a4a5-6db383219a32\") " pod="openstack/glance-99dc-account-create-update-qf55d" Dec 06 06:15:38 crc kubenswrapper[4809]: I1206 06:15:38.298900 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c342dde-c9c4-4ead-a4a5-6db383219a32-operator-scripts\") pod \"glance-99dc-account-create-update-qf55d\" (UID: \"1c342dde-c9c4-4ead-a4a5-6db383219a32\") " pod="openstack/glance-99dc-account-create-update-qf55d" Dec 06 06:15:38 crc kubenswrapper[4809]: I1206 06:15:38.360946 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmwbc\" (UniqueName: \"kubernetes.io/projected/1c342dde-c9c4-4ead-a4a5-6db383219a32-kube-api-access-pmwbc\") pod \"glance-99dc-account-create-update-qf55d\" (UID: \"1c342dde-c9c4-4ead-a4a5-6db383219a32\") " pod="openstack/glance-99dc-account-create-update-qf55d" Dec 06 06:15:38 crc kubenswrapper[4809]: I1206 06:15:38.460976 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-qz5fx"] Dec 06 06:15:38 crc kubenswrapper[4809]: I1206 06:15:38.461887 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-xrxv5" Dec 06 06:15:38 crc kubenswrapper[4809]: I1206 06:15:38.494786 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-99dc-account-create-update-qf55d" Dec 06 06:15:38 crc kubenswrapper[4809]: I1206 06:15:38.646679 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-e6aa-account-create-update-pz4lp"] Dec 06 06:15:38 crc kubenswrapper[4809]: I1206 06:15:38.690681 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-wgmjm" event={"ID":"5f4f2323-f99a-4d85-8e0f-a0eb968de27b","Type":"ContainerStarted","Data":"b665cf0868eba59053e3941bb7eaa0cd60aad27eda0f5fdb2dda3183bca31273"} Dec 06 06:15:38 crc kubenswrapper[4809]: I1206 06:15:38.698278 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-rksl7" event={"ID":"cc6d16d3-5a2f-4d15-ba47-a25e4608703b","Type":"ContainerStarted","Data":"8b81316c32d7dad976ae169ecdfdc91bb6ef750ed7c1ba63de301ba0edd9128b"} Dec 06 06:15:38 crc kubenswrapper[4809]: I1206 06:15:38.698316 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-rksl7" event={"ID":"cc6d16d3-5a2f-4d15-ba47-a25e4608703b","Type":"ContainerStarted","Data":"e008ba4635da577aac7f51d87d658eef61199d33de6f83c393e21a5636886d10"} Dec 06 06:15:38 crc kubenswrapper[4809]: I1206 06:15:38.711852 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-qz5fx" event={"ID":"69e67ebc-dcc4-4f25-9366-7d8a23256d58","Type":"ContainerStarted","Data":"db5a991c1e3bd14b8870ca877973da28e16a9a5ad7601683632682fe0f686f90"} Dec 06 06:15:38 crc kubenswrapper[4809]: I1206 06:15:38.721609 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-wgmjm" podStartSLOduration=8.307826127 podStartE2EDuration="17.72158551s" podCreationTimestamp="2025-12-06 06:15:21 +0000 UTC" firstStartedPulling="2025-12-06 06:15:27.37385642 +0000 UTC m=+1452.262839362" lastFinishedPulling="2025-12-06 06:15:36.787615803 +0000 UTC m=+1461.676598745" observedRunningTime="2025-12-06 06:15:38.711422396 +0000 UTC m=+1463.600405338" watchObservedRunningTime="2025-12-06 06:15:38.72158551 +0000 UTC m=+1463.610568452" Dec 06 06:15:38 crc kubenswrapper[4809]: I1206 06:15:38.731817 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-c863-account-create-update-29vzq" event={"ID":"2afacc78-69cc-4fd2-b857-60a53a29c650","Type":"ContainerStarted","Data":"fddb70bee35c5a86f4a9f121024eb4be70cdb178cdf33ffdd0cc174764e2e383"} Dec 06 06:15:38 crc kubenswrapper[4809]: I1206 06:15:38.751197 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-openstack-db-create-rksl7" podStartSLOduration=9.751170697 podStartE2EDuration="9.751170697s" podCreationTimestamp="2025-12-06 06:15:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:15:38.745421813 +0000 UTC m=+1463.634404755" watchObservedRunningTime="2025-12-06 06:15:38.751170697 +0000 UTC m=+1463.640153639" Dec 06 06:15:38 crc kubenswrapper[4809]: I1206 06:15:38.871486 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-d99d-account-create-update-l7vkv"] Dec 06 06:15:39 crc kubenswrapper[4809]: I1206 06:15:39.024391 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-8cd6q"] Dec 06 06:15:39 crc kubenswrapper[4809]: I1206 06:15:39.170811 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-99dc-account-create-update-qf55d"] Dec 06 06:15:39 crc kubenswrapper[4809]: I1206 06:15:39.181741 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-xrxv5"] Dec 06 06:15:39 crc kubenswrapper[4809]: I1206 06:15:39.403301 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8bfcbfb-3946-420c-b90f-f390887145f0" path="/var/lib/kubelet/pods/a8bfcbfb-3946-420c-b90f-f390887145f0/volumes" Dec 06 06:15:39 crc kubenswrapper[4809]: I1206 06:15:39.405442 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba14ef06-3f68-45a3-8fde-36980cd2d194" path="/var/lib/kubelet/pods/ba14ef06-3f68-45a3-8fde-36980cd2d194/volumes" Dec 06 06:15:39 crc kubenswrapper[4809]: I1206 06:15:39.406201 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 06 06:15:39 crc kubenswrapper[4809]: I1206 06:15:39.750170 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-qz5fx" event={"ID":"69e67ebc-dcc4-4f25-9366-7d8a23256d58","Type":"ContainerStarted","Data":"2138523ae14f02ce6e2b49b458eee3b5e22269e264269047d930c802c93186c8"} Dec 06 06:15:39 crc kubenswrapper[4809]: I1206 06:15:39.753990 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-c863-account-create-update-29vzq" event={"ID":"2afacc78-69cc-4fd2-b857-60a53a29c650","Type":"ContainerStarted","Data":"a4bac04ada57d950a028ed0c265dc011d7f9f458934984fd696ff4b0f3ae6177"} Dec 06 06:15:39 crc kubenswrapper[4809]: I1206 06:15:39.758733 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-8cd6q" event={"ID":"6b5b8648-fb4d-4b44-9e6d-0fc714e55add","Type":"ContainerStarted","Data":"eed039fa9517b32bd3f5995a4fe2d65bba6ab99deeb6d677d0a7a6498e2d596f"} Dec 06 06:15:39 crc kubenswrapper[4809]: I1206 06:15:39.760623 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-99dc-account-create-update-qf55d" event={"ID":"1c342dde-c9c4-4ead-a4a5-6db383219a32","Type":"ContainerStarted","Data":"3fd12289e7b2b235cdb57e742b514ec97b2d36c7e4fc9d295eba64e96f21c1a4"} Dec 06 06:15:39 crc kubenswrapper[4809]: I1206 06:15:39.760651 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-99dc-account-create-update-qf55d" event={"ID":"1c342dde-c9c4-4ead-a4a5-6db383219a32","Type":"ContainerStarted","Data":"811aa18d7ced9e3e3e3c108a66229b8e1f56138b907c64f23fa136b77cd53039"} Dec 06 06:15:39 crc kubenswrapper[4809]: I1206 06:15:39.762910 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d99d-account-create-update-l7vkv" event={"ID":"53bf9e4c-33f6-4069-b6f1-8f99be5dbe67","Type":"ContainerStarted","Data":"e6e0eac6101c3a70a8f725eed6c9a714e4351da79309ead2e22db9074a7c17fd"} Dec 06 06:15:39 crc kubenswrapper[4809]: I1206 06:15:39.762955 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d99d-account-create-update-l7vkv" event={"ID":"53bf9e4c-33f6-4069-b6f1-8f99be5dbe67","Type":"ContainerStarted","Data":"c941eee8ea0a59c8cb9cb20f5cddcf562ac4e289330eb0776d0ff119605a86e1"} Dec 06 06:15:39 crc kubenswrapper[4809]: I1206 06:15:39.765110 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-e6aa-account-create-update-pz4lp" event={"ID":"fd020248-5e26-4df3-8a57-420613ffc902","Type":"ContainerStarted","Data":"642f19e6ec8cda9ad8da9265a06e7016d564c6600da0eeaaa72682a1f0da5669"} Dec 06 06:15:39 crc kubenswrapper[4809]: I1206 06:15:39.765144 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-e6aa-account-create-update-pz4lp" event={"ID":"fd020248-5e26-4df3-8a57-420613ffc902","Type":"ContainerStarted","Data":"bfa512558a362167737f91fa34c11b207c34c3f2a49173e25fe146d3b72a0c7d"} Dec 06 06:15:39 crc kubenswrapper[4809]: I1206 06:15:39.772070 4809 generic.go:334] "Generic (PLEG): container finished" podID="cc6d16d3-5a2f-4d15-ba47-a25e4608703b" containerID="8b81316c32d7dad976ae169ecdfdc91bb6ef750ed7c1ba63de301ba0edd9128b" exitCode=0 Dec 06 06:15:39 crc kubenswrapper[4809]: I1206 06:15:39.772245 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-rksl7" event={"ID":"cc6d16d3-5a2f-4d15-ba47-a25e4608703b","Type":"ContainerDied","Data":"8b81316c32d7dad976ae169ecdfdc91bb6ef750ed7c1ba63de301ba0edd9128b"} Dec 06 06:15:39 crc kubenswrapper[4809]: I1206 06:15:39.776807 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-xrxv5" event={"ID":"aa09ddab-61ad-4c63-af1c-f974502055f1","Type":"ContainerStarted","Data":"51024097e224727b62b17160225d13777def691fcec814a356b5e421b3ebdb1e"} Dec 06 06:15:39 crc kubenswrapper[4809]: I1206 06:15:39.792289 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-qz5fx" podStartSLOduration=2.79225348 podStartE2EDuration="2.79225348s" podCreationTimestamp="2025-12-06 06:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:15:39.788251992 +0000 UTC m=+1464.677234944" watchObservedRunningTime="2025-12-06 06:15:39.79225348 +0000 UTC m=+1464.681236422" Dec 06 06:15:39 crc kubenswrapper[4809]: I1206 06:15:39.837799 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-c863-account-create-update-29vzq" podStartSLOduration=10.837783347 podStartE2EDuration="10.837783347s" podCreationTimestamp="2025-12-06 06:15:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:15:39.833691916 +0000 UTC m=+1464.722674858" watchObservedRunningTime="2025-12-06 06:15:39.837783347 +0000 UTC m=+1464.726766289" Dec 06 06:15:39 crc kubenswrapper[4809]: I1206 06:15:39.864955 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-99dc-account-create-update-qf55d" podStartSLOduration=1.864923149 podStartE2EDuration="1.864923149s" podCreationTimestamp="2025-12-06 06:15:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:15:39.859530504 +0000 UTC m=+1464.748513446" watchObservedRunningTime="2025-12-06 06:15:39.864923149 +0000 UTC m=+1464.753906091" Dec 06 06:15:39 crc kubenswrapper[4809]: I1206 06:15:39.887403 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-d99d-account-create-update-l7vkv" podStartSLOduration=2.887375625 podStartE2EDuration="2.887375625s" podCreationTimestamp="2025-12-06 06:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:15:39.877189899 +0000 UTC m=+1464.766172841" watchObservedRunningTime="2025-12-06 06:15:39.887375625 +0000 UTC m=+1464.776358567" Dec 06 06:15:39 crc kubenswrapper[4809]: I1206 06:15:39.907401 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-xrxv5" podStartSLOduration=2.907379904 podStartE2EDuration="2.907379904s" podCreationTimestamp="2025-12-06 06:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:15:39.895200225 +0000 UTC m=+1464.784183157" watchObservedRunningTime="2025-12-06 06:15:39.907379904 +0000 UTC m=+1464.796362846" Dec 06 06:15:39 crc kubenswrapper[4809]: I1206 06:15:39.928694 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-e6aa-account-create-update-pz4lp" podStartSLOduration=2.928665627 podStartE2EDuration="2.928665627s" podCreationTimestamp="2025-12-06 06:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:15:39.922485611 +0000 UTC m=+1464.811468553" watchObservedRunningTime="2025-12-06 06:15:39.928665627 +0000 UTC m=+1464.817648569" Dec 06 06:15:40 crc kubenswrapper[4809]: I1206 06:15:40.786337 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-xrxv5" event={"ID":"aa09ddab-61ad-4c63-af1c-f974502055f1","Type":"ContainerStarted","Data":"afa603635846d9db85859e61469f8b9bd848799c589de4b7f49b27be5c74bd2e"} Dec 06 06:15:40 crc kubenswrapper[4809]: I1206 06:15:40.788120 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-8cd6q" event={"ID":"6b5b8648-fb4d-4b44-9e6d-0fc714e55add","Type":"ContainerStarted","Data":"d8104ebb13ec1ef97f7b517163b0dfd873d1d121fae861dfbdb52a22e511eb38"} Dec 06 06:15:40 crc kubenswrapper[4809]: I1206 06:15:40.804690 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-8cd6q" podStartSLOduration=3.804665978 podStartE2EDuration="3.804665978s" podCreationTimestamp="2025-12-06 06:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:15:40.803768074 +0000 UTC m=+1465.692751016" watchObservedRunningTime="2025-12-06 06:15:40.804665978 +0000 UTC m=+1465.693648920" Dec 06 06:15:41 crc kubenswrapper[4809]: I1206 06:15:41.215154 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-rksl7" Dec 06 06:15:41 crc kubenswrapper[4809]: I1206 06:15:41.389513 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9njqg\" (UniqueName: \"kubernetes.io/projected/cc6d16d3-5a2f-4d15-ba47-a25e4608703b-kube-api-access-9njqg\") pod \"cc6d16d3-5a2f-4d15-ba47-a25e4608703b\" (UID: \"cc6d16d3-5a2f-4d15-ba47-a25e4608703b\") " Dec 06 06:15:41 crc kubenswrapper[4809]: I1206 06:15:41.389693 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc6d16d3-5a2f-4d15-ba47-a25e4608703b-operator-scripts\") pod \"cc6d16d3-5a2f-4d15-ba47-a25e4608703b\" (UID: \"cc6d16d3-5a2f-4d15-ba47-a25e4608703b\") " Dec 06 06:15:41 crc kubenswrapper[4809]: I1206 06:15:41.390967 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc6d16d3-5a2f-4d15-ba47-a25e4608703b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cc6d16d3-5a2f-4d15-ba47-a25e4608703b" (UID: "cc6d16d3-5a2f-4d15-ba47-a25e4608703b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:15:41 crc kubenswrapper[4809]: I1206 06:15:41.403565 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc6d16d3-5a2f-4d15-ba47-a25e4608703b-kube-api-access-9njqg" (OuterVolumeSpecName: "kube-api-access-9njqg") pod "cc6d16d3-5a2f-4d15-ba47-a25e4608703b" (UID: "cc6d16d3-5a2f-4d15-ba47-a25e4608703b"). InnerVolumeSpecName "kube-api-access-9njqg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:15:41 crc kubenswrapper[4809]: I1206 06:15:41.492007 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9njqg\" (UniqueName: \"kubernetes.io/projected/cc6d16d3-5a2f-4d15-ba47-a25e4608703b-kube-api-access-9njqg\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:41 crc kubenswrapper[4809]: I1206 06:15:41.492354 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc6d16d3-5a2f-4d15-ba47-a25e4608703b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:41 crc kubenswrapper[4809]: I1206 06:15:41.801305 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-rksl7" event={"ID":"cc6d16d3-5a2f-4d15-ba47-a25e4608703b","Type":"ContainerDied","Data":"e008ba4635da577aac7f51d87d658eef61199d33de6f83c393e21a5636886d10"} Dec 06 06:15:41 crc kubenswrapper[4809]: I1206 06:15:41.801360 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e008ba4635da577aac7f51d87d658eef61199d33de6f83c393e21a5636886d10" Dec 06 06:15:41 crc kubenswrapper[4809]: I1206 06:15:41.801364 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-rksl7" Dec 06 06:15:41 crc kubenswrapper[4809]: I1206 06:15:41.803211 4809 generic.go:334] "Generic (PLEG): container finished" podID="53bf9e4c-33f6-4069-b6f1-8f99be5dbe67" containerID="e6e0eac6101c3a70a8f725eed6c9a714e4351da79309ead2e22db9074a7c17fd" exitCode=0 Dec 06 06:15:41 crc kubenswrapper[4809]: I1206 06:15:41.803296 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d99d-account-create-update-l7vkv" event={"ID":"53bf9e4c-33f6-4069-b6f1-8f99be5dbe67","Type":"ContainerDied","Data":"e6e0eac6101c3a70a8f725eed6c9a714e4351da79309ead2e22db9074a7c17fd"} Dec 06 06:15:41 crc kubenswrapper[4809]: I1206 06:15:41.805315 4809 generic.go:334] "Generic (PLEG): container finished" podID="fd020248-5e26-4df3-8a57-420613ffc902" containerID="642f19e6ec8cda9ad8da9265a06e7016d564c6600da0eeaaa72682a1f0da5669" exitCode=0 Dec 06 06:15:41 crc kubenswrapper[4809]: I1206 06:15:41.805357 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-e6aa-account-create-update-pz4lp" event={"ID":"fd020248-5e26-4df3-8a57-420613ffc902","Type":"ContainerDied","Data":"642f19e6ec8cda9ad8da9265a06e7016d564c6600da0eeaaa72682a1f0da5669"} Dec 06 06:15:41 crc kubenswrapper[4809]: I1206 06:15:41.809610 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f6536b25-1cc1-4f74-bf46-613965ddf185","Type":"ContainerStarted","Data":"c229460d835203ea61a85721814803837875570bf40da7b5ebd574d3453b2ebf"} Dec 06 06:15:41 crc kubenswrapper[4809]: I1206 06:15:41.812076 4809 generic.go:334] "Generic (PLEG): container finished" podID="aa09ddab-61ad-4c63-af1c-f974502055f1" containerID="afa603635846d9db85859e61469f8b9bd848799c589de4b7f49b27be5c74bd2e" exitCode=0 Dec 06 06:15:41 crc kubenswrapper[4809]: I1206 06:15:41.812281 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-xrxv5" event={"ID":"aa09ddab-61ad-4c63-af1c-f974502055f1","Type":"ContainerDied","Data":"afa603635846d9db85859e61469f8b9bd848799c589de4b7f49b27be5c74bd2e"} Dec 06 06:15:41 crc kubenswrapper[4809]: I1206 06:15:41.814554 4809 generic.go:334] "Generic (PLEG): container finished" podID="69e67ebc-dcc4-4f25-9366-7d8a23256d58" containerID="2138523ae14f02ce6e2b49b458eee3b5e22269e264269047d930c802c93186c8" exitCode=0 Dec 06 06:15:41 crc kubenswrapper[4809]: I1206 06:15:41.814616 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-qz5fx" event={"ID":"69e67ebc-dcc4-4f25-9366-7d8a23256d58","Type":"ContainerDied","Data":"2138523ae14f02ce6e2b49b458eee3b5e22269e264269047d930c802c93186c8"} Dec 06 06:15:41 crc kubenswrapper[4809]: I1206 06:15:41.816600 4809 generic.go:334] "Generic (PLEG): container finished" podID="2afacc78-69cc-4fd2-b857-60a53a29c650" containerID="a4bac04ada57d950a028ed0c265dc011d7f9f458934984fd696ff4b0f3ae6177" exitCode=0 Dec 06 06:15:41 crc kubenswrapper[4809]: I1206 06:15:41.816665 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-c863-account-create-update-29vzq" event={"ID":"2afacc78-69cc-4fd2-b857-60a53a29c650","Type":"ContainerDied","Data":"a4bac04ada57d950a028ed0c265dc011d7f9f458934984fd696ff4b0f3ae6177"} Dec 06 06:15:41 crc kubenswrapper[4809]: I1206 06:15:41.817997 4809 generic.go:334] "Generic (PLEG): container finished" podID="6b5b8648-fb4d-4b44-9e6d-0fc714e55add" containerID="d8104ebb13ec1ef97f7b517163b0dfd873d1d121fae861dfbdb52a22e511eb38" exitCode=0 Dec 06 06:15:41 crc kubenswrapper[4809]: I1206 06:15:41.818025 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-8cd6q" event={"ID":"6b5b8648-fb4d-4b44-9e6d-0fc714e55add","Type":"ContainerDied","Data":"d8104ebb13ec1ef97f7b517163b0dfd873d1d121fae861dfbdb52a22e511eb38"} Dec 06 06:15:41 crc kubenswrapper[4809]: I1206 06:15:41.819481 4809 generic.go:334] "Generic (PLEG): container finished" podID="1c342dde-c9c4-4ead-a4a5-6db383219a32" containerID="3fd12289e7b2b235cdb57e742b514ec97b2d36c7e4fc9d295eba64e96f21c1a4" exitCode=0 Dec 06 06:15:41 crc kubenswrapper[4809]: I1206 06:15:41.819521 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-99dc-account-create-update-qf55d" event={"ID":"1c342dde-c9c4-4ead-a4a5-6db383219a32","Type":"ContainerDied","Data":"3fd12289e7b2b235cdb57e742b514ec97b2d36c7e4fc9d295eba64e96f21c1a4"} Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.244353 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-8cd6q" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.377701 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-xhlzm" podUID="6b64391f-74d7-4b69-a898-f7375057c6d5" containerName="ovn-controller" probeResult="failure" output=< Dec 06 06:15:43 crc kubenswrapper[4809]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 06 06:15:43 crc kubenswrapper[4809]: > Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.437892 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cwfvt\" (UniqueName: \"kubernetes.io/projected/6b5b8648-fb4d-4b44-9e6d-0fc714e55add-kube-api-access-cwfvt\") pod \"6b5b8648-fb4d-4b44-9e6d-0fc714e55add\" (UID: \"6b5b8648-fb4d-4b44-9e6d-0fc714e55add\") " Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.438017 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6b5b8648-fb4d-4b44-9e6d-0fc714e55add-operator-scripts\") pod \"6b5b8648-fb4d-4b44-9e6d-0fc714e55add\" (UID: \"6b5b8648-fb4d-4b44-9e6d-0fc714e55add\") " Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.439368 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b5b8648-fb4d-4b44-9e6d-0fc714e55add-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6b5b8648-fb4d-4b44-9e6d-0fc714e55add" (UID: "6b5b8648-fb4d-4b44-9e6d-0fc714e55add"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.444967 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b5b8648-fb4d-4b44-9e6d-0fc714e55add-kube-api-access-cwfvt" (OuterVolumeSpecName: "kube-api-access-cwfvt") pod "6b5b8648-fb4d-4b44-9e6d-0fc714e55add" (UID: "6b5b8648-fb4d-4b44-9e6d-0fc714e55add"). InnerVolumeSpecName "kube-api-access-cwfvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.522471 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-jgqjw" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.522859 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-jgqjw" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.540572 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cwfvt\" (UniqueName: \"kubernetes.io/projected/6b5b8648-fb4d-4b44-9e6d-0fc714e55add-kube-api-access-cwfvt\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.540601 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6b5b8648-fb4d-4b44-9e6d-0fc714e55add-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.544298 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-e6aa-account-create-update-pz4lp" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.564554 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-qz5fx" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.572263 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-d99d-account-create-update-l7vkv" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.642173 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fd020248-5e26-4df3-8a57-420613ffc902-operator-scripts\") pod \"fd020248-5e26-4df3-8a57-420613ffc902\" (UID: \"fd020248-5e26-4df3-8a57-420613ffc902\") " Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.642376 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42kts\" (UniqueName: \"kubernetes.io/projected/fd020248-5e26-4df3-8a57-420613ffc902-kube-api-access-42kts\") pod \"fd020248-5e26-4df3-8a57-420613ffc902\" (UID: \"fd020248-5e26-4df3-8a57-420613ffc902\") " Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.643952 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd020248-5e26-4df3-8a57-420613ffc902-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fd020248-5e26-4df3-8a57-420613ffc902" (UID: "fd020248-5e26-4df3-8a57-420613ffc902"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.652409 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd020248-5e26-4df3-8a57-420613ffc902-kube-api-access-42kts" (OuterVolumeSpecName: "kube-api-access-42kts") pod "fd020248-5e26-4df3-8a57-420613ffc902" (UID: "fd020248-5e26-4df3-8a57-420613ffc902"). InnerVolumeSpecName "kube-api-access-42kts". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.744256 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ctck\" (UniqueName: \"kubernetes.io/projected/53bf9e4c-33f6-4069-b6f1-8f99be5dbe67-kube-api-access-2ctck\") pod \"53bf9e4c-33f6-4069-b6f1-8f99be5dbe67\" (UID: \"53bf9e4c-33f6-4069-b6f1-8f99be5dbe67\") " Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.744346 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/69e67ebc-dcc4-4f25-9366-7d8a23256d58-operator-scripts\") pod \"69e67ebc-dcc4-4f25-9366-7d8a23256d58\" (UID: \"69e67ebc-dcc4-4f25-9366-7d8a23256d58\") " Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.744404 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zstq8\" (UniqueName: \"kubernetes.io/projected/69e67ebc-dcc4-4f25-9366-7d8a23256d58-kube-api-access-zstq8\") pod \"69e67ebc-dcc4-4f25-9366-7d8a23256d58\" (UID: \"69e67ebc-dcc4-4f25-9366-7d8a23256d58\") " Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.745892 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53bf9e4c-33f6-4069-b6f1-8f99be5dbe67-operator-scripts\") pod \"53bf9e4c-33f6-4069-b6f1-8f99be5dbe67\" (UID: \"53bf9e4c-33f6-4069-b6f1-8f99be5dbe67\") " Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.746740 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42kts\" (UniqueName: \"kubernetes.io/projected/fd020248-5e26-4df3-8a57-420613ffc902-kube-api-access-42kts\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.746767 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fd020248-5e26-4df3-8a57-420613ffc902-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.747194 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69e67ebc-dcc4-4f25-9366-7d8a23256d58-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "69e67ebc-dcc4-4f25-9366-7d8a23256d58" (UID: "69e67ebc-dcc4-4f25-9366-7d8a23256d58"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.750958 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53bf9e4c-33f6-4069-b6f1-8f99be5dbe67-kube-api-access-2ctck" (OuterVolumeSpecName: "kube-api-access-2ctck") pod "53bf9e4c-33f6-4069-b6f1-8f99be5dbe67" (UID: "53bf9e4c-33f6-4069-b6f1-8f99be5dbe67"). InnerVolumeSpecName "kube-api-access-2ctck". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.751430 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53bf9e4c-33f6-4069-b6f1-8f99be5dbe67-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "53bf9e4c-33f6-4069-b6f1-8f99be5dbe67" (UID: "53bf9e4c-33f6-4069-b6f1-8f99be5dbe67"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.753686 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69e67ebc-dcc4-4f25-9366-7d8a23256d58-kube-api-access-zstq8" (OuterVolumeSpecName: "kube-api-access-zstq8") pod "69e67ebc-dcc4-4f25-9366-7d8a23256d58" (UID: "69e67ebc-dcc4-4f25-9366-7d8a23256d58"). InnerVolumeSpecName "kube-api-access-zstq8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.848316 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-99dc-account-create-update-qf55d" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.849356 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ctck\" (UniqueName: \"kubernetes.io/projected/53bf9e4c-33f6-4069-b6f1-8f99be5dbe67-kube-api-access-2ctck\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.849388 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/69e67ebc-dcc4-4f25-9366-7d8a23256d58-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.849400 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zstq8\" (UniqueName: \"kubernetes.io/projected/69e67ebc-dcc4-4f25-9366-7d8a23256d58-kube-api-access-zstq8\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.849412 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53bf9e4c-33f6-4069-b6f1-8f99be5dbe67-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.858285 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-c863-account-create-update-29vzq" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.858584 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-99dc-account-create-update-qf55d" event={"ID":"1c342dde-c9c4-4ead-a4a5-6db383219a32","Type":"ContainerDied","Data":"811aa18d7ced9e3e3e3c108a66229b8e1f56138b907c64f23fa136b77cd53039"} Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.858627 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="811aa18d7ced9e3e3e3c108a66229b8e1f56138b907c64f23fa136b77cd53039" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.858709 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-99dc-account-create-update-qf55d" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.865612 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d99d-account-create-update-l7vkv" event={"ID":"53bf9e4c-33f6-4069-b6f1-8f99be5dbe67","Type":"ContainerDied","Data":"c941eee8ea0a59c8cb9cb20f5cddcf562ac4e289330eb0776d0ff119605a86e1"} Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.865663 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c941eee8ea0a59c8cb9cb20f5cddcf562ac4e289330eb0776d0ff119605a86e1" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.865748 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-d99d-account-create-update-l7vkv" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.888335 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-e6aa-account-create-update-pz4lp" event={"ID":"fd020248-5e26-4df3-8a57-420613ffc902","Type":"ContainerDied","Data":"bfa512558a362167737f91fa34c11b207c34c3f2a49173e25fe146d3b72a0c7d"} Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.888457 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bfa512558a362167737f91fa34c11b207c34c3f2a49173e25fe146d3b72a0c7d" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.888612 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-xrxv5" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.888745 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-e6aa-account-create-update-pz4lp" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.895492 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-xrxv5" event={"ID":"aa09ddab-61ad-4c63-af1c-f974502055f1","Type":"ContainerDied","Data":"51024097e224727b62b17160225d13777def691fcec814a356b5e421b3ebdb1e"} Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.895530 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="51024097e224727b62b17160225d13777def691fcec814a356b5e421b3ebdb1e" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.901911 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-qz5fx" event={"ID":"69e67ebc-dcc4-4f25-9366-7d8a23256d58","Type":"ContainerDied","Data":"db5a991c1e3bd14b8870ca877973da28e16a9a5ad7601683632682fe0f686f90"} Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.902116 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db5a991c1e3bd14b8870ca877973da28e16a9a5ad7601683632682fe0f686f90" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.902266 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-qz5fx" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.910847 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-c863-account-create-update-29vzq" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.911408 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-c863-account-create-update-29vzq" event={"ID":"2afacc78-69cc-4fd2-b857-60a53a29c650","Type":"ContainerDied","Data":"fddb70bee35c5a86f4a9f121024eb4be70cdb178cdf33ffdd0cc174764e2e383"} Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.911447 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fddb70bee35c5a86f4a9f121024eb4be70cdb178cdf33ffdd0cc174764e2e383" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.926860 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-8cd6q" event={"ID":"6b5b8648-fb4d-4b44-9e6d-0fc714e55add","Type":"ContainerDied","Data":"eed039fa9517b32bd3f5995a4fe2d65bba6ab99deeb6d677d0a7a6498e2d596f"} Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.927121 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eed039fa9517b32bd3f5995a4fe2d65bba6ab99deeb6d677d0a7a6498e2d596f" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.927280 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-8cd6q" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.939840 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-xhlzm-config-lfqg7"] Dec 06 06:15:43 crc kubenswrapper[4809]: E1206 06:15:43.940623 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2afacc78-69cc-4fd2-b857-60a53a29c650" containerName="mariadb-account-create-update" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.940646 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2afacc78-69cc-4fd2-b857-60a53a29c650" containerName="mariadb-account-create-update" Dec 06 06:15:43 crc kubenswrapper[4809]: E1206 06:15:43.940684 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c342dde-c9c4-4ead-a4a5-6db383219a32" containerName="mariadb-account-create-update" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.940693 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c342dde-c9c4-4ead-a4a5-6db383219a32" containerName="mariadb-account-create-update" Dec 06 06:15:43 crc kubenswrapper[4809]: E1206 06:15:43.940725 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b5b8648-fb4d-4b44-9e6d-0fc714e55add" containerName="mariadb-database-create" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.940752 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b5b8648-fb4d-4b44-9e6d-0fc714e55add" containerName="mariadb-database-create" Dec 06 06:15:43 crc kubenswrapper[4809]: E1206 06:15:43.940765 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd020248-5e26-4df3-8a57-420613ffc902" containerName="mariadb-account-create-update" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.940774 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd020248-5e26-4df3-8a57-420613ffc902" containerName="mariadb-account-create-update" Dec 06 06:15:43 crc kubenswrapper[4809]: E1206 06:15:43.940787 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53bf9e4c-33f6-4069-b6f1-8f99be5dbe67" containerName="mariadb-account-create-update" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.940795 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="53bf9e4c-33f6-4069-b6f1-8f99be5dbe67" containerName="mariadb-account-create-update" Dec 06 06:15:43 crc kubenswrapper[4809]: E1206 06:15:43.940824 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69e67ebc-dcc4-4f25-9366-7d8a23256d58" containerName="mariadb-database-create" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.940833 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="69e67ebc-dcc4-4f25-9366-7d8a23256d58" containerName="mariadb-database-create" Dec 06 06:15:43 crc kubenswrapper[4809]: E1206 06:15:43.940860 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa09ddab-61ad-4c63-af1c-f974502055f1" containerName="mariadb-database-create" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.940868 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa09ddab-61ad-4c63-af1c-f974502055f1" containerName="mariadb-database-create" Dec 06 06:15:43 crc kubenswrapper[4809]: E1206 06:15:43.940881 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc6d16d3-5a2f-4d15-ba47-a25e4608703b" containerName="mariadb-database-create" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.940904 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc6d16d3-5a2f-4d15-ba47-a25e4608703b" containerName="mariadb-database-create" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.941282 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd020248-5e26-4df3-8a57-420613ffc902" containerName="mariadb-account-create-update" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.941311 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa09ddab-61ad-4c63-af1c-f974502055f1" containerName="mariadb-database-create" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.941357 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="2afacc78-69cc-4fd2-b857-60a53a29c650" containerName="mariadb-account-create-update" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.941368 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="69e67ebc-dcc4-4f25-9366-7d8a23256d58" containerName="mariadb-database-create" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.941383 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="53bf9e4c-33f6-4069-b6f1-8f99be5dbe67" containerName="mariadb-account-create-update" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.941396 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b5b8648-fb4d-4b44-9e6d-0fc714e55add" containerName="mariadb-database-create" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.941432 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc6d16d3-5a2f-4d15-ba47-a25e4608703b" containerName="mariadb-database-create" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.941445 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c342dde-c9c4-4ead-a4a5-6db383219a32" containerName="mariadb-account-create-update" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.942644 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xhlzm-config-lfqg7" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.945348 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.955866 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmwbc\" (UniqueName: \"kubernetes.io/projected/1c342dde-c9c4-4ead-a4a5-6db383219a32-kube-api-access-pmwbc\") pod \"1c342dde-c9c4-4ead-a4a5-6db383219a32\" (UID: \"1c342dde-c9c4-4ead-a4a5-6db383219a32\") " Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.956111 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c342dde-c9c4-4ead-a4a5-6db383219a32-operator-scripts\") pod \"1c342dde-c9c4-4ead-a4a5-6db383219a32\" (UID: \"1c342dde-c9c4-4ead-a4a5-6db383219a32\") " Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.958002 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c342dde-c9c4-4ead-a4a5-6db383219a32-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1c342dde-c9c4-4ead-a4a5-6db383219a32" (UID: "1c342dde-c9c4-4ead-a4a5-6db383219a32"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.959080 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d56c7bbe-8035-4412-bd07-ed4197deca90-var-run-ovn\") pod \"ovn-controller-xhlzm-config-lfqg7\" (UID: \"d56c7bbe-8035-4412-bd07-ed4197deca90\") " pod="openstack/ovn-controller-xhlzm-config-lfqg7" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.959192 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d56c7bbe-8035-4412-bd07-ed4197deca90-var-run\") pod \"ovn-controller-xhlzm-config-lfqg7\" (UID: \"d56c7bbe-8035-4412-bd07-ed4197deca90\") " pod="openstack/ovn-controller-xhlzm-config-lfqg7" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.959317 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d56c7bbe-8035-4412-bd07-ed4197deca90-scripts\") pod \"ovn-controller-xhlzm-config-lfqg7\" (UID: \"d56c7bbe-8035-4412-bd07-ed4197deca90\") " pod="openstack/ovn-controller-xhlzm-config-lfqg7" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.959400 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d56c7bbe-8035-4412-bd07-ed4197deca90-additional-scripts\") pod \"ovn-controller-xhlzm-config-lfqg7\" (UID: \"d56c7bbe-8035-4412-bd07-ed4197deca90\") " pod="openstack/ovn-controller-xhlzm-config-lfqg7" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.959442 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjzzn\" (UniqueName: \"kubernetes.io/projected/d56c7bbe-8035-4412-bd07-ed4197deca90-kube-api-access-xjzzn\") pod \"ovn-controller-xhlzm-config-lfqg7\" (UID: \"d56c7bbe-8035-4412-bd07-ed4197deca90\") " pod="openstack/ovn-controller-xhlzm-config-lfqg7" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.959476 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d56c7bbe-8035-4412-bd07-ed4197deca90-var-log-ovn\") pod \"ovn-controller-xhlzm-config-lfqg7\" (UID: \"d56c7bbe-8035-4412-bd07-ed4197deca90\") " pod="openstack/ovn-controller-xhlzm-config-lfqg7" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.959600 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c342dde-c9c4-4ead-a4a5-6db383219a32-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.982501 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c342dde-c9c4-4ead-a4a5-6db383219a32-kube-api-access-pmwbc" (OuterVolumeSpecName: "kube-api-access-pmwbc") pod "1c342dde-c9c4-4ead-a4a5-6db383219a32" (UID: "1c342dde-c9c4-4ead-a4a5-6db383219a32"). InnerVolumeSpecName "kube-api-access-pmwbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:15:43 crc kubenswrapper[4809]: I1206 06:15:43.994030 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-xhlzm-config-lfqg7"] Dec 06 06:15:44 crc kubenswrapper[4809]: I1206 06:15:44.060267 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qswm8\" (UniqueName: \"kubernetes.io/projected/aa09ddab-61ad-4c63-af1c-f974502055f1-kube-api-access-qswm8\") pod \"aa09ddab-61ad-4c63-af1c-f974502055f1\" (UID: \"aa09ddab-61ad-4c63-af1c-f974502055f1\") " Dec 06 06:15:44 crc kubenswrapper[4809]: I1206 06:15:44.060323 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ljkj\" (UniqueName: \"kubernetes.io/projected/2afacc78-69cc-4fd2-b857-60a53a29c650-kube-api-access-7ljkj\") pod \"2afacc78-69cc-4fd2-b857-60a53a29c650\" (UID: \"2afacc78-69cc-4fd2-b857-60a53a29c650\") " Dec 06 06:15:44 crc kubenswrapper[4809]: I1206 06:15:44.060816 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aa09ddab-61ad-4c63-af1c-f974502055f1-operator-scripts\") pod \"aa09ddab-61ad-4c63-af1c-f974502055f1\" (UID: \"aa09ddab-61ad-4c63-af1c-f974502055f1\") " Dec 06 06:15:44 crc kubenswrapper[4809]: I1206 06:15:44.060898 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2afacc78-69cc-4fd2-b857-60a53a29c650-operator-scripts\") pod \"2afacc78-69cc-4fd2-b857-60a53a29c650\" (UID: \"2afacc78-69cc-4fd2-b857-60a53a29c650\") " Dec 06 06:15:44 crc kubenswrapper[4809]: I1206 06:15:44.061303 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2afacc78-69cc-4fd2-b857-60a53a29c650-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2afacc78-69cc-4fd2-b857-60a53a29c650" (UID: "2afacc78-69cc-4fd2-b857-60a53a29c650"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:15:44 crc kubenswrapper[4809]: I1206 06:15:44.061317 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa09ddab-61ad-4c63-af1c-f974502055f1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "aa09ddab-61ad-4c63-af1c-f974502055f1" (UID: "aa09ddab-61ad-4c63-af1c-f974502055f1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:15:44 crc kubenswrapper[4809]: I1206 06:15:44.061667 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d56c7bbe-8035-4412-bd07-ed4197deca90-var-run-ovn\") pod \"ovn-controller-xhlzm-config-lfqg7\" (UID: \"d56c7bbe-8035-4412-bd07-ed4197deca90\") " pod="openstack/ovn-controller-xhlzm-config-lfqg7" Dec 06 06:15:44 crc kubenswrapper[4809]: I1206 06:15:44.062028 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d56c7bbe-8035-4412-bd07-ed4197deca90-var-run-ovn\") pod \"ovn-controller-xhlzm-config-lfqg7\" (UID: \"d56c7bbe-8035-4412-bd07-ed4197deca90\") " pod="openstack/ovn-controller-xhlzm-config-lfqg7" Dec 06 06:15:44 crc kubenswrapper[4809]: I1206 06:15:44.062271 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d56c7bbe-8035-4412-bd07-ed4197deca90-var-run\") pod \"ovn-controller-xhlzm-config-lfqg7\" (UID: \"d56c7bbe-8035-4412-bd07-ed4197deca90\") " pod="openstack/ovn-controller-xhlzm-config-lfqg7" Dec 06 06:15:44 crc kubenswrapper[4809]: I1206 06:15:44.062334 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d56c7bbe-8035-4412-bd07-ed4197deca90-var-run\") pod \"ovn-controller-xhlzm-config-lfqg7\" (UID: \"d56c7bbe-8035-4412-bd07-ed4197deca90\") " pod="openstack/ovn-controller-xhlzm-config-lfqg7" Dec 06 06:15:44 crc kubenswrapper[4809]: I1206 06:15:44.062653 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d56c7bbe-8035-4412-bd07-ed4197deca90-scripts\") pod \"ovn-controller-xhlzm-config-lfqg7\" (UID: \"d56c7bbe-8035-4412-bd07-ed4197deca90\") " pod="openstack/ovn-controller-xhlzm-config-lfqg7" Dec 06 06:15:44 crc kubenswrapper[4809]: I1206 06:15:44.062811 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d56c7bbe-8035-4412-bd07-ed4197deca90-additional-scripts\") pod \"ovn-controller-xhlzm-config-lfqg7\" (UID: \"d56c7bbe-8035-4412-bd07-ed4197deca90\") " pod="openstack/ovn-controller-xhlzm-config-lfqg7" Dec 06 06:15:44 crc kubenswrapper[4809]: I1206 06:15:44.062928 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjzzn\" (UniqueName: \"kubernetes.io/projected/d56c7bbe-8035-4412-bd07-ed4197deca90-kube-api-access-xjzzn\") pod \"ovn-controller-xhlzm-config-lfqg7\" (UID: \"d56c7bbe-8035-4412-bd07-ed4197deca90\") " pod="openstack/ovn-controller-xhlzm-config-lfqg7" Dec 06 06:15:44 crc kubenswrapper[4809]: I1206 06:15:44.063619 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d56c7bbe-8035-4412-bd07-ed4197deca90-var-log-ovn\") pod \"ovn-controller-xhlzm-config-lfqg7\" (UID: \"d56c7bbe-8035-4412-bd07-ed4197deca90\") " pod="openstack/ovn-controller-xhlzm-config-lfqg7" Dec 06 06:15:44 crc kubenswrapper[4809]: I1206 06:15:44.065497 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d56c7bbe-8035-4412-bd07-ed4197deca90-scripts\") pod \"ovn-controller-xhlzm-config-lfqg7\" (UID: \"d56c7bbe-8035-4412-bd07-ed4197deca90\") " pod="openstack/ovn-controller-xhlzm-config-lfqg7" Dec 06 06:15:44 crc kubenswrapper[4809]: I1206 06:15:44.064728 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d56c7bbe-8035-4412-bd07-ed4197deca90-additional-scripts\") pod \"ovn-controller-xhlzm-config-lfqg7\" (UID: \"d56c7bbe-8035-4412-bd07-ed4197deca90\") " pod="openstack/ovn-controller-xhlzm-config-lfqg7" Dec 06 06:15:44 crc kubenswrapper[4809]: I1206 06:15:44.064258 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d56c7bbe-8035-4412-bd07-ed4197deca90-var-log-ovn\") pod \"ovn-controller-xhlzm-config-lfqg7\" (UID: \"d56c7bbe-8035-4412-bd07-ed4197deca90\") " pod="openstack/ovn-controller-xhlzm-config-lfqg7" Dec 06 06:15:44 crc kubenswrapper[4809]: I1206 06:15:44.064947 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2afacc78-69cc-4fd2-b857-60a53a29c650-kube-api-access-7ljkj" (OuterVolumeSpecName: "kube-api-access-7ljkj") pod "2afacc78-69cc-4fd2-b857-60a53a29c650" (UID: "2afacc78-69cc-4fd2-b857-60a53a29c650"). InnerVolumeSpecName "kube-api-access-7ljkj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:15:44 crc kubenswrapper[4809]: I1206 06:15:44.065765 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa09ddab-61ad-4c63-af1c-f974502055f1-kube-api-access-qswm8" (OuterVolumeSpecName: "kube-api-access-qswm8") pod "aa09ddab-61ad-4c63-af1c-f974502055f1" (UID: "aa09ddab-61ad-4c63-af1c-f974502055f1"). InnerVolumeSpecName "kube-api-access-qswm8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:15:44 crc kubenswrapper[4809]: I1206 06:15:44.066071 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmwbc\" (UniqueName: \"kubernetes.io/projected/1c342dde-c9c4-4ead-a4a5-6db383219a32-kube-api-access-pmwbc\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:44 crc kubenswrapper[4809]: I1206 06:15:44.066820 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qswm8\" (UniqueName: \"kubernetes.io/projected/aa09ddab-61ad-4c63-af1c-f974502055f1-kube-api-access-qswm8\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:44 crc kubenswrapper[4809]: I1206 06:15:44.066911 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ljkj\" (UniqueName: \"kubernetes.io/projected/2afacc78-69cc-4fd2-b857-60a53a29c650-kube-api-access-7ljkj\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:44 crc kubenswrapper[4809]: I1206 06:15:44.067615 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aa09ddab-61ad-4c63-af1c-f974502055f1-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:44 crc kubenswrapper[4809]: I1206 06:15:44.067716 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2afacc78-69cc-4fd2-b857-60a53a29c650-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:44 crc kubenswrapper[4809]: I1206 06:15:44.081511 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjzzn\" (UniqueName: \"kubernetes.io/projected/d56c7bbe-8035-4412-bd07-ed4197deca90-kube-api-access-xjzzn\") pod \"ovn-controller-xhlzm-config-lfqg7\" (UID: \"d56c7bbe-8035-4412-bd07-ed4197deca90\") " pod="openstack/ovn-controller-xhlzm-config-lfqg7" Dec 06 06:15:44 crc kubenswrapper[4809]: I1206 06:15:44.283231 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xhlzm-config-lfqg7" Dec 06 06:15:44 crc kubenswrapper[4809]: I1206 06:15:44.909806 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-cj7v2"] Dec 06 06:15:44 crc kubenswrapper[4809]: I1206 06:15:44.911829 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-cj7v2" Dec 06 06:15:44 crc kubenswrapper[4809]: I1206 06:15:44.925157 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-cj7v2"] Dec 06 06:15:44 crc kubenswrapper[4809]: I1206 06:15:44.943105 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-xrxv5" Dec 06 06:15:44 crc kubenswrapper[4809]: I1206 06:15:44.996787 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fdf6cf2e-69cc-4152-9e07-736e0fc4512d-operator-scripts\") pod \"mysqld-exporter-openstack-cell1-db-create-cj7v2\" (UID: \"fdf6cf2e-69cc-4152-9e07-736e0fc4512d\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-cj7v2" Dec 06 06:15:44 crc kubenswrapper[4809]: I1206 06:15:44.996950 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6whdk\" (UniqueName: \"kubernetes.io/projected/fdf6cf2e-69cc-4152-9e07-736e0fc4512d-kube-api-access-6whdk\") pod \"mysqld-exporter-openstack-cell1-db-create-cj7v2\" (UID: \"fdf6cf2e-69cc-4152-9e07-736e0fc4512d\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-cj7v2" Dec 06 06:15:45 crc kubenswrapper[4809]: I1206 06:15:45.006370 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-xhlzm-config-lfqg7"] Dec 06 06:15:45 crc kubenswrapper[4809]: I1206 06:15:45.103033 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6whdk\" (UniqueName: \"kubernetes.io/projected/fdf6cf2e-69cc-4152-9e07-736e0fc4512d-kube-api-access-6whdk\") pod \"mysqld-exporter-openstack-cell1-db-create-cj7v2\" (UID: \"fdf6cf2e-69cc-4152-9e07-736e0fc4512d\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-cj7v2" Dec 06 06:15:45 crc kubenswrapper[4809]: I1206 06:15:45.103398 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fdf6cf2e-69cc-4152-9e07-736e0fc4512d-operator-scripts\") pod \"mysqld-exporter-openstack-cell1-db-create-cj7v2\" (UID: \"fdf6cf2e-69cc-4152-9e07-736e0fc4512d\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-cj7v2" Dec 06 06:15:45 crc kubenswrapper[4809]: I1206 06:15:45.104218 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fdf6cf2e-69cc-4152-9e07-736e0fc4512d-operator-scripts\") pod \"mysqld-exporter-openstack-cell1-db-create-cj7v2\" (UID: \"fdf6cf2e-69cc-4152-9e07-736e0fc4512d\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-cj7v2" Dec 06 06:15:45 crc kubenswrapper[4809]: I1206 06:15:45.126049 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-e068-account-create-update-8bklf"] Dec 06 06:15:45 crc kubenswrapper[4809]: I1206 06:15:45.127534 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-e068-account-create-update-8bklf" Dec 06 06:15:45 crc kubenswrapper[4809]: I1206 06:15:45.130852 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-openstack-cell1-db-secret" Dec 06 06:15:45 crc kubenswrapper[4809]: I1206 06:15:45.152597 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-e068-account-create-update-8bklf"] Dec 06 06:15:45 crc kubenswrapper[4809]: I1206 06:15:45.164039 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6whdk\" (UniqueName: \"kubernetes.io/projected/fdf6cf2e-69cc-4152-9e07-736e0fc4512d-kube-api-access-6whdk\") pod \"mysqld-exporter-openstack-cell1-db-create-cj7v2\" (UID: \"fdf6cf2e-69cc-4152-9e07-736e0fc4512d\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-cj7v2" Dec 06 06:15:45 crc kubenswrapper[4809]: I1206 06:15:45.206711 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcvk2\" (UniqueName: \"kubernetes.io/projected/bdf1730f-cf76-4ea5-b89f-fee3ca2e5533-kube-api-access-hcvk2\") pod \"mysqld-exporter-e068-account-create-update-8bklf\" (UID: \"bdf1730f-cf76-4ea5-b89f-fee3ca2e5533\") " pod="openstack/mysqld-exporter-e068-account-create-update-8bklf" Dec 06 06:15:45 crc kubenswrapper[4809]: I1206 06:15:45.206794 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bdf1730f-cf76-4ea5-b89f-fee3ca2e5533-operator-scripts\") pod \"mysqld-exporter-e068-account-create-update-8bklf\" (UID: \"bdf1730f-cf76-4ea5-b89f-fee3ca2e5533\") " pod="openstack/mysqld-exporter-e068-account-create-update-8bklf" Dec 06 06:15:45 crc kubenswrapper[4809]: I1206 06:15:45.235849 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-cj7v2" Dec 06 06:15:45 crc kubenswrapper[4809]: I1206 06:15:45.308356 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcvk2\" (UniqueName: \"kubernetes.io/projected/bdf1730f-cf76-4ea5-b89f-fee3ca2e5533-kube-api-access-hcvk2\") pod \"mysqld-exporter-e068-account-create-update-8bklf\" (UID: \"bdf1730f-cf76-4ea5-b89f-fee3ca2e5533\") " pod="openstack/mysqld-exporter-e068-account-create-update-8bklf" Dec 06 06:15:45 crc kubenswrapper[4809]: I1206 06:15:45.309716 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bdf1730f-cf76-4ea5-b89f-fee3ca2e5533-operator-scripts\") pod \"mysqld-exporter-e068-account-create-update-8bklf\" (UID: \"bdf1730f-cf76-4ea5-b89f-fee3ca2e5533\") " pod="openstack/mysqld-exporter-e068-account-create-update-8bklf" Dec 06 06:15:45 crc kubenswrapper[4809]: I1206 06:15:45.310644 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bdf1730f-cf76-4ea5-b89f-fee3ca2e5533-operator-scripts\") pod \"mysqld-exporter-e068-account-create-update-8bklf\" (UID: \"bdf1730f-cf76-4ea5-b89f-fee3ca2e5533\") " pod="openstack/mysqld-exporter-e068-account-create-update-8bklf" Dec 06 06:15:45 crc kubenswrapper[4809]: I1206 06:15:45.337018 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcvk2\" (UniqueName: \"kubernetes.io/projected/bdf1730f-cf76-4ea5-b89f-fee3ca2e5533-kube-api-access-hcvk2\") pod \"mysqld-exporter-e068-account-create-update-8bklf\" (UID: \"bdf1730f-cf76-4ea5-b89f-fee3ca2e5533\") " pod="openstack/mysqld-exporter-e068-account-create-update-8bklf" Dec 06 06:15:45 crc kubenswrapper[4809]: I1206 06:15:45.499727 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-e068-account-create-update-8bklf" Dec 06 06:15:45 crc kubenswrapper[4809]: I1206 06:15:45.804401 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-cj7v2"] Dec 06 06:15:45 crc kubenswrapper[4809]: I1206 06:15:45.952860 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-xhlzm-config-lfqg7" event={"ID":"d56c7bbe-8035-4412-bd07-ed4197deca90","Type":"ContainerStarted","Data":"e4e8d2c19a1b6d0c77a22e69bd9bb836d9d3511590e0116c8d1a49be5bc196be"} Dec 06 06:15:45 crc kubenswrapper[4809]: I1206 06:15:45.952911 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-xhlzm-config-lfqg7" event={"ID":"d56c7bbe-8035-4412-bd07-ed4197deca90","Type":"ContainerStarted","Data":"6ba2978523f97263bcee3ede91ee85b2b00c906196b5b3212da7503bc15a94ce"} Dec 06 06:15:45 crc kubenswrapper[4809]: I1206 06:15:45.954909 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-cj7v2" event={"ID":"fdf6cf2e-69cc-4152-9e07-736e0fc4512d","Type":"ContainerStarted","Data":"a418fc6a196b6657a486c4a483cf557407e3030393905587e520c93d4d1ac7c5"} Dec 06 06:15:45 crc kubenswrapper[4809]: I1206 06:15:45.976978 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-xhlzm-config-lfqg7" podStartSLOduration=2.976956433 podStartE2EDuration="2.976956433s" podCreationTimestamp="2025-12-06 06:15:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:15:45.971842115 +0000 UTC m=+1470.860825067" watchObservedRunningTime="2025-12-06 06:15:45.976956433 +0000 UTC m=+1470.865939375" Dec 06 06:15:46 crc kubenswrapper[4809]: I1206 06:15:46.063683 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-e068-account-create-update-8bklf"] Dec 06 06:15:46 crc kubenswrapper[4809]: I1206 06:15:46.975492 4809 generic.go:334] "Generic (PLEG): container finished" podID="bdf1730f-cf76-4ea5-b89f-fee3ca2e5533" containerID="b69305cf8058c4689e05fd5338f34098ca60dc761ac06609ef9e3f17d5e6a45e" exitCode=0 Dec 06 06:15:46 crc kubenswrapper[4809]: I1206 06:15:46.975659 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-e068-account-create-update-8bklf" event={"ID":"bdf1730f-cf76-4ea5-b89f-fee3ca2e5533","Type":"ContainerDied","Data":"b69305cf8058c4689e05fd5338f34098ca60dc761ac06609ef9e3f17d5e6a45e"} Dec 06 06:15:46 crc kubenswrapper[4809]: I1206 06:15:46.975786 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-e068-account-create-update-8bklf" event={"ID":"bdf1730f-cf76-4ea5-b89f-fee3ca2e5533","Type":"ContainerStarted","Data":"db2b20900989f3349d24d3414e423f27bd8e82b9a61ca2edf2cece40a9421423"} Dec 06 06:15:46 crc kubenswrapper[4809]: I1206 06:15:46.992505 4809 generic.go:334] "Generic (PLEG): container finished" podID="d56c7bbe-8035-4412-bd07-ed4197deca90" containerID="e4e8d2c19a1b6d0c77a22e69bd9bb836d9d3511590e0116c8d1a49be5bc196be" exitCode=0 Dec 06 06:15:46 crc kubenswrapper[4809]: I1206 06:15:46.992651 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-xhlzm-config-lfqg7" event={"ID":"d56c7bbe-8035-4412-bd07-ed4197deca90","Type":"ContainerDied","Data":"e4e8d2c19a1b6d0c77a22e69bd9bb836d9d3511590e0116c8d1a49be5bc196be"} Dec 06 06:15:46 crc kubenswrapper[4809]: I1206 06:15:46.998197 4809 generic.go:334] "Generic (PLEG): container finished" podID="fdf6cf2e-69cc-4152-9e07-736e0fc4512d" containerID="4f3773f4d9b40904f6c35e5546fd01aa5b09918b9be31362ce262e7fc4b010d7" exitCode=0 Dec 06 06:15:46 crc kubenswrapper[4809]: I1206 06:15:46.998295 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-cj7v2" event={"ID":"fdf6cf2e-69cc-4152-9e07-736e0fc4512d","Type":"ContainerDied","Data":"4f3773f4d9b40904f6c35e5546fd01aa5b09918b9be31362ce262e7fc4b010d7"} Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.281195 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-mc9z8"] Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.288057 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-mc9z8" Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.290545 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-rzj8j" Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.290747 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.314363 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-mc9z8"] Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.357234 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-xhlzm" Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.391194 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/413a30cf-4b57-44d7-b9bf-3d400bf4b897-db-sync-config-data\") pod \"glance-db-sync-mc9z8\" (UID: \"413a30cf-4b57-44d7-b9bf-3d400bf4b897\") " pod="openstack/glance-db-sync-mc9z8" Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.391542 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/413a30cf-4b57-44d7-b9bf-3d400bf4b897-config-data\") pod \"glance-db-sync-mc9z8\" (UID: \"413a30cf-4b57-44d7-b9bf-3d400bf4b897\") " pod="openstack/glance-db-sync-mc9z8" Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.391702 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/413a30cf-4b57-44d7-b9bf-3d400bf4b897-combined-ca-bundle\") pod \"glance-db-sync-mc9z8\" (UID: \"413a30cf-4b57-44d7-b9bf-3d400bf4b897\") " pod="openstack/glance-db-sync-mc9z8" Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.391778 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75lxt\" (UniqueName: \"kubernetes.io/projected/413a30cf-4b57-44d7-b9bf-3d400bf4b897-kube-api-access-75lxt\") pod \"glance-db-sync-mc9z8\" (UID: \"413a30cf-4b57-44d7-b9bf-3d400bf4b897\") " pod="openstack/glance-db-sync-mc9z8" Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.496673 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/413a30cf-4b57-44d7-b9bf-3d400bf4b897-combined-ca-bundle\") pod \"glance-db-sync-mc9z8\" (UID: \"413a30cf-4b57-44d7-b9bf-3d400bf4b897\") " pod="openstack/glance-db-sync-mc9z8" Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.496739 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75lxt\" (UniqueName: \"kubernetes.io/projected/413a30cf-4b57-44d7-b9bf-3d400bf4b897-kube-api-access-75lxt\") pod \"glance-db-sync-mc9z8\" (UID: \"413a30cf-4b57-44d7-b9bf-3d400bf4b897\") " pod="openstack/glance-db-sync-mc9z8" Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.497011 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/413a30cf-4b57-44d7-b9bf-3d400bf4b897-db-sync-config-data\") pod \"glance-db-sync-mc9z8\" (UID: \"413a30cf-4b57-44d7-b9bf-3d400bf4b897\") " pod="openstack/glance-db-sync-mc9z8" Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.497071 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/413a30cf-4b57-44d7-b9bf-3d400bf4b897-config-data\") pod \"glance-db-sync-mc9z8\" (UID: \"413a30cf-4b57-44d7-b9bf-3d400bf4b897\") " pod="openstack/glance-db-sync-mc9z8" Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.507179 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/413a30cf-4b57-44d7-b9bf-3d400bf4b897-combined-ca-bundle\") pod \"glance-db-sync-mc9z8\" (UID: \"413a30cf-4b57-44d7-b9bf-3d400bf4b897\") " pod="openstack/glance-db-sync-mc9z8" Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.511462 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/413a30cf-4b57-44d7-b9bf-3d400bf4b897-config-data\") pod \"glance-db-sync-mc9z8\" (UID: \"413a30cf-4b57-44d7-b9bf-3d400bf4b897\") " pod="openstack/glance-db-sync-mc9z8" Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.519643 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/413a30cf-4b57-44d7-b9bf-3d400bf4b897-db-sync-config-data\") pod \"glance-db-sync-mc9z8\" (UID: \"413a30cf-4b57-44d7-b9bf-3d400bf4b897\") " pod="openstack/glance-db-sync-mc9z8" Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.529058 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75lxt\" (UniqueName: \"kubernetes.io/projected/413a30cf-4b57-44d7-b9bf-3d400bf4b897-kube-api-access-75lxt\") pod \"glance-db-sync-mc9z8\" (UID: \"413a30cf-4b57-44d7-b9bf-3d400bf4b897\") " pod="openstack/glance-db-sync-mc9z8" Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.605955 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-e068-account-create-update-8bklf" Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.618715 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-mc9z8" Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.689859 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-cj7v2" Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.690763 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xhlzm-config-lfqg7" Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.703927 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bdf1730f-cf76-4ea5-b89f-fee3ca2e5533-operator-scripts\") pod \"bdf1730f-cf76-4ea5-b89f-fee3ca2e5533\" (UID: \"bdf1730f-cf76-4ea5-b89f-fee3ca2e5533\") " Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.704226 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hcvk2\" (UniqueName: \"kubernetes.io/projected/bdf1730f-cf76-4ea5-b89f-fee3ca2e5533-kube-api-access-hcvk2\") pod \"bdf1730f-cf76-4ea5-b89f-fee3ca2e5533\" (UID: \"bdf1730f-cf76-4ea5-b89f-fee3ca2e5533\") " Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.705227 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bdf1730f-cf76-4ea5-b89f-fee3ca2e5533-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bdf1730f-cf76-4ea5-b89f-fee3ca2e5533" (UID: "bdf1730f-cf76-4ea5-b89f-fee3ca2e5533"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.711031 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bdf1730f-cf76-4ea5-b89f-fee3ca2e5533-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.715746 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdf1730f-cf76-4ea5-b89f-fee3ca2e5533-kube-api-access-hcvk2" (OuterVolumeSpecName: "kube-api-access-hcvk2") pod "bdf1730f-cf76-4ea5-b89f-fee3ca2e5533" (UID: "bdf1730f-cf76-4ea5-b89f-fee3ca2e5533"). InnerVolumeSpecName "kube-api-access-hcvk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.824893 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fdf6cf2e-69cc-4152-9e07-736e0fc4512d-operator-scripts\") pod \"fdf6cf2e-69cc-4152-9e07-736e0fc4512d\" (UID: \"fdf6cf2e-69cc-4152-9e07-736e0fc4512d\") " Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.825075 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d56c7bbe-8035-4412-bd07-ed4197deca90-var-log-ovn\") pod \"d56c7bbe-8035-4412-bd07-ed4197deca90\" (UID: \"d56c7bbe-8035-4412-bd07-ed4197deca90\") " Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.825232 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6whdk\" (UniqueName: \"kubernetes.io/projected/fdf6cf2e-69cc-4152-9e07-736e0fc4512d-kube-api-access-6whdk\") pod \"fdf6cf2e-69cc-4152-9e07-736e0fc4512d\" (UID: \"fdf6cf2e-69cc-4152-9e07-736e0fc4512d\") " Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.825281 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d56c7bbe-8035-4412-bd07-ed4197deca90-var-run-ovn\") pod \"d56c7bbe-8035-4412-bd07-ed4197deca90\" (UID: \"d56c7bbe-8035-4412-bd07-ed4197deca90\") " Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.825300 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d56c7bbe-8035-4412-bd07-ed4197deca90-additional-scripts\") pod \"d56c7bbe-8035-4412-bd07-ed4197deca90\" (UID: \"d56c7bbe-8035-4412-bd07-ed4197deca90\") " Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.825360 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d56c7bbe-8035-4412-bd07-ed4197deca90-scripts\") pod \"d56c7bbe-8035-4412-bd07-ed4197deca90\" (UID: \"d56c7bbe-8035-4412-bd07-ed4197deca90\") " Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.825427 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d56c7bbe-8035-4412-bd07-ed4197deca90-var-run\") pod \"d56c7bbe-8035-4412-bd07-ed4197deca90\" (UID: \"d56c7bbe-8035-4412-bd07-ed4197deca90\") " Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.825512 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xjzzn\" (UniqueName: \"kubernetes.io/projected/d56c7bbe-8035-4412-bd07-ed4197deca90-kube-api-access-xjzzn\") pod \"d56c7bbe-8035-4412-bd07-ed4197deca90\" (UID: \"d56c7bbe-8035-4412-bd07-ed4197deca90\") " Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.826335 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hcvk2\" (UniqueName: \"kubernetes.io/projected/bdf1730f-cf76-4ea5-b89f-fee3ca2e5533-kube-api-access-hcvk2\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.827552 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d56c7bbe-8035-4412-bd07-ed4197deca90-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "d56c7bbe-8035-4412-bd07-ed4197deca90" (UID: "d56c7bbe-8035-4412-bd07-ed4197deca90"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.827699 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d56c7bbe-8035-4412-bd07-ed4197deca90-var-run" (OuterVolumeSpecName: "var-run") pod "d56c7bbe-8035-4412-bd07-ed4197deca90" (UID: "d56c7bbe-8035-4412-bd07-ed4197deca90"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.827504 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d56c7bbe-8035-4412-bd07-ed4197deca90-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "d56c7bbe-8035-4412-bd07-ed4197deca90" (UID: "d56c7bbe-8035-4412-bd07-ed4197deca90"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.828379 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fdf6cf2e-69cc-4152-9e07-736e0fc4512d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fdf6cf2e-69cc-4152-9e07-736e0fc4512d" (UID: "fdf6cf2e-69cc-4152-9e07-736e0fc4512d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.828727 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d56c7bbe-8035-4412-bd07-ed4197deca90-scripts" (OuterVolumeSpecName: "scripts") pod "d56c7bbe-8035-4412-bd07-ed4197deca90" (UID: "d56c7bbe-8035-4412-bd07-ed4197deca90"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.831171 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d56c7bbe-8035-4412-bd07-ed4197deca90-kube-api-access-xjzzn" (OuterVolumeSpecName: "kube-api-access-xjzzn") pod "d56c7bbe-8035-4412-bd07-ed4197deca90" (UID: "d56c7bbe-8035-4412-bd07-ed4197deca90"). InnerVolumeSpecName "kube-api-access-xjzzn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.831663 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdf6cf2e-69cc-4152-9e07-736e0fc4512d-kube-api-access-6whdk" (OuterVolumeSpecName: "kube-api-access-6whdk") pod "fdf6cf2e-69cc-4152-9e07-736e0fc4512d" (UID: "fdf6cf2e-69cc-4152-9e07-736e0fc4512d"). InnerVolumeSpecName "kube-api-access-6whdk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.834302 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d56c7bbe-8035-4412-bd07-ed4197deca90-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "d56c7bbe-8035-4412-bd07-ed4197deca90" (UID: "d56c7bbe-8035-4412-bd07-ed4197deca90"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.929045 4809 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d56c7bbe-8035-4412-bd07-ed4197deca90-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.929080 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6whdk\" (UniqueName: \"kubernetes.io/projected/fdf6cf2e-69cc-4152-9e07-736e0fc4512d-kube-api-access-6whdk\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.929095 4809 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d56c7bbe-8035-4412-bd07-ed4197deca90-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.929107 4809 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d56c7bbe-8035-4412-bd07-ed4197deca90-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.929118 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d56c7bbe-8035-4412-bd07-ed4197deca90-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.929128 4809 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d56c7bbe-8035-4412-bd07-ed4197deca90-var-run\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.929139 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xjzzn\" (UniqueName: \"kubernetes.io/projected/d56c7bbe-8035-4412-bd07-ed4197deca90-kube-api-access-xjzzn\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:48 crc kubenswrapper[4809]: I1206 06:15:48.929152 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fdf6cf2e-69cc-4152-9e07-736e0fc4512d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:49 crc kubenswrapper[4809]: I1206 06:15:49.028923 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-cj7v2" event={"ID":"fdf6cf2e-69cc-4152-9e07-736e0fc4512d","Type":"ContainerDied","Data":"a418fc6a196b6657a486c4a483cf557407e3030393905587e520c93d4d1ac7c5"} Dec 06 06:15:49 crc kubenswrapper[4809]: I1206 06:15:49.028980 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a418fc6a196b6657a486c4a483cf557407e3030393905587e520c93d4d1ac7c5" Dec 06 06:15:49 crc kubenswrapper[4809]: I1206 06:15:49.028981 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-cj7v2" Dec 06 06:15:49 crc kubenswrapper[4809]: I1206 06:15:49.030457 4809 generic.go:334] "Generic (PLEG): container finished" podID="5f4f2323-f99a-4d85-8e0f-a0eb968de27b" containerID="b665cf0868eba59053e3941bb7eaa0cd60aad27eda0f5fdb2dda3183bca31273" exitCode=0 Dec 06 06:15:49 crc kubenswrapper[4809]: I1206 06:15:49.030512 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-wgmjm" event={"ID":"5f4f2323-f99a-4d85-8e0f-a0eb968de27b","Type":"ContainerDied","Data":"b665cf0868eba59053e3941bb7eaa0cd60aad27eda0f5fdb2dda3183bca31273"} Dec 06 06:15:49 crc kubenswrapper[4809]: I1206 06:15:49.033307 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f6536b25-1cc1-4f74-bf46-613965ddf185","Type":"ContainerStarted","Data":"4bed7b24aeea6ed8999cf873386e191f0e016021ab10c642384864d9c8b535c6"} Dec 06 06:15:49 crc kubenswrapper[4809]: I1206 06:15:49.034607 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-e068-account-create-update-8bklf" event={"ID":"bdf1730f-cf76-4ea5-b89f-fee3ca2e5533","Type":"ContainerDied","Data":"db2b20900989f3349d24d3414e423f27bd8e82b9a61ca2edf2cece40a9421423"} Dec 06 06:15:49 crc kubenswrapper[4809]: I1206 06:15:49.034640 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db2b20900989f3349d24d3414e423f27bd8e82b9a61ca2edf2cece40a9421423" Dec 06 06:15:49 crc kubenswrapper[4809]: I1206 06:15:49.034690 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-e068-account-create-update-8bklf" Dec 06 06:15:49 crc kubenswrapper[4809]: I1206 06:15:49.036409 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-xhlzm-config-lfqg7" event={"ID":"d56c7bbe-8035-4412-bd07-ed4197deca90","Type":"ContainerDied","Data":"6ba2978523f97263bcee3ede91ee85b2b00c906196b5b3212da7503bc15a94ce"} Dec 06 06:15:49 crc kubenswrapper[4809]: I1206 06:15:49.036498 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6ba2978523f97263bcee3ede91ee85b2b00c906196b5b3212da7503bc15a94ce" Dec 06 06:15:49 crc kubenswrapper[4809]: I1206 06:15:49.036589 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xhlzm-config-lfqg7" Dec 06 06:15:49 crc kubenswrapper[4809]: I1206 06:15:49.102182 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=20.842488968 podStartE2EDuration="1m10.10216329s" podCreationTimestamp="2025-12-06 06:14:39 +0000 UTC" firstStartedPulling="2025-12-06 06:14:58.910420396 +0000 UTC m=+1423.799403338" lastFinishedPulling="2025-12-06 06:15:48.170094718 +0000 UTC m=+1473.059077660" observedRunningTime="2025-12-06 06:15:49.094898765 +0000 UTC m=+1473.983881717" watchObservedRunningTime="2025-12-06 06:15:49.10216329 +0000 UTC m=+1473.991146232" Dec 06 06:15:49 crc kubenswrapper[4809]: I1206 06:15:49.205792 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-xhlzm-config-lfqg7"] Dec 06 06:15:49 crc kubenswrapper[4809]: I1206 06:15:49.213801 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-xhlzm-config-lfqg7"] Dec 06 06:15:49 crc kubenswrapper[4809]: I1206 06:15:49.273306 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-mc9z8"] Dec 06 06:15:49 crc kubenswrapper[4809]: I1206 06:15:49.329471 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-xhlzm-config-jnsd4"] Dec 06 06:15:49 crc kubenswrapper[4809]: E1206 06:15:49.330039 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d56c7bbe-8035-4412-bd07-ed4197deca90" containerName="ovn-config" Dec 06 06:15:49 crc kubenswrapper[4809]: I1206 06:15:49.330057 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d56c7bbe-8035-4412-bd07-ed4197deca90" containerName="ovn-config" Dec 06 06:15:49 crc kubenswrapper[4809]: E1206 06:15:49.330074 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdf6cf2e-69cc-4152-9e07-736e0fc4512d" containerName="mariadb-database-create" Dec 06 06:15:49 crc kubenswrapper[4809]: I1206 06:15:49.330082 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdf6cf2e-69cc-4152-9e07-736e0fc4512d" containerName="mariadb-database-create" Dec 06 06:15:49 crc kubenswrapper[4809]: E1206 06:15:49.330098 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdf1730f-cf76-4ea5-b89f-fee3ca2e5533" containerName="mariadb-account-create-update" Dec 06 06:15:49 crc kubenswrapper[4809]: I1206 06:15:49.330106 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdf1730f-cf76-4ea5-b89f-fee3ca2e5533" containerName="mariadb-account-create-update" Dec 06 06:15:49 crc kubenswrapper[4809]: I1206 06:15:49.330314 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdf1730f-cf76-4ea5-b89f-fee3ca2e5533" containerName="mariadb-account-create-update" Dec 06 06:15:49 crc kubenswrapper[4809]: I1206 06:15:49.330329 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdf6cf2e-69cc-4152-9e07-736e0fc4512d" containerName="mariadb-database-create" Dec 06 06:15:49 crc kubenswrapper[4809]: I1206 06:15:49.330341 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="d56c7bbe-8035-4412-bd07-ed4197deca90" containerName="ovn-config" Dec 06 06:15:49 crc kubenswrapper[4809]: I1206 06:15:49.331078 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xhlzm-config-jnsd4" Dec 06 06:15:49 crc kubenswrapper[4809]: I1206 06:15:49.335245 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 06 06:15:49 crc kubenswrapper[4809]: I1206 06:15:49.360411 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-xhlzm-config-jnsd4"] Dec 06 06:15:49 crc kubenswrapper[4809]: I1206 06:15:49.408634 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d56c7bbe-8035-4412-bd07-ed4197deca90" path="/var/lib/kubelet/pods/d56c7bbe-8035-4412-bd07-ed4197deca90/volumes" Dec 06 06:15:49 crc kubenswrapper[4809]: I1206 06:15:49.469225 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjkgp\" (UniqueName: \"kubernetes.io/projected/7ebbbe4b-4be4-4141-9be9-b73ac12741cb-kube-api-access-qjkgp\") pod \"ovn-controller-xhlzm-config-jnsd4\" (UID: \"7ebbbe4b-4be4-4141-9be9-b73ac12741cb\") " pod="openstack/ovn-controller-xhlzm-config-jnsd4" Dec 06 06:15:49 crc kubenswrapper[4809]: I1206 06:15:49.469333 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7ebbbe4b-4be4-4141-9be9-b73ac12741cb-additional-scripts\") pod \"ovn-controller-xhlzm-config-jnsd4\" (UID: \"7ebbbe4b-4be4-4141-9be9-b73ac12741cb\") " pod="openstack/ovn-controller-xhlzm-config-jnsd4" Dec 06 06:15:49 crc kubenswrapper[4809]: I1206 06:15:49.469406 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7ebbbe4b-4be4-4141-9be9-b73ac12741cb-scripts\") pod \"ovn-controller-xhlzm-config-jnsd4\" (UID: \"7ebbbe4b-4be4-4141-9be9-b73ac12741cb\") " pod="openstack/ovn-controller-xhlzm-config-jnsd4" Dec 06 06:15:49 crc kubenswrapper[4809]: I1206 06:15:49.469456 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7ebbbe4b-4be4-4141-9be9-b73ac12741cb-var-run\") pod \"ovn-controller-xhlzm-config-jnsd4\" (UID: \"7ebbbe4b-4be4-4141-9be9-b73ac12741cb\") " pod="openstack/ovn-controller-xhlzm-config-jnsd4" Dec 06 06:15:49 crc kubenswrapper[4809]: I1206 06:15:49.469516 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7ebbbe4b-4be4-4141-9be9-b73ac12741cb-var-run-ovn\") pod \"ovn-controller-xhlzm-config-jnsd4\" (UID: \"7ebbbe4b-4be4-4141-9be9-b73ac12741cb\") " pod="openstack/ovn-controller-xhlzm-config-jnsd4" Dec 06 06:15:49 crc kubenswrapper[4809]: I1206 06:15:49.469563 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7ebbbe4b-4be4-4141-9be9-b73ac12741cb-var-log-ovn\") pod \"ovn-controller-xhlzm-config-jnsd4\" (UID: \"7ebbbe4b-4be4-4141-9be9-b73ac12741cb\") " pod="openstack/ovn-controller-xhlzm-config-jnsd4" Dec 06 06:15:49 crc kubenswrapper[4809]: I1206 06:15:49.572298 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7ebbbe4b-4be4-4141-9be9-b73ac12741cb-var-run-ovn\") pod \"ovn-controller-xhlzm-config-jnsd4\" (UID: \"7ebbbe4b-4be4-4141-9be9-b73ac12741cb\") " pod="openstack/ovn-controller-xhlzm-config-jnsd4" Dec 06 06:15:49 crc kubenswrapper[4809]: I1206 06:15:49.572395 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7ebbbe4b-4be4-4141-9be9-b73ac12741cb-var-log-ovn\") pod \"ovn-controller-xhlzm-config-jnsd4\" (UID: \"7ebbbe4b-4be4-4141-9be9-b73ac12741cb\") " pod="openstack/ovn-controller-xhlzm-config-jnsd4" Dec 06 06:15:49 crc kubenswrapper[4809]: I1206 06:15:49.572482 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjkgp\" (UniqueName: \"kubernetes.io/projected/7ebbbe4b-4be4-4141-9be9-b73ac12741cb-kube-api-access-qjkgp\") pod \"ovn-controller-xhlzm-config-jnsd4\" (UID: \"7ebbbe4b-4be4-4141-9be9-b73ac12741cb\") " pod="openstack/ovn-controller-xhlzm-config-jnsd4" Dec 06 06:15:49 crc kubenswrapper[4809]: I1206 06:15:49.572541 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7ebbbe4b-4be4-4141-9be9-b73ac12741cb-additional-scripts\") pod \"ovn-controller-xhlzm-config-jnsd4\" (UID: \"7ebbbe4b-4be4-4141-9be9-b73ac12741cb\") " pod="openstack/ovn-controller-xhlzm-config-jnsd4" Dec 06 06:15:49 crc kubenswrapper[4809]: I1206 06:15:49.572593 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7ebbbe4b-4be4-4141-9be9-b73ac12741cb-scripts\") pod \"ovn-controller-xhlzm-config-jnsd4\" (UID: \"7ebbbe4b-4be4-4141-9be9-b73ac12741cb\") " pod="openstack/ovn-controller-xhlzm-config-jnsd4" Dec 06 06:15:49 crc kubenswrapper[4809]: I1206 06:15:49.572624 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7ebbbe4b-4be4-4141-9be9-b73ac12741cb-var-run\") pod \"ovn-controller-xhlzm-config-jnsd4\" (UID: \"7ebbbe4b-4be4-4141-9be9-b73ac12741cb\") " pod="openstack/ovn-controller-xhlzm-config-jnsd4" Dec 06 06:15:49 crc kubenswrapper[4809]: I1206 06:15:49.572948 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7ebbbe4b-4be4-4141-9be9-b73ac12741cb-var-run\") pod \"ovn-controller-xhlzm-config-jnsd4\" (UID: \"7ebbbe4b-4be4-4141-9be9-b73ac12741cb\") " pod="openstack/ovn-controller-xhlzm-config-jnsd4" Dec 06 06:15:49 crc kubenswrapper[4809]: I1206 06:15:49.572757 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7ebbbe4b-4be4-4141-9be9-b73ac12741cb-var-log-ovn\") pod \"ovn-controller-xhlzm-config-jnsd4\" (UID: \"7ebbbe4b-4be4-4141-9be9-b73ac12741cb\") " pod="openstack/ovn-controller-xhlzm-config-jnsd4" Dec 06 06:15:49 crc kubenswrapper[4809]: I1206 06:15:49.572667 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7ebbbe4b-4be4-4141-9be9-b73ac12741cb-var-run-ovn\") pod \"ovn-controller-xhlzm-config-jnsd4\" (UID: \"7ebbbe4b-4be4-4141-9be9-b73ac12741cb\") " pod="openstack/ovn-controller-xhlzm-config-jnsd4" Dec 06 06:15:49 crc kubenswrapper[4809]: I1206 06:15:49.573466 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7ebbbe4b-4be4-4141-9be9-b73ac12741cb-additional-scripts\") pod \"ovn-controller-xhlzm-config-jnsd4\" (UID: \"7ebbbe4b-4be4-4141-9be9-b73ac12741cb\") " pod="openstack/ovn-controller-xhlzm-config-jnsd4" Dec 06 06:15:49 crc kubenswrapper[4809]: I1206 06:15:49.574656 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7ebbbe4b-4be4-4141-9be9-b73ac12741cb-scripts\") pod \"ovn-controller-xhlzm-config-jnsd4\" (UID: \"7ebbbe4b-4be4-4141-9be9-b73ac12741cb\") " pod="openstack/ovn-controller-xhlzm-config-jnsd4" Dec 06 06:15:49 crc kubenswrapper[4809]: I1206 06:15:49.590522 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjkgp\" (UniqueName: \"kubernetes.io/projected/7ebbbe4b-4be4-4141-9be9-b73ac12741cb-kube-api-access-qjkgp\") pod \"ovn-controller-xhlzm-config-jnsd4\" (UID: \"7ebbbe4b-4be4-4141-9be9-b73ac12741cb\") " pod="openstack/ovn-controller-xhlzm-config-jnsd4" Dec 06 06:15:49 crc kubenswrapper[4809]: I1206 06:15:49.705373 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xhlzm-config-jnsd4" Dec 06 06:15:50 crc kubenswrapper[4809]: I1206 06:15:50.046499 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-mc9z8" event={"ID":"413a30cf-4b57-44d7-b9bf-3d400bf4b897","Type":"ContainerStarted","Data":"b9cf3c39badf7c33ac4e9c51dbb256bdb00b7f12f045eb32a6b04feef00fdd38"} Dec 06 06:15:50 crc kubenswrapper[4809]: I1206 06:15:50.213116 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-xhlzm-config-jnsd4"] Dec 06 06:15:50 crc kubenswrapper[4809]: I1206 06:15:50.469457 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-0"] Dec 06 06:15:50 crc kubenswrapper[4809]: I1206 06:15:50.471341 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 06 06:15:50 crc kubenswrapper[4809]: I1206 06:15:50.475403 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-config-data" Dec 06 06:15:50 crc kubenswrapper[4809]: I1206 06:15:50.500322 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lptf5\" (UniqueName: \"kubernetes.io/projected/3563a545-b965-4c74-a183-9f8da5876aee-kube-api-access-lptf5\") pod \"mysqld-exporter-0\" (UID: \"3563a545-b965-4c74-a183-9f8da5876aee\") " pod="openstack/mysqld-exporter-0" Dec 06 06:15:50 crc kubenswrapper[4809]: I1206 06:15:50.500556 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3563a545-b965-4c74-a183-9f8da5876aee-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"3563a545-b965-4c74-a183-9f8da5876aee\") " pod="openstack/mysqld-exporter-0" Dec 06 06:15:50 crc kubenswrapper[4809]: I1206 06:15:50.500638 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3563a545-b965-4c74-a183-9f8da5876aee-config-data\") pod \"mysqld-exporter-0\" (UID: \"3563a545-b965-4c74-a183-9f8da5876aee\") " pod="openstack/mysqld-exporter-0" Dec 06 06:15:50 crc kubenswrapper[4809]: I1206 06:15:50.515786 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 06 06:15:50 crc kubenswrapper[4809]: I1206 06:15:50.602861 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3563a545-b965-4c74-a183-9f8da5876aee-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"3563a545-b965-4c74-a183-9f8da5876aee\") " pod="openstack/mysqld-exporter-0" Dec 06 06:15:50 crc kubenswrapper[4809]: I1206 06:15:50.603006 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3563a545-b965-4c74-a183-9f8da5876aee-config-data\") pod \"mysqld-exporter-0\" (UID: \"3563a545-b965-4c74-a183-9f8da5876aee\") " pod="openstack/mysqld-exporter-0" Dec 06 06:15:50 crc kubenswrapper[4809]: I1206 06:15:50.603079 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lptf5\" (UniqueName: \"kubernetes.io/projected/3563a545-b965-4c74-a183-9f8da5876aee-kube-api-access-lptf5\") pod \"mysqld-exporter-0\" (UID: \"3563a545-b965-4c74-a183-9f8da5876aee\") " pod="openstack/mysqld-exporter-0" Dec 06 06:15:50 crc kubenswrapper[4809]: I1206 06:15:50.612909 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3563a545-b965-4c74-a183-9f8da5876aee-config-data\") pod \"mysqld-exporter-0\" (UID: \"3563a545-b965-4c74-a183-9f8da5876aee\") " pod="openstack/mysqld-exporter-0" Dec 06 06:15:50 crc kubenswrapper[4809]: I1206 06:15:50.614363 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3563a545-b965-4c74-a183-9f8da5876aee-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"3563a545-b965-4c74-a183-9f8da5876aee\") " pod="openstack/mysqld-exporter-0" Dec 06 06:15:50 crc kubenswrapper[4809]: I1206 06:15:50.624265 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lptf5\" (UniqueName: \"kubernetes.io/projected/3563a545-b965-4c74-a183-9f8da5876aee-kube-api-access-lptf5\") pod \"mysqld-exporter-0\" (UID: \"3563a545-b965-4c74-a183-9f8da5876aee\") " pod="openstack/mysqld-exporter-0" Dec 06 06:15:50 crc kubenswrapper[4809]: I1206 06:15:50.723193 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-wgmjm" Dec 06 06:15:50 crc kubenswrapper[4809]: I1206 06:15:50.808305 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 06 06:15:50 crc kubenswrapper[4809]: I1206 06:15:50.912097 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5f4f2323-f99a-4d85-8e0f-a0eb968de27b-etc-swift\") pod \"5f4f2323-f99a-4d85-8e0f-a0eb968de27b\" (UID: \"5f4f2323-f99a-4d85-8e0f-a0eb968de27b\") " Dec 06 06:15:50 crc kubenswrapper[4809]: I1206 06:15:50.912202 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f4f2323-f99a-4d85-8e0f-a0eb968de27b-combined-ca-bundle\") pod \"5f4f2323-f99a-4d85-8e0f-a0eb968de27b\" (UID: \"5f4f2323-f99a-4d85-8e0f-a0eb968de27b\") " Dec 06 06:15:50 crc kubenswrapper[4809]: I1206 06:15:50.912232 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5f4f2323-f99a-4d85-8e0f-a0eb968de27b-scripts\") pod \"5f4f2323-f99a-4d85-8e0f-a0eb968de27b\" (UID: \"5f4f2323-f99a-4d85-8e0f-a0eb968de27b\") " Dec 06 06:15:50 crc kubenswrapper[4809]: I1206 06:15:50.912274 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5f4f2323-f99a-4d85-8e0f-a0eb968de27b-dispersionconf\") pod \"5f4f2323-f99a-4d85-8e0f-a0eb968de27b\" (UID: \"5f4f2323-f99a-4d85-8e0f-a0eb968de27b\") " Dec 06 06:15:50 crc kubenswrapper[4809]: I1206 06:15:50.912348 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jr2rt\" (UniqueName: \"kubernetes.io/projected/5f4f2323-f99a-4d85-8e0f-a0eb968de27b-kube-api-access-jr2rt\") pod \"5f4f2323-f99a-4d85-8e0f-a0eb968de27b\" (UID: \"5f4f2323-f99a-4d85-8e0f-a0eb968de27b\") " Dec 06 06:15:50 crc kubenswrapper[4809]: I1206 06:15:50.912385 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5f4f2323-f99a-4d85-8e0f-a0eb968de27b-ring-data-devices\") pod \"5f4f2323-f99a-4d85-8e0f-a0eb968de27b\" (UID: \"5f4f2323-f99a-4d85-8e0f-a0eb968de27b\") " Dec 06 06:15:50 crc kubenswrapper[4809]: I1206 06:15:50.912415 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5f4f2323-f99a-4d85-8e0f-a0eb968de27b-swiftconf\") pod \"5f4f2323-f99a-4d85-8e0f-a0eb968de27b\" (UID: \"5f4f2323-f99a-4d85-8e0f-a0eb968de27b\") " Dec 06 06:15:50 crc kubenswrapper[4809]: I1206 06:15:50.915002 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f4f2323-f99a-4d85-8e0f-a0eb968de27b-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "5f4f2323-f99a-4d85-8e0f-a0eb968de27b" (UID: "5f4f2323-f99a-4d85-8e0f-a0eb968de27b"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:15:50 crc kubenswrapper[4809]: I1206 06:15:50.917096 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f4f2323-f99a-4d85-8e0f-a0eb968de27b-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "5f4f2323-f99a-4d85-8e0f-a0eb968de27b" (UID: "5f4f2323-f99a-4d85-8e0f-a0eb968de27b"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:15:50 crc kubenswrapper[4809]: I1206 06:15:50.922488 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f4f2323-f99a-4d85-8e0f-a0eb968de27b-kube-api-access-jr2rt" (OuterVolumeSpecName: "kube-api-access-jr2rt") pod "5f4f2323-f99a-4d85-8e0f-a0eb968de27b" (UID: "5f4f2323-f99a-4d85-8e0f-a0eb968de27b"). InnerVolumeSpecName "kube-api-access-jr2rt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:15:50 crc kubenswrapper[4809]: I1206 06:15:50.925376 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f4f2323-f99a-4d85-8e0f-a0eb968de27b-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "5f4f2323-f99a-4d85-8e0f-a0eb968de27b" (UID: "5f4f2323-f99a-4d85-8e0f-a0eb968de27b"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:15:50 crc kubenswrapper[4809]: I1206 06:15:50.945518 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f4f2323-f99a-4d85-8e0f-a0eb968de27b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5f4f2323-f99a-4d85-8e0f-a0eb968de27b" (UID: "5f4f2323-f99a-4d85-8e0f-a0eb968de27b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:15:50 crc kubenswrapper[4809]: I1206 06:15:50.956091 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f4f2323-f99a-4d85-8e0f-a0eb968de27b-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "5f4f2323-f99a-4d85-8e0f-a0eb968de27b" (UID: "5f4f2323-f99a-4d85-8e0f-a0eb968de27b"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:15:50 crc kubenswrapper[4809]: I1206 06:15:50.975553 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f4f2323-f99a-4d85-8e0f-a0eb968de27b-scripts" (OuterVolumeSpecName: "scripts") pod "5f4f2323-f99a-4d85-8e0f-a0eb968de27b" (UID: "5f4f2323-f99a-4d85-8e0f-a0eb968de27b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:15:50 crc kubenswrapper[4809]: I1206 06:15:50.991432 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 06 06:15:51 crc kubenswrapper[4809]: I1206 06:15:51.016365 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f4f2323-f99a-4d85-8e0f-a0eb968de27b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:51 crc kubenswrapper[4809]: I1206 06:15:51.016400 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5f4f2323-f99a-4d85-8e0f-a0eb968de27b-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:51 crc kubenswrapper[4809]: I1206 06:15:51.016411 4809 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5f4f2323-f99a-4d85-8e0f-a0eb968de27b-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:51 crc kubenswrapper[4809]: I1206 06:15:51.016425 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jr2rt\" (UniqueName: \"kubernetes.io/projected/5f4f2323-f99a-4d85-8e0f-a0eb968de27b-kube-api-access-jr2rt\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:51 crc kubenswrapper[4809]: I1206 06:15:51.016439 4809 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5f4f2323-f99a-4d85-8e0f-a0eb968de27b-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:51 crc kubenswrapper[4809]: I1206 06:15:51.016449 4809 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5f4f2323-f99a-4d85-8e0f-a0eb968de27b-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:51 crc kubenswrapper[4809]: I1206 06:15:51.016459 4809 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5f4f2323-f99a-4d85-8e0f-a0eb968de27b-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:51 crc kubenswrapper[4809]: I1206 06:15:51.058543 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-wgmjm" event={"ID":"5f4f2323-f99a-4d85-8e0f-a0eb968de27b","Type":"ContainerDied","Data":"629f7d646a99a3a7368147301a35395ab73534330c4109dbfa174e596452420c"} Dec 06 06:15:51 crc kubenswrapper[4809]: I1206 06:15:51.058592 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="629f7d646a99a3a7368147301a35395ab73534330c4109dbfa174e596452420c" Dec 06 06:15:51 crc kubenswrapper[4809]: I1206 06:15:51.058659 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-wgmjm" Dec 06 06:15:51 crc kubenswrapper[4809]: I1206 06:15:51.061556 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-xhlzm-config-jnsd4" event={"ID":"7ebbbe4b-4be4-4141-9be9-b73ac12741cb","Type":"ContainerStarted","Data":"f353b221a72c907f42a2fc4d8be666369112112697a23d895465668987d9ac5b"} Dec 06 06:15:51 crc kubenswrapper[4809]: I1206 06:15:51.061602 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-xhlzm-config-jnsd4" event={"ID":"7ebbbe4b-4be4-4141-9be9-b73ac12741cb","Type":"ContainerStarted","Data":"f03f76dfcb404327cc8204bca44ad667e95ca5c7a42e1ab3492fad7494f81711"} Dec 06 06:15:51 crc kubenswrapper[4809]: I1206 06:15:51.131361 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-xhlzm-config-jnsd4" podStartSLOduration=2.131329285 podStartE2EDuration="2.131329285s" podCreationTimestamp="2025-12-06 06:15:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:15:51.09112023 +0000 UTC m=+1475.980103192" watchObservedRunningTime="2025-12-06 06:15:51.131329285 +0000 UTC m=+1476.020312257" Dec 06 06:15:51 crc kubenswrapper[4809]: I1206 06:15:51.295613 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 06 06:15:51 crc kubenswrapper[4809]: W1206 06:15:51.300710 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3563a545_b965_4c74_a183_9f8da5876aee.slice/crio-3a302852a8edd1b50ba0f27dab81fb73d1cf4525ada49dec4f7accf4058a1d9b WatchSource:0}: Error finding container 3a302852a8edd1b50ba0f27dab81fb73d1cf4525ada49dec4f7accf4058a1d9b: Status 404 returned error can't find the container with id 3a302852a8edd1b50ba0f27dab81fb73d1cf4525ada49dec4f7accf4058a1d9b Dec 06 06:15:52 crc kubenswrapper[4809]: I1206 06:15:52.090047 4809 generic.go:334] "Generic (PLEG): container finished" podID="7ebbbe4b-4be4-4141-9be9-b73ac12741cb" containerID="f353b221a72c907f42a2fc4d8be666369112112697a23d895465668987d9ac5b" exitCode=0 Dec 06 06:15:52 crc kubenswrapper[4809]: I1206 06:15:52.090268 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-xhlzm-config-jnsd4" event={"ID":"7ebbbe4b-4be4-4141-9be9-b73ac12741cb","Type":"ContainerDied","Data":"f353b221a72c907f42a2fc4d8be666369112112697a23d895465668987d9ac5b"} Dec 06 06:15:52 crc kubenswrapper[4809]: I1206 06:15:52.093753 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"3563a545-b965-4c74-a183-9f8da5876aee","Type":"ContainerStarted","Data":"3a302852a8edd1b50ba0f27dab81fb73d1cf4525ada49dec4f7accf4058a1d9b"} Dec 06 06:15:52 crc kubenswrapper[4809]: I1206 06:15:52.769336 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cfd13e5d-ca28-4061-8ca5-2c4566332bf0-etc-swift\") pod \"swift-storage-0\" (UID: \"cfd13e5d-ca28-4061-8ca5-2c4566332bf0\") " pod="openstack/swift-storage-0" Dec 06 06:15:52 crc kubenswrapper[4809]: I1206 06:15:52.777232 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cfd13e5d-ca28-4061-8ca5-2c4566332bf0-etc-swift\") pod \"swift-storage-0\" (UID: \"cfd13e5d-ca28-4061-8ca5-2c4566332bf0\") " pod="openstack/swift-storage-0" Dec 06 06:15:52 crc kubenswrapper[4809]: I1206 06:15:52.953793 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.119029 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-xhlzm-config-jnsd4" event={"ID":"7ebbbe4b-4be4-4141-9be9-b73ac12741cb","Type":"ContainerDied","Data":"f03f76dfcb404327cc8204bca44ad667e95ca5c7a42e1ab3492fad7494f81711"} Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.119468 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f03f76dfcb404327cc8204bca44ad667e95ca5c7a42e1ab3492fad7494f81711" Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.129230 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xhlzm-config-jnsd4" Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.173171 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.206974 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7ebbbe4b-4be4-4141-9be9-b73ac12741cb-scripts\") pod \"7ebbbe4b-4be4-4141-9be9-b73ac12741cb\" (UID: \"7ebbbe4b-4be4-4141-9be9-b73ac12741cb\") " Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.207090 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7ebbbe4b-4be4-4141-9be9-b73ac12741cb-var-run\") pod \"7ebbbe4b-4be4-4141-9be9-b73ac12741cb\" (UID: \"7ebbbe4b-4be4-4141-9be9-b73ac12741cb\") " Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.207127 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qjkgp\" (UniqueName: \"kubernetes.io/projected/7ebbbe4b-4be4-4141-9be9-b73ac12741cb-kube-api-access-qjkgp\") pod \"7ebbbe4b-4be4-4141-9be9-b73ac12741cb\" (UID: \"7ebbbe4b-4be4-4141-9be9-b73ac12741cb\") " Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.207339 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7ebbbe4b-4be4-4141-9be9-b73ac12741cb-var-run-ovn\") pod \"7ebbbe4b-4be4-4141-9be9-b73ac12741cb\" (UID: \"7ebbbe4b-4be4-4141-9be9-b73ac12741cb\") " Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.207374 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7ebbbe4b-4be4-4141-9be9-b73ac12741cb-additional-scripts\") pod \"7ebbbe4b-4be4-4141-9be9-b73ac12741cb\" (UID: \"7ebbbe4b-4be4-4141-9be9-b73ac12741cb\") " Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.207406 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7ebbbe4b-4be4-4141-9be9-b73ac12741cb-var-log-ovn\") pod \"7ebbbe4b-4be4-4141-9be9-b73ac12741cb\" (UID: \"7ebbbe4b-4be4-4141-9be9-b73ac12741cb\") " Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.207867 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7ebbbe4b-4be4-4141-9be9-b73ac12741cb-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "7ebbbe4b-4be4-4141-9be9-b73ac12741cb" (UID: "7ebbbe4b-4be4-4141-9be9-b73ac12741cb"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.207903 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7ebbbe4b-4be4-4141-9be9-b73ac12741cb-var-run" (OuterVolumeSpecName: "var-run") pod "7ebbbe4b-4be4-4141-9be9-b73ac12741cb" (UID: "7ebbbe4b-4be4-4141-9be9-b73ac12741cb"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.208190 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ebbbe4b-4be4-4141-9be9-b73ac12741cb-scripts" (OuterVolumeSpecName: "scripts") pod "7ebbbe4b-4be4-4141-9be9-b73ac12741cb" (UID: "7ebbbe4b-4be4-4141-9be9-b73ac12741cb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.208225 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7ebbbe4b-4be4-4141-9be9-b73ac12741cb-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "7ebbbe4b-4be4-4141-9be9-b73ac12741cb" (UID: "7ebbbe4b-4be4-4141-9be9-b73ac12741cb"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.209794 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ebbbe4b-4be4-4141-9be9-b73ac12741cb-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "7ebbbe4b-4be4-4141-9be9-b73ac12741cb" (UID: "7ebbbe4b-4be4-4141-9be9-b73ac12741cb"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.234307 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ebbbe4b-4be4-4141-9be9-b73ac12741cb-kube-api-access-qjkgp" (OuterVolumeSpecName: "kube-api-access-qjkgp") pod "7ebbbe4b-4be4-4141-9be9-b73ac12741cb" (UID: "7ebbbe4b-4be4-4141-9be9-b73ac12741cb"). InnerVolumeSpecName "kube-api-access-qjkgp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.310086 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7ebbbe4b-4be4-4141-9be9-b73ac12741cb-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.310119 4809 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7ebbbe4b-4be4-4141-9be9-b73ac12741cb-var-run\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.310129 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qjkgp\" (UniqueName: \"kubernetes.io/projected/7ebbbe4b-4be4-4141-9be9-b73ac12741cb-kube-api-access-qjkgp\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.310139 4809 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7ebbbe4b-4be4-4141-9be9-b73ac12741cb-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.310148 4809 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7ebbbe4b-4be4-4141-9be9-b73ac12741cb-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.310157 4809 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7ebbbe4b-4be4-4141-9be9-b73ac12741cb-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.551718 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-8f52w"] Dec 06 06:15:54 crc kubenswrapper[4809]: E1206 06:15:54.552176 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f4f2323-f99a-4d85-8e0f-a0eb968de27b" containerName="swift-ring-rebalance" Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.552193 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f4f2323-f99a-4d85-8e0f-a0eb968de27b" containerName="swift-ring-rebalance" Dec 06 06:15:54 crc kubenswrapper[4809]: E1206 06:15:54.552211 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ebbbe4b-4be4-4141-9be9-b73ac12741cb" containerName="ovn-config" Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.552218 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ebbbe4b-4be4-4141-9be9-b73ac12741cb" containerName="ovn-config" Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.552445 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ebbbe4b-4be4-4141-9be9-b73ac12741cb" containerName="ovn-config" Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.552473 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f4f2323-f99a-4d85-8e0f-a0eb968de27b" containerName="swift-ring-rebalance" Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.553186 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-8f52w" Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.571659 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-8f52w"] Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.618646 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac97c219-a010-4070-84fe-ae41b3083627-operator-scripts\") pod \"heat-db-create-8f52w\" (UID: \"ac97c219-a010-4070-84fe-ae41b3083627\") " pod="openstack/heat-db-create-8f52w" Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.618737 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghq45\" (UniqueName: \"kubernetes.io/projected/ac97c219-a010-4070-84fe-ae41b3083627-kube-api-access-ghq45\") pod \"heat-db-create-8f52w\" (UID: \"ac97c219-a010-4070-84fe-ae41b3083627\") " pod="openstack/heat-db-create-8f52w" Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.635609 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.658864 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-nr9js"] Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.660883 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-nr9js" Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.686610 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-nr9js"] Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.765737 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghq45\" (UniqueName: \"kubernetes.io/projected/ac97c219-a010-4070-84fe-ae41b3083627-kube-api-access-ghq45\") pod \"heat-db-create-8f52w\" (UID: \"ac97c219-a010-4070-84fe-ae41b3083627\") " pod="openstack/heat-db-create-8f52w" Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.766361 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/72b421c4-764d-464e-9e4b-5cbf6cdbd5d5-operator-scripts\") pod \"cinder-db-create-nr9js\" (UID: \"72b421c4-764d-464e-9e4b-5cbf6cdbd5d5\") " pod="openstack/cinder-db-create-nr9js" Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.766853 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxrsd\" (UniqueName: \"kubernetes.io/projected/72b421c4-764d-464e-9e4b-5cbf6cdbd5d5-kube-api-access-sxrsd\") pod \"cinder-db-create-nr9js\" (UID: \"72b421c4-764d-464e-9e4b-5cbf6cdbd5d5\") " pod="openstack/cinder-db-create-nr9js" Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.766974 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac97c219-a010-4070-84fe-ae41b3083627-operator-scripts\") pod \"heat-db-create-8f52w\" (UID: \"ac97c219-a010-4070-84fe-ae41b3083627\") " pod="openstack/heat-db-create-8f52w" Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.767985 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac97c219-a010-4070-84fe-ae41b3083627-operator-scripts\") pod \"heat-db-create-8f52w\" (UID: \"ac97c219-a010-4070-84fe-ae41b3083627\") " pod="openstack/heat-db-create-8f52w" Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.829723 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghq45\" (UniqueName: \"kubernetes.io/projected/ac97c219-a010-4070-84fe-ae41b3083627-kube-api-access-ghq45\") pod \"heat-db-create-8f52w\" (UID: \"ac97c219-a010-4070-84fe-ae41b3083627\") " pod="openstack/heat-db-create-8f52w" Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.868761 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/72b421c4-764d-464e-9e4b-5cbf6cdbd5d5-operator-scripts\") pod \"cinder-db-create-nr9js\" (UID: \"72b421c4-764d-464e-9e4b-5cbf6cdbd5d5\") " pod="openstack/cinder-db-create-nr9js" Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.868901 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxrsd\" (UniqueName: \"kubernetes.io/projected/72b421c4-764d-464e-9e4b-5cbf6cdbd5d5-kube-api-access-sxrsd\") pod \"cinder-db-create-nr9js\" (UID: \"72b421c4-764d-464e-9e4b-5cbf6cdbd5d5\") " pod="openstack/cinder-db-create-nr9js" Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.869701 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/72b421c4-764d-464e-9e4b-5cbf6cdbd5d5-operator-scripts\") pod \"cinder-db-create-nr9js\" (UID: \"72b421c4-764d-464e-9e4b-5cbf6cdbd5d5\") " pod="openstack/cinder-db-create-nr9js" Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.877638 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-8f52w" Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.886967 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-8nl86"] Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.888741 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-8nl86" Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.906438 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxrsd\" (UniqueName: \"kubernetes.io/projected/72b421c4-764d-464e-9e4b-5cbf6cdbd5d5-kube-api-access-sxrsd\") pod \"cinder-db-create-nr9js\" (UID: \"72b421c4-764d-464e-9e4b-5cbf6cdbd5d5\") " pod="openstack/cinder-db-create-nr9js" Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.944380 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-8nl86"] Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.975893 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngwm7\" (UniqueName: \"kubernetes.io/projected/9eb5001b-9536-430d-8fae-cea0ecabd5ad-kube-api-access-ngwm7\") pod \"barbican-db-create-8nl86\" (UID: \"9eb5001b-9536-430d-8fae-cea0ecabd5ad\") " pod="openstack/barbican-db-create-8nl86" Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.976007 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9eb5001b-9536-430d-8fae-cea0ecabd5ad-operator-scripts\") pod \"barbican-db-create-8nl86\" (UID: \"9eb5001b-9536-430d-8fae-cea0ecabd5ad\") " pod="openstack/barbican-db-create-8nl86" Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.983582 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-d911-account-create-update-wsk86"] Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.984569 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-nr9js" Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.985262 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-d911-account-create-update-wsk86" Dec 06 06:15:54 crc kubenswrapper[4809]: I1206 06:15:54.987980 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.010750 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-d911-account-create-update-wsk86"] Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.038421 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-bh98g"] Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.040787 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-bh98g" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.047140 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.047177 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-2h2rg" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.047366 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.047459 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.078649 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngwm7\" (UniqueName: \"kubernetes.io/projected/9eb5001b-9536-430d-8fae-cea0ecabd5ad-kube-api-access-ngwm7\") pod \"barbican-db-create-8nl86\" (UID: \"9eb5001b-9536-430d-8fae-cea0ecabd5ad\") " pod="openstack/barbican-db-create-8nl86" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.078692 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51864703-13bc-484f-abb0-3a59bc5195ed-operator-scripts\") pod \"cinder-d911-account-create-update-wsk86\" (UID: \"51864703-13bc-484f-abb0-3a59bc5195ed\") " pod="openstack/cinder-d911-account-create-update-wsk86" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.078751 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75l2t\" (UniqueName: \"kubernetes.io/projected/51864703-13bc-484f-abb0-3a59bc5195ed-kube-api-access-75l2t\") pod \"cinder-d911-account-create-update-wsk86\" (UID: \"51864703-13bc-484f-abb0-3a59bc5195ed\") " pod="openstack/cinder-d911-account-create-update-wsk86" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.078781 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9eb5001b-9536-430d-8fae-cea0ecabd5ad-operator-scripts\") pod \"barbican-db-create-8nl86\" (UID: \"9eb5001b-9536-430d-8fae-cea0ecabd5ad\") " pod="openstack/barbican-db-create-8nl86" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.079826 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9eb5001b-9536-430d-8fae-cea0ecabd5ad-operator-scripts\") pod \"barbican-db-create-8nl86\" (UID: \"9eb5001b-9536-430d-8fae-cea0ecabd5ad\") " pod="openstack/barbican-db-create-8nl86" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.081030 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-bh98g"] Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.114596 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngwm7\" (UniqueName: \"kubernetes.io/projected/9eb5001b-9536-430d-8fae-cea0ecabd5ad-kube-api-access-ngwm7\") pod \"barbican-db-create-8nl86\" (UID: \"9eb5001b-9536-430d-8fae-cea0ecabd5ad\") " pod="openstack/barbican-db-create-8nl86" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.119007 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-fbc7-account-create-update-95k9l"] Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.120459 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-fbc7-account-create-update-95k9l" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.127511 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.132196 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xhlzm-config-jnsd4" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.146468 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-fbc7-account-create-update-95k9l"] Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.180672 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llnjw\" (UniqueName: \"kubernetes.io/projected/b0c71ada-5bd2-467e-8d93-cbda68cafcb2-kube-api-access-llnjw\") pod \"keystone-db-sync-bh98g\" (UID: \"b0c71ada-5bd2-467e-8d93-cbda68cafcb2\") " pod="openstack/keystone-db-sync-bh98g" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.180736 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0c71ada-5bd2-467e-8d93-cbda68cafcb2-combined-ca-bundle\") pod \"keystone-db-sync-bh98g\" (UID: \"b0c71ada-5bd2-467e-8d93-cbda68cafcb2\") " pod="openstack/keystone-db-sync-bh98g" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.180763 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmgn2\" (UniqueName: \"kubernetes.io/projected/5749399f-d0d9-40ba-ae30-448d7311a953-kube-api-access-hmgn2\") pod \"heat-fbc7-account-create-update-95k9l\" (UID: \"5749399f-d0d9-40ba-ae30-448d7311a953\") " pod="openstack/heat-fbc7-account-create-update-95k9l" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.180826 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51864703-13bc-484f-abb0-3a59bc5195ed-operator-scripts\") pod \"cinder-d911-account-create-update-wsk86\" (UID: \"51864703-13bc-484f-abb0-3a59bc5195ed\") " pod="openstack/cinder-d911-account-create-update-wsk86" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.180913 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0c71ada-5bd2-467e-8d93-cbda68cafcb2-config-data\") pod \"keystone-db-sync-bh98g\" (UID: \"b0c71ada-5bd2-467e-8d93-cbda68cafcb2\") " pod="openstack/keystone-db-sync-bh98g" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.181114 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5749399f-d0d9-40ba-ae30-448d7311a953-operator-scripts\") pod \"heat-fbc7-account-create-update-95k9l\" (UID: \"5749399f-d0d9-40ba-ae30-448d7311a953\") " pod="openstack/heat-fbc7-account-create-update-95k9l" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.181150 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75l2t\" (UniqueName: \"kubernetes.io/projected/51864703-13bc-484f-abb0-3a59bc5195ed-kube-api-access-75l2t\") pod \"cinder-d911-account-create-update-wsk86\" (UID: \"51864703-13bc-484f-abb0-3a59bc5195ed\") " pod="openstack/cinder-d911-account-create-update-wsk86" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.186565 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51864703-13bc-484f-abb0-3a59bc5195ed-operator-scripts\") pod \"cinder-d911-account-create-update-wsk86\" (UID: \"51864703-13bc-484f-abb0-3a59bc5195ed\") " pod="openstack/cinder-d911-account-create-update-wsk86" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.192499 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-7ec9-account-create-update-llsx9"] Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.193957 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-7ec9-account-create-update-llsx9" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.195975 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.227998 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-7ec9-account-create-update-llsx9"] Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.228524 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75l2t\" (UniqueName: \"kubernetes.io/projected/51864703-13bc-484f-abb0-3a59bc5195ed-kube-api-access-75l2t\") pod \"cinder-d911-account-create-update-wsk86\" (UID: \"51864703-13bc-484f-abb0-3a59bc5195ed\") " pod="openstack/cinder-d911-account-create-update-wsk86" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.283328 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0c71ada-5bd2-467e-8d93-cbda68cafcb2-config-data\") pod \"keystone-db-sync-bh98g\" (UID: \"b0c71ada-5bd2-467e-8d93-cbda68cafcb2\") " pod="openstack/keystone-db-sync-bh98g" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.283461 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5749399f-d0d9-40ba-ae30-448d7311a953-operator-scripts\") pod \"heat-fbc7-account-create-update-95k9l\" (UID: \"5749399f-d0d9-40ba-ae30-448d7311a953\") " pod="openstack/heat-fbc7-account-create-update-95k9l" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.283571 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32aaf34e-00f0-4de8-9a07-ead18a18c8a9-operator-scripts\") pod \"barbican-7ec9-account-create-update-llsx9\" (UID: \"32aaf34e-00f0-4de8-9a07-ead18a18c8a9\") " pod="openstack/barbican-7ec9-account-create-update-llsx9" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.283652 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llnjw\" (UniqueName: \"kubernetes.io/projected/b0c71ada-5bd2-467e-8d93-cbda68cafcb2-kube-api-access-llnjw\") pod \"keystone-db-sync-bh98g\" (UID: \"b0c71ada-5bd2-467e-8d93-cbda68cafcb2\") " pod="openstack/keystone-db-sync-bh98g" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.283705 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0c71ada-5bd2-467e-8d93-cbda68cafcb2-combined-ca-bundle\") pod \"keystone-db-sync-bh98g\" (UID: \"b0c71ada-5bd2-467e-8d93-cbda68cafcb2\") " pod="openstack/keystone-db-sync-bh98g" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.283732 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nt7g\" (UniqueName: \"kubernetes.io/projected/32aaf34e-00f0-4de8-9a07-ead18a18c8a9-kube-api-access-9nt7g\") pod \"barbican-7ec9-account-create-update-llsx9\" (UID: \"32aaf34e-00f0-4de8-9a07-ead18a18c8a9\") " pod="openstack/barbican-7ec9-account-create-update-llsx9" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.283766 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmgn2\" (UniqueName: \"kubernetes.io/projected/5749399f-d0d9-40ba-ae30-448d7311a953-kube-api-access-hmgn2\") pod \"heat-fbc7-account-create-update-95k9l\" (UID: \"5749399f-d0d9-40ba-ae30-448d7311a953\") " pod="openstack/heat-fbc7-account-create-update-95k9l" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.285432 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5749399f-d0d9-40ba-ae30-448d7311a953-operator-scripts\") pod \"heat-fbc7-account-create-update-95k9l\" (UID: \"5749399f-d0d9-40ba-ae30-448d7311a953\") " pod="openstack/heat-fbc7-account-create-update-95k9l" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.292768 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-8nl86" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.298735 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0c71ada-5bd2-467e-8d93-cbda68cafcb2-combined-ca-bundle\") pod \"keystone-db-sync-bh98g\" (UID: \"b0c71ada-5bd2-467e-8d93-cbda68cafcb2\") " pod="openstack/keystone-db-sync-bh98g" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.308390 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmgn2\" (UniqueName: \"kubernetes.io/projected/5749399f-d0d9-40ba-ae30-448d7311a953-kube-api-access-hmgn2\") pod \"heat-fbc7-account-create-update-95k9l\" (UID: \"5749399f-d0d9-40ba-ae30-448d7311a953\") " pod="openstack/heat-fbc7-account-create-update-95k9l" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.308477 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-pc5wh"] Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.310168 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-pc5wh" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.315880 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llnjw\" (UniqueName: \"kubernetes.io/projected/b0c71ada-5bd2-467e-8d93-cbda68cafcb2-kube-api-access-llnjw\") pod \"keystone-db-sync-bh98g\" (UID: \"b0c71ada-5bd2-467e-8d93-cbda68cafcb2\") " pod="openstack/keystone-db-sync-bh98g" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.315965 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-pc5wh"] Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.316851 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0c71ada-5bd2-467e-8d93-cbda68cafcb2-config-data\") pod \"keystone-db-sync-bh98g\" (UID: \"b0c71ada-5bd2-467e-8d93-cbda68cafcb2\") " pod="openstack/keystone-db-sync-bh98g" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.321094 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-d911-account-create-update-wsk86" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.330152 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-xhlzm-config-jnsd4"] Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.346350 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-xhlzm-config-jnsd4"] Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.385447 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9138f866-bbfa-45f4-b3b8-a55b783e1c9c-operator-scripts\") pod \"neutron-db-create-pc5wh\" (UID: \"9138f866-bbfa-45f4-b3b8-a55b783e1c9c\") " pod="openstack/neutron-db-create-pc5wh" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.385923 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32aaf34e-00f0-4de8-9a07-ead18a18c8a9-operator-scripts\") pod \"barbican-7ec9-account-create-update-llsx9\" (UID: \"32aaf34e-00f0-4de8-9a07-ead18a18c8a9\") " pod="openstack/barbican-7ec9-account-create-update-llsx9" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.386018 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fntj2\" (UniqueName: \"kubernetes.io/projected/9138f866-bbfa-45f4-b3b8-a55b783e1c9c-kube-api-access-fntj2\") pod \"neutron-db-create-pc5wh\" (UID: \"9138f866-bbfa-45f4-b3b8-a55b783e1c9c\") " pod="openstack/neutron-db-create-pc5wh" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.386083 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nt7g\" (UniqueName: \"kubernetes.io/projected/32aaf34e-00f0-4de8-9a07-ead18a18c8a9-kube-api-access-9nt7g\") pod \"barbican-7ec9-account-create-update-llsx9\" (UID: \"32aaf34e-00f0-4de8-9a07-ead18a18c8a9\") " pod="openstack/barbican-7ec9-account-create-update-llsx9" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.387417 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32aaf34e-00f0-4de8-9a07-ead18a18c8a9-operator-scripts\") pod \"barbican-7ec9-account-create-update-llsx9\" (UID: \"32aaf34e-00f0-4de8-9a07-ead18a18c8a9\") " pod="openstack/barbican-7ec9-account-create-update-llsx9" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.398954 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-bh98g" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.409132 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nt7g\" (UniqueName: \"kubernetes.io/projected/32aaf34e-00f0-4de8-9a07-ead18a18c8a9-kube-api-access-9nt7g\") pod \"barbican-7ec9-account-create-update-llsx9\" (UID: \"32aaf34e-00f0-4de8-9a07-ead18a18c8a9\") " pod="openstack/barbican-7ec9-account-create-update-llsx9" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.410048 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-7ec9-account-create-update-llsx9" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.413167 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ebbbe4b-4be4-4141-9be9-b73ac12741cb" path="/var/lib/kubelet/pods/7ebbbe4b-4be4-4141-9be9-b73ac12741cb/volumes" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.479855 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-112c-account-create-update-kskg6"] Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.481621 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-112c-account-create-update-kskg6" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.486195 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.487515 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fntj2\" (UniqueName: \"kubernetes.io/projected/9138f866-bbfa-45f4-b3b8-a55b783e1c9c-kube-api-access-fntj2\") pod \"neutron-db-create-pc5wh\" (UID: \"9138f866-bbfa-45f4-b3b8-a55b783e1c9c\") " pod="openstack/neutron-db-create-pc5wh" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.487659 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9138f866-bbfa-45f4-b3b8-a55b783e1c9c-operator-scripts\") pod \"neutron-db-create-pc5wh\" (UID: \"9138f866-bbfa-45f4-b3b8-a55b783e1c9c\") " pod="openstack/neutron-db-create-pc5wh" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.489308 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9138f866-bbfa-45f4-b3b8-a55b783e1c9c-operator-scripts\") pod \"neutron-db-create-pc5wh\" (UID: \"9138f866-bbfa-45f4-b3b8-a55b783e1c9c\") " pod="openstack/neutron-db-create-pc5wh" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.508601 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-fbc7-account-create-update-95k9l" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.524794 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-112c-account-create-update-kskg6"] Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.538951 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fntj2\" (UniqueName: \"kubernetes.io/projected/9138f866-bbfa-45f4-b3b8-a55b783e1c9c-kube-api-access-fntj2\") pod \"neutron-db-create-pc5wh\" (UID: \"9138f866-bbfa-45f4-b3b8-a55b783e1c9c\") " pod="openstack/neutron-db-create-pc5wh" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.589721 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6b837741-4518-43ae-a8a4-befdbcb593b6-operator-scripts\") pod \"neutron-112c-account-create-update-kskg6\" (UID: \"6b837741-4518-43ae-a8a4-befdbcb593b6\") " pod="openstack/neutron-112c-account-create-update-kskg6" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.597122 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxx74\" (UniqueName: \"kubernetes.io/projected/6b837741-4518-43ae-a8a4-befdbcb593b6-kube-api-access-zxx74\") pod \"neutron-112c-account-create-update-kskg6\" (UID: \"6b837741-4518-43ae-a8a4-befdbcb593b6\") " pod="openstack/neutron-112c-account-create-update-kskg6" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.700837 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxx74\" (UniqueName: \"kubernetes.io/projected/6b837741-4518-43ae-a8a4-befdbcb593b6-kube-api-access-zxx74\") pod \"neutron-112c-account-create-update-kskg6\" (UID: \"6b837741-4518-43ae-a8a4-befdbcb593b6\") " pod="openstack/neutron-112c-account-create-update-kskg6" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.700915 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6b837741-4518-43ae-a8a4-befdbcb593b6-operator-scripts\") pod \"neutron-112c-account-create-update-kskg6\" (UID: \"6b837741-4518-43ae-a8a4-befdbcb593b6\") " pod="openstack/neutron-112c-account-create-update-kskg6" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.701763 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6b837741-4518-43ae-a8a4-befdbcb593b6-operator-scripts\") pod \"neutron-112c-account-create-update-kskg6\" (UID: \"6b837741-4518-43ae-a8a4-befdbcb593b6\") " pod="openstack/neutron-112c-account-create-update-kskg6" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.724817 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-pc5wh" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.742408 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxx74\" (UniqueName: \"kubernetes.io/projected/6b837741-4518-43ae-a8a4-befdbcb593b6-kube-api-access-zxx74\") pod \"neutron-112c-account-create-update-kskg6\" (UID: \"6b837741-4518-43ae-a8a4-befdbcb593b6\") " pod="openstack/neutron-112c-account-create-update-kskg6" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.810560 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-112c-account-create-update-kskg6" Dec 06 06:15:55 crc kubenswrapper[4809]: I1206 06:15:55.991676 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 06 06:15:56 crc kubenswrapper[4809]: I1206 06:15:55.999598 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 06 06:15:56 crc kubenswrapper[4809]: I1206 06:15:56.062714 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-8f52w"] Dec 06 06:15:56 crc kubenswrapper[4809]: I1206 06:15:56.144828 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-8f52w" event={"ID":"ac97c219-a010-4070-84fe-ae41b3083627","Type":"ContainerStarted","Data":"42100265eb4c03cd4b71e3c069ee4f57b34c3753ef55f4da743f5a20bb50ade3"} Dec 06 06:15:56 crc kubenswrapper[4809]: I1206 06:15:56.146905 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 06 06:15:56 crc kubenswrapper[4809]: I1206 06:15:56.262980 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-nr9js"] Dec 06 06:15:56 crc kubenswrapper[4809]: I1206 06:15:56.620998 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 06 06:15:56 crc kubenswrapper[4809]: I1206 06:15:56.660716 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-d911-account-create-update-wsk86"] Dec 06 06:15:56 crc kubenswrapper[4809]: I1206 06:15:56.687260 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-bh98g"] Dec 06 06:15:56 crc kubenswrapper[4809]: I1206 06:15:56.772236 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-8nl86"] Dec 06 06:15:56 crc kubenswrapper[4809]: I1206 06:15:56.903447 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-fbc7-account-create-update-95k9l"] Dec 06 06:15:56 crc kubenswrapper[4809]: I1206 06:15:56.919998 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-pc5wh"] Dec 06 06:15:56 crc kubenswrapper[4809]: I1206 06:15:56.931852 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-7ec9-account-create-update-llsx9"] Dec 06 06:15:57 crc kubenswrapper[4809]: I1206 06:15:57.042567 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-112c-account-create-update-kskg6"] Dec 06 06:15:57 crc kubenswrapper[4809]: W1206 06:15:57.098044 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6b837741_4518_43ae_a8a4_befdbcb593b6.slice/crio-f7a03479986b4b2148bdca0861a5f601461030e68f5e97e5716ae7e02ab69648 WatchSource:0}: Error finding container f7a03479986b4b2148bdca0861a5f601461030e68f5e97e5716ae7e02ab69648: Status 404 returned error can't find the container with id f7a03479986b4b2148bdca0861a5f601461030e68f5e97e5716ae7e02ab69648 Dec 06 06:15:57 crc kubenswrapper[4809]: I1206 06:15:57.194059 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-fbc7-account-create-update-95k9l" event={"ID":"5749399f-d0d9-40ba-ae30-448d7311a953","Type":"ContainerStarted","Data":"f1953ef660ae4593051d0bbf14b8c068e14603609c556195b3b733f79a6134a5"} Dec 06 06:15:57 crc kubenswrapper[4809]: I1206 06:15:57.198866 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-8nl86" event={"ID":"9eb5001b-9536-430d-8fae-cea0ecabd5ad","Type":"ContainerStarted","Data":"a64d68ef4d1fe8cf67b5eb8ebaee95ec7def14806a620f37cbcf5ed7a03301ce"} Dec 06 06:15:57 crc kubenswrapper[4809]: I1206 06:15:57.204290 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-7ec9-account-create-update-llsx9" event={"ID":"32aaf34e-00f0-4de8-9a07-ead18a18c8a9","Type":"ContainerStarted","Data":"78c191a48f3dea312b25e50acf52d595fbcabb9aaa8e481293af69d2b68d5abd"} Dec 06 06:15:57 crc kubenswrapper[4809]: I1206 06:15:57.205293 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-bh98g" event={"ID":"b0c71ada-5bd2-467e-8d93-cbda68cafcb2","Type":"ContainerStarted","Data":"4225bdf2ba24105a7bb38ed868e40ffd53979a45758912e45e18320d4eea950a"} Dec 06 06:15:57 crc kubenswrapper[4809]: I1206 06:15:57.209652 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-d911-account-create-update-wsk86" event={"ID":"51864703-13bc-484f-abb0-3a59bc5195ed","Type":"ContainerStarted","Data":"23a9d1eeb04c59f4bd584b496f77e2cbe9a8544a19aaadb2bd3169fbbaec3dc7"} Dec 06 06:15:57 crc kubenswrapper[4809]: I1206 06:15:57.216179 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cfd13e5d-ca28-4061-8ca5-2c4566332bf0","Type":"ContainerStarted","Data":"5af0314e9a61a208d115b58c31bd46a31079fa297be9dc9e6ce7aaf8c414c257"} Dec 06 06:15:57 crc kubenswrapper[4809]: I1206 06:15:57.217574 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-pc5wh" event={"ID":"9138f866-bbfa-45f4-b3b8-a55b783e1c9c","Type":"ContainerStarted","Data":"fd0b19b9f3b28d463d4df50138e748ac95da69581d1ac6fb37dacb6e629f6437"} Dec 06 06:15:57 crc kubenswrapper[4809]: I1206 06:15:57.231826 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-8f52w" event={"ID":"ac97c219-a010-4070-84fe-ae41b3083627","Type":"ContainerStarted","Data":"5518c89e0684238e8a60dbd595ff31fe19ce292ee3631f1c85a374d4bf4bec0a"} Dec 06 06:15:57 crc kubenswrapper[4809]: I1206 06:15:57.246907 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"3563a545-b965-4c74-a183-9f8da5876aee","Type":"ContainerStarted","Data":"2416c7e9cd8e39fe724afee7991341b25c4e5b7738b872ea62c8a977a8721b0c"} Dec 06 06:15:57 crc kubenswrapper[4809]: I1206 06:15:57.252138 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-112c-account-create-update-kskg6" event={"ID":"6b837741-4518-43ae-a8a4-befdbcb593b6","Type":"ContainerStarted","Data":"f7a03479986b4b2148bdca0861a5f601461030e68f5e97e5716ae7e02ab69648"} Dec 06 06:15:57 crc kubenswrapper[4809]: I1206 06:15:57.255740 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-nr9js" event={"ID":"72b421c4-764d-464e-9e4b-5cbf6cdbd5d5","Type":"ContainerStarted","Data":"51ead11faa1c75494f637cd9efed4980c8d186a829c27baffcd145a90d6e1c3c"} Dec 06 06:15:59 crc kubenswrapper[4809]: I1206 06:15:59.320273 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-pc5wh" event={"ID":"9138f866-bbfa-45f4-b3b8-a55b783e1c9c","Type":"ContainerStarted","Data":"c558b3048d490cfd96438f3b8b709314761686af8dce1c3fd2f186fe2ee23899"} Dec 06 06:15:59 crc kubenswrapper[4809]: I1206 06:15:59.341465 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-d911-account-create-update-wsk86" event={"ID":"51864703-13bc-484f-abb0-3a59bc5195ed","Type":"ContainerStarted","Data":"955742661ebbcc1286b13ee089c7a6dad4bb1892566b14c31a44cbd5b6bb2a04"} Dec 06 06:15:59 crc kubenswrapper[4809]: I1206 06:15:59.351623 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-8nl86" event={"ID":"9eb5001b-9536-430d-8fae-cea0ecabd5ad","Type":"ContainerStarted","Data":"e714667d5d6c52e4a0da5d11fdf1b19a24567dded8567c6053f2f5a1cf78cdc2"} Dec 06 06:15:59 crc kubenswrapper[4809]: I1206 06:15:59.352983 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-fbc7-account-create-update-95k9l" event={"ID":"5749399f-d0d9-40ba-ae30-448d7311a953","Type":"ContainerStarted","Data":"5d6a3110494f958e2436dec41c34e634cee07ccb9796e45f2a9580c2a9604055"} Dec 06 06:15:59 crc kubenswrapper[4809]: I1206 06:15:59.354548 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-112c-account-create-update-kskg6" event={"ID":"6b837741-4518-43ae-a8a4-befdbcb593b6","Type":"ContainerStarted","Data":"4079a9916fc3a595c2f6cb41013ebe78e1024ab9a5322ccdeaa48049a5d43714"} Dec 06 06:15:59 crc kubenswrapper[4809]: I1206 06:15:59.362835 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-7ec9-account-create-update-llsx9" event={"ID":"32aaf34e-00f0-4de8-9a07-ead18a18c8a9","Type":"ContainerStarted","Data":"54466e843436c80dbb0c46cb440bc4569458c697cf5eb5b3e0d61a937450d64e"} Dec 06 06:15:59 crc kubenswrapper[4809]: I1206 06:15:59.377040 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-nr9js" event={"ID":"72b421c4-764d-464e-9e4b-5cbf6cdbd5d5","Type":"ContainerStarted","Data":"e586d2395ccc1ce18aea94ffe162c27ddadeed3797ac587c97054312fd868f8d"} Dec 06 06:15:59 crc kubenswrapper[4809]: I1206 06:15:59.416458 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-7ec9-account-create-update-llsx9" podStartSLOduration=4.416434013 podStartE2EDuration="4.416434013s" podCreationTimestamp="2025-12-06 06:15:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:15:59.389747813 +0000 UTC m=+1484.278730755" watchObservedRunningTime="2025-12-06 06:15:59.416434013 +0000 UTC m=+1484.305416945" Dec 06 06:15:59 crc kubenswrapper[4809]: I1206 06:15:59.419759 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-nr9js" podStartSLOduration=5.419743903 podStartE2EDuration="5.419743903s" podCreationTimestamp="2025-12-06 06:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:15:59.41038818 +0000 UTC m=+1484.299371122" watchObservedRunningTime="2025-12-06 06:15:59.419743903 +0000 UTC m=+1484.308726835" Dec 06 06:15:59 crc kubenswrapper[4809]: I1206 06:15:59.436618 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-create-8f52w" podStartSLOduration=5.436599067 podStartE2EDuration="5.436599067s" podCreationTimestamp="2025-12-06 06:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:15:59.429196808 +0000 UTC m=+1484.318179750" watchObservedRunningTime="2025-12-06 06:15:59.436599067 +0000 UTC m=+1484.325582009" Dec 06 06:15:59 crc kubenswrapper[4809]: I1206 06:15:59.460611 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-0" podStartSLOduration=5.239145028 podStartE2EDuration="9.460588704s" podCreationTimestamp="2025-12-06 06:15:50 +0000 UTC" firstStartedPulling="2025-12-06 06:15:51.303955749 +0000 UTC m=+1476.192938691" lastFinishedPulling="2025-12-06 06:15:55.525399425 +0000 UTC m=+1480.414382367" observedRunningTime="2025-12-06 06:15:59.451844688 +0000 UTC m=+1484.340827630" watchObservedRunningTime="2025-12-06 06:15:59.460588704 +0000 UTC m=+1484.349571646" Dec 06 06:16:00 crc kubenswrapper[4809]: I1206 06:16:00.410408 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-fbc7-account-create-update-95k9l" podStartSLOduration=6.410392624 podStartE2EDuration="6.410392624s" podCreationTimestamp="2025-12-06 06:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:16:00.409580702 +0000 UTC m=+1485.298563724" watchObservedRunningTime="2025-12-06 06:16:00.410392624 +0000 UTC m=+1485.299375566" Dec 06 06:16:00 crc kubenswrapper[4809]: I1206 06:16:00.448952 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-112c-account-create-update-kskg6" podStartSLOduration=5.448910773 podStartE2EDuration="5.448910773s" podCreationTimestamp="2025-12-06 06:15:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:16:00.427435614 +0000 UTC m=+1485.316418576" watchObservedRunningTime="2025-12-06 06:16:00.448910773 +0000 UTC m=+1485.337893715" Dec 06 06:16:00 crc kubenswrapper[4809]: I1206 06:16:00.467683 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 06 06:16:00 crc kubenswrapper[4809]: I1206 06:16:00.468009 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="f6536b25-1cc1-4f74-bf46-613965ddf185" containerName="prometheus" containerID="cri-o://b586d911296037f57743670c3c8e87969fa4c7f025ce88b226bef7b8edc0a9ee" gracePeriod=600 Dec 06 06:16:00 crc kubenswrapper[4809]: I1206 06:16:00.468188 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="f6536b25-1cc1-4f74-bf46-613965ddf185" containerName="thanos-sidecar" containerID="cri-o://4bed7b24aeea6ed8999cf873386e191f0e016021ab10c642384864d9c8b535c6" gracePeriod=600 Dec 06 06:16:00 crc kubenswrapper[4809]: I1206 06:16:00.468247 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="f6536b25-1cc1-4f74-bf46-613965ddf185" containerName="config-reloader" containerID="cri-o://c229460d835203ea61a85721814803837875570bf40da7b5ebd574d3453b2ebf" gracePeriod=600 Dec 06 06:16:00 crc kubenswrapper[4809]: I1206 06:16:00.472383 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-d911-account-create-update-wsk86" podStartSLOduration=6.472360875 podStartE2EDuration="6.472360875s" podCreationTimestamp="2025-12-06 06:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:16:00.453319031 +0000 UTC m=+1485.342301973" watchObservedRunningTime="2025-12-06 06:16:00.472360875 +0000 UTC m=+1485.361343817" Dec 06 06:16:00 crc kubenswrapper[4809]: I1206 06:16:00.486154 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-pc5wh" podStartSLOduration=5.486133186 podStartE2EDuration="5.486133186s" podCreationTimestamp="2025-12-06 06:15:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:16:00.467251597 +0000 UTC m=+1485.356234539" watchObservedRunningTime="2025-12-06 06:16:00.486133186 +0000 UTC m=+1485.375116128" Dec 06 06:16:00 crc kubenswrapper[4809]: I1206 06:16:00.507730 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-8nl86" podStartSLOduration=6.507711738 podStartE2EDuration="6.507711738s" podCreationTimestamp="2025-12-06 06:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:16:00.491380898 +0000 UTC m=+1485.380363850" watchObservedRunningTime="2025-12-06 06:16:00.507711738 +0000 UTC m=+1485.396694680" Dec 06 06:16:00 crc kubenswrapper[4809]: E1206 06:16:00.906056 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf6536b25_1cc1_4f74_bf46_613965ddf185.slice/crio-c229460d835203ea61a85721814803837875570bf40da7b5ebd574d3453b2ebf.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podac97c219_a010_4070_84fe_ae41b3083627.slice/crio-5518c89e0684238e8a60dbd595ff31fe19ce292ee3631f1c85a374d4bf4bec0a.scope\": RecentStats: unable to find data in memory cache]" Dec 06 06:16:00 crc kubenswrapper[4809]: I1206 06:16:00.991921 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="f6536b25-1cc1-4f74-bf46-613965ddf185" containerName="prometheus" probeResult="failure" output="Get \"http://10.217.0.134:9090/-/ready\": dial tcp 10.217.0.134:9090: connect: connection refused" Dec 06 06:16:01 crc kubenswrapper[4809]: I1206 06:16:01.414238 4809 generic.go:334] "Generic (PLEG): container finished" podID="ac97c219-a010-4070-84fe-ae41b3083627" containerID="5518c89e0684238e8a60dbd595ff31fe19ce292ee3631f1c85a374d4bf4bec0a" exitCode=0 Dec 06 06:16:01 crc kubenswrapper[4809]: I1206 06:16:01.416159 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-8f52w" event={"ID":"ac97c219-a010-4070-84fe-ae41b3083627","Type":"ContainerDied","Data":"5518c89e0684238e8a60dbd595ff31fe19ce292ee3631f1c85a374d4bf4bec0a"} Dec 06 06:16:01 crc kubenswrapper[4809]: I1206 06:16:01.440896 4809 generic.go:334] "Generic (PLEG): container finished" podID="f6536b25-1cc1-4f74-bf46-613965ddf185" containerID="4bed7b24aeea6ed8999cf873386e191f0e016021ab10c642384864d9c8b535c6" exitCode=0 Dec 06 06:16:01 crc kubenswrapper[4809]: I1206 06:16:01.440941 4809 generic.go:334] "Generic (PLEG): container finished" podID="f6536b25-1cc1-4f74-bf46-613965ddf185" containerID="c229460d835203ea61a85721814803837875570bf40da7b5ebd574d3453b2ebf" exitCode=0 Dec 06 06:16:01 crc kubenswrapper[4809]: I1206 06:16:01.440950 4809 generic.go:334] "Generic (PLEG): container finished" podID="f6536b25-1cc1-4f74-bf46-613965ddf185" containerID="b586d911296037f57743670c3c8e87969fa4c7f025ce88b226bef7b8edc0a9ee" exitCode=0 Dec 06 06:16:01 crc kubenswrapper[4809]: I1206 06:16:01.440973 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f6536b25-1cc1-4f74-bf46-613965ddf185","Type":"ContainerDied","Data":"4bed7b24aeea6ed8999cf873386e191f0e016021ab10c642384864d9c8b535c6"} Dec 06 06:16:01 crc kubenswrapper[4809]: I1206 06:16:01.441120 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f6536b25-1cc1-4f74-bf46-613965ddf185","Type":"ContainerDied","Data":"c229460d835203ea61a85721814803837875570bf40da7b5ebd574d3453b2ebf"} Dec 06 06:16:01 crc kubenswrapper[4809]: I1206 06:16:01.441131 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f6536b25-1cc1-4f74-bf46-613965ddf185","Type":"ContainerDied","Data":"b586d911296037f57743670c3c8e87969fa4c7f025ce88b226bef7b8edc0a9ee"} Dec 06 06:16:02 crc kubenswrapper[4809]: I1206 06:16:02.465169 4809 generic.go:334] "Generic (PLEG): container finished" podID="51864703-13bc-484f-abb0-3a59bc5195ed" containerID="955742661ebbcc1286b13ee089c7a6dad4bb1892566b14c31a44cbd5b6bb2a04" exitCode=0 Dec 06 06:16:02 crc kubenswrapper[4809]: I1206 06:16:02.465246 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-d911-account-create-update-wsk86" event={"ID":"51864703-13bc-484f-abb0-3a59bc5195ed","Type":"ContainerDied","Data":"955742661ebbcc1286b13ee089c7a6dad4bb1892566b14c31a44cbd5b6bb2a04"} Dec 06 06:16:02 crc kubenswrapper[4809]: I1206 06:16:02.467860 4809 generic.go:334] "Generic (PLEG): container finished" podID="5749399f-d0d9-40ba-ae30-448d7311a953" containerID="5d6a3110494f958e2436dec41c34e634cee07ccb9796e45f2a9580c2a9604055" exitCode=0 Dec 06 06:16:02 crc kubenswrapper[4809]: I1206 06:16:02.467892 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-fbc7-account-create-update-95k9l" event={"ID":"5749399f-d0d9-40ba-ae30-448d7311a953","Type":"ContainerDied","Data":"5d6a3110494f958e2436dec41c34e634cee07ccb9796e45f2a9580c2a9604055"} Dec 06 06:16:02 crc kubenswrapper[4809]: I1206 06:16:02.470210 4809 generic.go:334] "Generic (PLEG): container finished" podID="9eb5001b-9536-430d-8fae-cea0ecabd5ad" containerID="e714667d5d6c52e4a0da5d11fdf1b19a24567dded8567c6053f2f5a1cf78cdc2" exitCode=0 Dec 06 06:16:02 crc kubenswrapper[4809]: I1206 06:16:02.470293 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-8nl86" event={"ID":"9eb5001b-9536-430d-8fae-cea0ecabd5ad","Type":"ContainerDied","Data":"e714667d5d6c52e4a0da5d11fdf1b19a24567dded8567c6053f2f5a1cf78cdc2"} Dec 06 06:16:02 crc kubenswrapper[4809]: I1206 06:16:02.477800 4809 generic.go:334] "Generic (PLEG): container finished" podID="6b837741-4518-43ae-a8a4-befdbcb593b6" containerID="4079a9916fc3a595c2f6cb41013ebe78e1024ab9a5322ccdeaa48049a5d43714" exitCode=0 Dec 06 06:16:02 crc kubenswrapper[4809]: I1206 06:16:02.477866 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-112c-account-create-update-kskg6" event={"ID":"6b837741-4518-43ae-a8a4-befdbcb593b6","Type":"ContainerDied","Data":"4079a9916fc3a595c2f6cb41013ebe78e1024ab9a5322ccdeaa48049a5d43714"} Dec 06 06:16:02 crc kubenswrapper[4809]: I1206 06:16:02.488063 4809 generic.go:334] "Generic (PLEG): container finished" podID="32aaf34e-00f0-4de8-9a07-ead18a18c8a9" containerID="54466e843436c80dbb0c46cb440bc4569458c697cf5eb5b3e0d61a937450d64e" exitCode=0 Dec 06 06:16:02 crc kubenswrapper[4809]: I1206 06:16:02.488136 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-7ec9-account-create-update-llsx9" event={"ID":"32aaf34e-00f0-4de8-9a07-ead18a18c8a9","Type":"ContainerDied","Data":"54466e843436c80dbb0c46cb440bc4569458c697cf5eb5b3e0d61a937450d64e"} Dec 06 06:16:02 crc kubenswrapper[4809]: I1206 06:16:02.491546 4809 generic.go:334] "Generic (PLEG): container finished" podID="72b421c4-764d-464e-9e4b-5cbf6cdbd5d5" containerID="e586d2395ccc1ce18aea94ffe162c27ddadeed3797ac587c97054312fd868f8d" exitCode=0 Dec 06 06:16:02 crc kubenswrapper[4809]: I1206 06:16:02.491639 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-nr9js" event={"ID":"72b421c4-764d-464e-9e4b-5cbf6cdbd5d5","Type":"ContainerDied","Data":"e586d2395ccc1ce18aea94ffe162c27ddadeed3797ac587c97054312fd868f8d"} Dec 06 06:16:02 crc kubenswrapper[4809]: I1206 06:16:02.495117 4809 generic.go:334] "Generic (PLEG): container finished" podID="9138f866-bbfa-45f4-b3b8-a55b783e1c9c" containerID="c558b3048d490cfd96438f3b8b709314761686af8dce1c3fd2f186fe2ee23899" exitCode=0 Dec 06 06:16:02 crc kubenswrapper[4809]: I1206 06:16:02.495155 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-pc5wh" event={"ID":"9138f866-bbfa-45f4-b3b8-a55b783e1c9c","Type":"ContainerDied","Data":"c558b3048d490cfd96438f3b8b709314761686af8dce1c3fd2f186fe2ee23899"} Dec 06 06:16:05 crc kubenswrapper[4809]: I1206 06:16:05.992337 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="f6536b25-1cc1-4f74-bf46-613965ddf185" containerName="prometheus" probeResult="failure" output="Get \"http://10.217.0.134:9090/-/ready\": dial tcp 10.217.0.134:9090: connect: connection refused" Dec 06 06:16:08 crc kubenswrapper[4809]: E1206 06:16:08.798174 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api:current-podified" Dec 06 06:16:08 crc kubenswrapper[4809]: E1206 06:16:08.798706 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-75lxt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-mc9z8_openstack(413a30cf-4b57-44d7-b9bf-3d400bf4b897): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 06:16:08 crc kubenswrapper[4809]: E1206 06:16:08.799991 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-mc9z8" podUID="413a30cf-4b57-44d7-b9bf-3d400bf4b897" Dec 06 06:16:08 crc kubenswrapper[4809]: I1206 06:16:08.842271 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-7ec9-account-create-update-llsx9" Dec 06 06:16:08 crc kubenswrapper[4809]: I1206 06:16:08.850613 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-8f52w" Dec 06 06:16:08 crc kubenswrapper[4809]: I1206 06:16:08.991395 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ghq45\" (UniqueName: \"kubernetes.io/projected/ac97c219-a010-4070-84fe-ae41b3083627-kube-api-access-ghq45\") pod \"ac97c219-a010-4070-84fe-ae41b3083627\" (UID: \"ac97c219-a010-4070-84fe-ae41b3083627\") " Dec 06 06:16:08 crc kubenswrapper[4809]: I1206 06:16:08.991539 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac97c219-a010-4070-84fe-ae41b3083627-operator-scripts\") pod \"ac97c219-a010-4070-84fe-ae41b3083627\" (UID: \"ac97c219-a010-4070-84fe-ae41b3083627\") " Dec 06 06:16:08 crc kubenswrapper[4809]: I1206 06:16:08.991702 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32aaf34e-00f0-4de8-9a07-ead18a18c8a9-operator-scripts\") pod \"32aaf34e-00f0-4de8-9a07-ead18a18c8a9\" (UID: \"32aaf34e-00f0-4de8-9a07-ead18a18c8a9\") " Dec 06 06:16:08 crc kubenswrapper[4809]: I1206 06:16:08.991886 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9nt7g\" (UniqueName: \"kubernetes.io/projected/32aaf34e-00f0-4de8-9a07-ead18a18c8a9-kube-api-access-9nt7g\") pod \"32aaf34e-00f0-4de8-9a07-ead18a18c8a9\" (UID: \"32aaf34e-00f0-4de8-9a07-ead18a18c8a9\") " Dec 06 06:16:08 crc kubenswrapper[4809]: I1206 06:16:08.992452 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32aaf34e-00f0-4de8-9a07-ead18a18c8a9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "32aaf34e-00f0-4de8-9a07-ead18a18c8a9" (UID: "32aaf34e-00f0-4de8-9a07-ead18a18c8a9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:16:08 crc kubenswrapper[4809]: I1206 06:16:08.992729 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32aaf34e-00f0-4de8-9a07-ead18a18c8a9-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:08 crc kubenswrapper[4809]: I1206 06:16:08.993055 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac97c219-a010-4070-84fe-ae41b3083627-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ac97c219-a010-4070-84fe-ae41b3083627" (UID: "ac97c219-a010-4070-84fe-ae41b3083627"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:16:08 crc kubenswrapper[4809]: I1206 06:16:08.998767 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac97c219-a010-4070-84fe-ae41b3083627-kube-api-access-ghq45" (OuterVolumeSpecName: "kube-api-access-ghq45") pod "ac97c219-a010-4070-84fe-ae41b3083627" (UID: "ac97c219-a010-4070-84fe-ae41b3083627"). InnerVolumeSpecName "kube-api-access-ghq45". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:16:08 crc kubenswrapper[4809]: I1206 06:16:08.999011 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32aaf34e-00f0-4de8-9a07-ead18a18c8a9-kube-api-access-9nt7g" (OuterVolumeSpecName: "kube-api-access-9nt7g") pod "32aaf34e-00f0-4de8-9a07-ead18a18c8a9" (UID: "32aaf34e-00f0-4de8-9a07-ead18a18c8a9"). InnerVolumeSpecName "kube-api-access-9nt7g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.095841 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9nt7g\" (UniqueName: \"kubernetes.io/projected/32aaf34e-00f0-4de8-9a07-ead18a18c8a9-kube-api-access-9nt7g\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.095875 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ghq45\" (UniqueName: \"kubernetes.io/projected/ac97c219-a010-4070-84fe-ae41b3083627-kube-api-access-ghq45\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.095886 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac97c219-a010-4070-84fe-ae41b3083627-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.220752 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-nr9js" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.226863 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-112c-account-create-update-kskg6" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.248743 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-fbc7-account-create-update-95k9l" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.271314 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-8nl86" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.299005 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-d911-account-create-update-wsk86" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.301385 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/72b421c4-764d-464e-9e4b-5cbf6cdbd5d5-operator-scripts\") pod \"72b421c4-764d-464e-9e4b-5cbf6cdbd5d5\" (UID: \"72b421c4-764d-464e-9e4b-5cbf6cdbd5d5\") " Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.301564 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxx74\" (UniqueName: \"kubernetes.io/projected/6b837741-4518-43ae-a8a4-befdbcb593b6-kube-api-access-zxx74\") pod \"6b837741-4518-43ae-a8a4-befdbcb593b6\" (UID: \"6b837741-4518-43ae-a8a4-befdbcb593b6\") " Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.301751 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6b837741-4518-43ae-a8a4-befdbcb593b6-operator-scripts\") pod \"6b837741-4518-43ae-a8a4-befdbcb593b6\" (UID: \"6b837741-4518-43ae-a8a4-befdbcb593b6\") " Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.301810 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sxrsd\" (UniqueName: \"kubernetes.io/projected/72b421c4-764d-464e-9e4b-5cbf6cdbd5d5-kube-api-access-sxrsd\") pod \"72b421c4-764d-464e-9e4b-5cbf6cdbd5d5\" (UID: \"72b421c4-764d-464e-9e4b-5cbf6cdbd5d5\") " Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.302266 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/72b421c4-764d-464e-9e4b-5cbf6cdbd5d5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "72b421c4-764d-464e-9e4b-5cbf6cdbd5d5" (UID: "72b421c4-764d-464e-9e4b-5cbf6cdbd5d5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.302501 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/72b421c4-764d-464e-9e4b-5cbf6cdbd5d5-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.302817 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b837741-4518-43ae-a8a4-befdbcb593b6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6b837741-4518-43ae-a8a4-befdbcb593b6" (UID: "6b837741-4518-43ae-a8a4-befdbcb593b6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.309512 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b837741-4518-43ae-a8a4-befdbcb593b6-kube-api-access-zxx74" (OuterVolumeSpecName: "kube-api-access-zxx74") pod "6b837741-4518-43ae-a8a4-befdbcb593b6" (UID: "6b837741-4518-43ae-a8a4-befdbcb593b6"). InnerVolumeSpecName "kube-api-access-zxx74". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.311888 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-pc5wh" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.313346 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72b421c4-764d-464e-9e4b-5cbf6cdbd5d5-kube-api-access-sxrsd" (OuterVolumeSpecName: "kube-api-access-sxrsd") pod "72b421c4-764d-464e-9e4b-5cbf6cdbd5d5" (UID: "72b421c4-764d-464e-9e4b-5cbf6cdbd5d5"). InnerVolumeSpecName "kube-api-access-sxrsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.404278 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngwm7\" (UniqueName: \"kubernetes.io/projected/9eb5001b-9536-430d-8fae-cea0ecabd5ad-kube-api-access-ngwm7\") pod \"9eb5001b-9536-430d-8fae-cea0ecabd5ad\" (UID: \"9eb5001b-9536-430d-8fae-cea0ecabd5ad\") " Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.404387 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51864703-13bc-484f-abb0-3a59bc5195ed-operator-scripts\") pod \"51864703-13bc-484f-abb0-3a59bc5195ed\" (UID: \"51864703-13bc-484f-abb0-3a59bc5195ed\") " Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.404430 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-75l2t\" (UniqueName: \"kubernetes.io/projected/51864703-13bc-484f-abb0-3a59bc5195ed-kube-api-access-75l2t\") pod \"51864703-13bc-484f-abb0-3a59bc5195ed\" (UID: \"51864703-13bc-484f-abb0-3a59bc5195ed\") " Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.404466 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5749399f-d0d9-40ba-ae30-448d7311a953-operator-scripts\") pod \"5749399f-d0d9-40ba-ae30-448d7311a953\" (UID: \"5749399f-d0d9-40ba-ae30-448d7311a953\") " Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.404579 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9138f866-bbfa-45f4-b3b8-a55b783e1c9c-operator-scripts\") pod \"9138f866-bbfa-45f4-b3b8-a55b783e1c9c\" (UID: \"9138f866-bbfa-45f4-b3b8-a55b783e1c9c\") " Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.404681 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fntj2\" (UniqueName: \"kubernetes.io/projected/9138f866-bbfa-45f4-b3b8-a55b783e1c9c-kube-api-access-fntj2\") pod \"9138f866-bbfa-45f4-b3b8-a55b783e1c9c\" (UID: \"9138f866-bbfa-45f4-b3b8-a55b783e1c9c\") " Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.404719 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hmgn2\" (UniqueName: \"kubernetes.io/projected/5749399f-d0d9-40ba-ae30-448d7311a953-kube-api-access-hmgn2\") pod \"5749399f-d0d9-40ba-ae30-448d7311a953\" (UID: \"5749399f-d0d9-40ba-ae30-448d7311a953\") " Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.404803 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9eb5001b-9536-430d-8fae-cea0ecabd5ad-operator-scripts\") pod \"9eb5001b-9536-430d-8fae-cea0ecabd5ad\" (UID: \"9eb5001b-9536-430d-8fae-cea0ecabd5ad\") " Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.406115 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6b837741-4518-43ae-a8a4-befdbcb593b6-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.406143 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sxrsd\" (UniqueName: \"kubernetes.io/projected/72b421c4-764d-464e-9e4b-5cbf6cdbd5d5-kube-api-access-sxrsd\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.406157 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxx74\" (UniqueName: \"kubernetes.io/projected/6b837741-4518-43ae-a8a4-befdbcb593b6-kube-api-access-zxx74\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.406468 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51864703-13bc-484f-abb0-3a59bc5195ed-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "51864703-13bc-484f-abb0-3a59bc5195ed" (UID: "51864703-13bc-484f-abb0-3a59bc5195ed"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.406523 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5749399f-d0d9-40ba-ae30-448d7311a953-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5749399f-d0d9-40ba-ae30-448d7311a953" (UID: "5749399f-d0d9-40ba-ae30-448d7311a953"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.407821 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9eb5001b-9536-430d-8fae-cea0ecabd5ad-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9eb5001b-9536-430d-8fae-cea0ecabd5ad" (UID: "9eb5001b-9536-430d-8fae-cea0ecabd5ad"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.411430 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9138f866-bbfa-45f4-b3b8-a55b783e1c9c-kube-api-access-fntj2" (OuterVolumeSpecName: "kube-api-access-fntj2") pod "9138f866-bbfa-45f4-b3b8-a55b783e1c9c" (UID: "9138f866-bbfa-45f4-b3b8-a55b783e1c9c"). InnerVolumeSpecName "kube-api-access-fntj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.413870 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9138f866-bbfa-45f4-b3b8-a55b783e1c9c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9138f866-bbfa-45f4-b3b8-a55b783e1c9c" (UID: "9138f866-bbfa-45f4-b3b8-a55b783e1c9c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.414450 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9eb5001b-9536-430d-8fae-cea0ecabd5ad-kube-api-access-ngwm7" (OuterVolumeSpecName: "kube-api-access-ngwm7") pod "9eb5001b-9536-430d-8fae-cea0ecabd5ad" (UID: "9eb5001b-9536-430d-8fae-cea0ecabd5ad"). InnerVolumeSpecName "kube-api-access-ngwm7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.417573 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5749399f-d0d9-40ba-ae30-448d7311a953-kube-api-access-hmgn2" (OuterVolumeSpecName: "kube-api-access-hmgn2") pod "5749399f-d0d9-40ba-ae30-448d7311a953" (UID: "5749399f-d0d9-40ba-ae30-448d7311a953"). InnerVolumeSpecName "kube-api-access-hmgn2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.434598 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51864703-13bc-484f-abb0-3a59bc5195ed-kube-api-access-75l2t" (OuterVolumeSpecName: "kube-api-access-75l2t") pod "51864703-13bc-484f-abb0-3a59bc5195ed" (UID: "51864703-13bc-484f-abb0-3a59bc5195ed"). InnerVolumeSpecName "kube-api-access-75l2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.508735 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hmgn2\" (UniqueName: \"kubernetes.io/projected/5749399f-d0d9-40ba-ae30-448d7311a953-kube-api-access-hmgn2\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.509019 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9eb5001b-9536-430d-8fae-cea0ecabd5ad-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.509102 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngwm7\" (UniqueName: \"kubernetes.io/projected/9eb5001b-9536-430d-8fae-cea0ecabd5ad-kube-api-access-ngwm7\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.509194 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51864703-13bc-484f-abb0-3a59bc5195ed-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.509264 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-75l2t\" (UniqueName: \"kubernetes.io/projected/51864703-13bc-484f-abb0-3a59bc5195ed-kube-api-access-75l2t\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.509332 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5749399f-d0d9-40ba-ae30-448d7311a953-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.509403 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9138f866-bbfa-45f4-b3b8-a55b783e1c9c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.509472 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fntj2\" (UniqueName: \"kubernetes.io/projected/9138f866-bbfa-45f4-b3b8-a55b783e1c9c-kube-api-access-fntj2\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.563381 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-d911-account-create-update-wsk86" event={"ID":"51864703-13bc-484f-abb0-3a59bc5195ed","Type":"ContainerDied","Data":"23a9d1eeb04c59f4bd584b496f77e2cbe9a8544a19aaadb2bd3169fbbaec3dc7"} Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.563417 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-d911-account-create-update-wsk86" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.563433 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="23a9d1eeb04c59f4bd584b496f77e2cbe9a8544a19aaadb2bd3169fbbaec3dc7" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.565435 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-8f52w" event={"ID":"ac97c219-a010-4070-84fe-ae41b3083627","Type":"ContainerDied","Data":"42100265eb4c03cd4b71e3c069ee4f57b34c3753ef55f4da743f5a20bb50ade3"} Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.565450 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-8f52w" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.565485 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42100265eb4c03cd4b71e3c069ee4f57b34c3753ef55f4da743f5a20bb50ade3" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.568465 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-fbc7-account-create-update-95k9l" event={"ID":"5749399f-d0d9-40ba-ae30-448d7311a953","Type":"ContainerDied","Data":"f1953ef660ae4593051d0bbf14b8c068e14603609c556195b3b733f79a6134a5"} Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.568491 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f1953ef660ae4593051d0bbf14b8c068e14603609c556195b3b733f79a6134a5" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.568544 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-fbc7-account-create-update-95k9l" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.575390 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-8nl86" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.575854 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-8nl86" event={"ID":"9eb5001b-9536-430d-8fae-cea0ecabd5ad","Type":"ContainerDied","Data":"a64d68ef4d1fe8cf67b5eb8ebaee95ec7def14806a620f37cbcf5ed7a03301ce"} Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.575876 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a64d68ef4d1fe8cf67b5eb8ebaee95ec7def14806a620f37cbcf5ed7a03301ce" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.580817 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-112c-account-create-update-kskg6" event={"ID":"6b837741-4518-43ae-a8a4-befdbcb593b6","Type":"ContainerDied","Data":"f7a03479986b4b2148bdca0861a5f601461030e68f5e97e5716ae7e02ab69648"} Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.580858 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f7a03479986b4b2148bdca0861a5f601461030e68f5e97e5716ae7e02ab69648" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.580964 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-112c-account-create-update-kskg6" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.585646 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-7ec9-account-create-update-llsx9" event={"ID":"32aaf34e-00f0-4de8-9a07-ead18a18c8a9","Type":"ContainerDied","Data":"78c191a48f3dea312b25e50acf52d595fbcabb9aaa8e481293af69d2b68d5abd"} Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.585690 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="78c191a48f3dea312b25e50acf52d595fbcabb9aaa8e481293af69d2b68d5abd" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.585766 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-7ec9-account-create-update-llsx9" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.589950 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-nr9js" event={"ID":"72b421c4-764d-464e-9e4b-5cbf6cdbd5d5","Type":"ContainerDied","Data":"51ead11faa1c75494f637cd9efed4980c8d186a829c27baffcd145a90d6e1c3c"} Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.589987 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="51ead11faa1c75494f637cd9efed4980c8d186a829c27baffcd145a90d6e1c3c" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.589972 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-nr9js" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.593489 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-pc5wh" event={"ID":"9138f866-bbfa-45f4-b3b8-a55b783e1c9c","Type":"ContainerDied","Data":"fd0b19b9f3b28d463d4df50138e748ac95da69581d1ac6fb37dacb6e629f6437"} Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.593545 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd0b19b9f3b28d463d4df50138e748ac95da69581d1ac6fb37dacb6e629f6437" Dec 06 06:16:09 crc kubenswrapper[4809]: I1206 06:16:09.593504 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-pc5wh" Dec 06 06:16:09 crc kubenswrapper[4809]: E1206 06:16:09.595281 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api:current-podified\\\"\"" pod="openstack/glance-db-sync-mc9z8" podUID="413a30cf-4b57-44d7-b9bf-3d400bf4b897" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.265823 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.373513 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f6536b25-1cc1-4f74-bf46-613965ddf185-tls-assets\") pod \"f6536b25-1cc1-4f74-bf46-613965ddf185\" (UID: \"f6536b25-1cc1-4f74-bf46-613965ddf185\") " Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.373650 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/f6536b25-1cc1-4f74-bf46-613965ddf185-prometheus-metric-storage-rulefiles-0\") pod \"f6536b25-1cc1-4f74-bf46-613965ddf185\" (UID: \"f6536b25-1cc1-4f74-bf46-613965ddf185\") " Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.373694 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f6536b25-1cc1-4f74-bf46-613965ddf185-config-out\") pod \"f6536b25-1cc1-4f74-bf46-613965ddf185\" (UID: \"f6536b25-1cc1-4f74-bf46-613965ddf185\") " Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.373741 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/f6536b25-1cc1-4f74-bf46-613965ddf185-thanos-prometheus-http-client-file\") pod \"f6536b25-1cc1-4f74-bf46-613965ddf185\" (UID: \"f6536b25-1cc1-4f74-bf46-613965ddf185\") " Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.373855 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f6536b25-1cc1-4f74-bf46-613965ddf185-web-config\") pod \"f6536b25-1cc1-4f74-bf46-613965ddf185\" (UID: \"f6536b25-1cc1-4f74-bf46-613965ddf185\") " Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.373892 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"f6536b25-1cc1-4f74-bf46-613965ddf185\" (UID: \"f6536b25-1cc1-4f74-bf46-613965ddf185\") " Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.373919 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8mz4q\" (UniqueName: \"kubernetes.io/projected/f6536b25-1cc1-4f74-bf46-613965ddf185-kube-api-access-8mz4q\") pod \"f6536b25-1cc1-4f74-bf46-613965ddf185\" (UID: \"f6536b25-1cc1-4f74-bf46-613965ddf185\") " Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.373976 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f6536b25-1cc1-4f74-bf46-613965ddf185-config\") pod \"f6536b25-1cc1-4f74-bf46-613965ddf185\" (UID: \"f6536b25-1cc1-4f74-bf46-613965ddf185\") " Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.374295 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6536b25-1cc1-4f74-bf46-613965ddf185-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "f6536b25-1cc1-4f74-bf46-613965ddf185" (UID: "f6536b25-1cc1-4f74-bf46-613965ddf185"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.374636 4809 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/f6536b25-1cc1-4f74-bf46-613965ddf185-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.384506 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6536b25-1cc1-4f74-bf46-613965ddf185-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "f6536b25-1cc1-4f74-bf46-613965ddf185" (UID: "f6536b25-1cc1-4f74-bf46-613965ddf185"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.384680 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6536b25-1cc1-4f74-bf46-613965ddf185-kube-api-access-8mz4q" (OuterVolumeSpecName: "kube-api-access-8mz4q") pod "f6536b25-1cc1-4f74-bf46-613965ddf185" (UID: "f6536b25-1cc1-4f74-bf46-613965ddf185"). InnerVolumeSpecName "kube-api-access-8mz4q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.386157 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6536b25-1cc1-4f74-bf46-613965ddf185-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "f6536b25-1cc1-4f74-bf46-613965ddf185" (UID: "f6536b25-1cc1-4f74-bf46-613965ddf185"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.388015 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "f6536b25-1cc1-4f74-bf46-613965ddf185" (UID: "f6536b25-1cc1-4f74-bf46-613965ddf185"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.395134 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6536b25-1cc1-4f74-bf46-613965ddf185-config-out" (OuterVolumeSpecName: "config-out") pod "f6536b25-1cc1-4f74-bf46-613965ddf185" (UID: "f6536b25-1cc1-4f74-bf46-613965ddf185"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.399254 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6536b25-1cc1-4f74-bf46-613965ddf185-config" (OuterVolumeSpecName: "config") pod "f6536b25-1cc1-4f74-bf46-613965ddf185" (UID: "f6536b25-1cc1-4f74-bf46-613965ddf185"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.477221 4809 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/f6536b25-1cc1-4f74-bf46-613965ddf185-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.477273 4809 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.477288 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8mz4q\" (UniqueName: \"kubernetes.io/projected/f6536b25-1cc1-4f74-bf46-613965ddf185-kube-api-access-8mz4q\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.477301 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/f6536b25-1cc1-4f74-bf46-613965ddf185-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.477313 4809 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f6536b25-1cc1-4f74-bf46-613965ddf185-tls-assets\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.477323 4809 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f6536b25-1cc1-4f74-bf46-613965ddf185-config-out\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.488040 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6536b25-1cc1-4f74-bf46-613965ddf185-web-config" (OuterVolumeSpecName: "web-config") pod "f6536b25-1cc1-4f74-bf46-613965ddf185" (UID: "f6536b25-1cc1-4f74-bf46-613965ddf185"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.538616 4809 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.579661 4809 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f6536b25-1cc1-4f74-bf46-613965ddf185-web-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.579749 4809 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.642036 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f6536b25-1cc1-4f74-bf46-613965ddf185","Type":"ContainerDied","Data":"44015bc9fdb90f164a86a172d0c7a083c8cbce1f698d4de2a734ce7aa5f22bce"} Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.642099 4809 scope.go:117] "RemoveContainer" containerID="4bed7b24aeea6ed8999cf873386e191f0e016021ab10c642384864d9c8b535c6" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.642211 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.662848 4809 scope.go:117] "RemoveContainer" containerID="c229460d835203ea61a85721814803837875570bf40da7b5ebd574d3453b2ebf" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.706464 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.718462 4809 scope.go:117] "RemoveContainer" containerID="b586d911296037f57743670c3c8e87969fa4c7f025ce88b226bef7b8edc0a9ee" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.735008 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.750317 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 06 06:16:12 crc kubenswrapper[4809]: E1206 06:16:12.750898 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac97c219-a010-4070-84fe-ae41b3083627" containerName="mariadb-database-create" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.750927 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac97c219-a010-4070-84fe-ae41b3083627" containerName="mariadb-database-create" Dec 06 06:16:12 crc kubenswrapper[4809]: E1206 06:16:12.750973 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9138f866-bbfa-45f4-b3b8-a55b783e1c9c" containerName="mariadb-database-create" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.750985 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="9138f866-bbfa-45f4-b3b8-a55b783e1c9c" containerName="mariadb-database-create" Dec 06 06:16:12 crc kubenswrapper[4809]: E1206 06:16:12.751004 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32aaf34e-00f0-4de8-9a07-ead18a18c8a9" containerName="mariadb-account-create-update" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.751015 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="32aaf34e-00f0-4de8-9a07-ead18a18c8a9" containerName="mariadb-account-create-update" Dec 06 06:16:12 crc kubenswrapper[4809]: E1206 06:16:12.751030 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9eb5001b-9536-430d-8fae-cea0ecabd5ad" containerName="mariadb-database-create" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.751038 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="9eb5001b-9536-430d-8fae-cea0ecabd5ad" containerName="mariadb-database-create" Dec 06 06:16:12 crc kubenswrapper[4809]: E1206 06:16:12.751053 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72b421c4-764d-464e-9e4b-5cbf6cdbd5d5" containerName="mariadb-database-create" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.751060 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="72b421c4-764d-464e-9e4b-5cbf6cdbd5d5" containerName="mariadb-database-create" Dec 06 06:16:12 crc kubenswrapper[4809]: E1206 06:16:12.751081 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6536b25-1cc1-4f74-bf46-613965ddf185" containerName="thanos-sidecar" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.751089 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6536b25-1cc1-4f74-bf46-613965ddf185" containerName="thanos-sidecar" Dec 06 06:16:12 crc kubenswrapper[4809]: E1206 06:16:12.751108 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b837741-4518-43ae-a8a4-befdbcb593b6" containerName="mariadb-account-create-update" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.751116 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b837741-4518-43ae-a8a4-befdbcb593b6" containerName="mariadb-account-create-update" Dec 06 06:16:12 crc kubenswrapper[4809]: E1206 06:16:12.751134 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6536b25-1cc1-4f74-bf46-613965ddf185" containerName="prometheus" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.751142 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6536b25-1cc1-4f74-bf46-613965ddf185" containerName="prometheus" Dec 06 06:16:12 crc kubenswrapper[4809]: E1206 06:16:12.751155 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6536b25-1cc1-4f74-bf46-613965ddf185" containerName="init-config-reloader" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.751163 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6536b25-1cc1-4f74-bf46-613965ddf185" containerName="init-config-reloader" Dec 06 06:16:12 crc kubenswrapper[4809]: E1206 06:16:12.751172 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5749399f-d0d9-40ba-ae30-448d7311a953" containerName="mariadb-account-create-update" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.751179 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="5749399f-d0d9-40ba-ae30-448d7311a953" containerName="mariadb-account-create-update" Dec 06 06:16:12 crc kubenswrapper[4809]: E1206 06:16:12.751191 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51864703-13bc-484f-abb0-3a59bc5195ed" containerName="mariadb-account-create-update" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.751198 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="51864703-13bc-484f-abb0-3a59bc5195ed" containerName="mariadb-account-create-update" Dec 06 06:16:12 crc kubenswrapper[4809]: E1206 06:16:12.751219 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6536b25-1cc1-4f74-bf46-613965ddf185" containerName="config-reloader" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.751228 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6536b25-1cc1-4f74-bf46-613965ddf185" containerName="config-reloader" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.751467 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac97c219-a010-4070-84fe-ae41b3083627" containerName="mariadb-database-create" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.751484 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="72b421c4-764d-464e-9e4b-5cbf6cdbd5d5" containerName="mariadb-database-create" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.751502 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="9eb5001b-9536-430d-8fae-cea0ecabd5ad" containerName="mariadb-database-create" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.751519 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b837741-4518-43ae-a8a4-befdbcb593b6" containerName="mariadb-account-create-update" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.751529 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6536b25-1cc1-4f74-bf46-613965ddf185" containerName="prometheus" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.751547 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="51864703-13bc-484f-abb0-3a59bc5195ed" containerName="mariadb-account-create-update" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.751560 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6536b25-1cc1-4f74-bf46-613965ddf185" containerName="thanos-sidecar" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.751572 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="5749399f-d0d9-40ba-ae30-448d7311a953" containerName="mariadb-account-create-update" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.751586 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="32aaf34e-00f0-4de8-9a07-ead18a18c8a9" containerName="mariadb-account-create-update" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.751601 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6536b25-1cc1-4f74-bf46-613965ddf185" containerName="config-reloader" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.751614 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="9138f866-bbfa-45f4-b3b8-a55b783e1c9c" containerName="mariadb-database-create" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.755351 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.761832 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.763046 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-cx7x2" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.765172 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.765349 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.765505 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.765638 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.768451 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.770911 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.800098 4809 scope.go:117] "RemoveContainer" containerID="df03251430e15c56ceb99e1135791e64f339c2265989799fec5b3d7450799a7d" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.888428 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ca8e11eb-2da1-4706-b589-650bcb817d4c-config\") pod \"prometheus-metric-storage-0\" (UID: \"ca8e11eb-2da1-4706-b589-650bcb817d4c\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.888493 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/ca8e11eb-2da1-4706-b589-650bcb817d4c-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"ca8e11eb-2da1-4706-b589-650bcb817d4c\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.888726 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ca8e11eb-2da1-4706-b589-650bcb817d4c-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"ca8e11eb-2da1-4706-b589-650bcb817d4c\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.888779 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kn8m2\" (UniqueName: \"kubernetes.io/projected/ca8e11eb-2da1-4706-b589-650bcb817d4c-kube-api-access-kn8m2\") pod \"prometheus-metric-storage-0\" (UID: \"ca8e11eb-2da1-4706-b589-650bcb817d4c\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.888810 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/ca8e11eb-2da1-4706-b589-650bcb817d4c-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"ca8e11eb-2da1-4706-b589-650bcb817d4c\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.889129 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"prometheus-metric-storage-0\" (UID: \"ca8e11eb-2da1-4706-b589-650bcb817d4c\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.889158 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/ca8e11eb-2da1-4706-b589-650bcb817d4c-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"ca8e11eb-2da1-4706-b589-650bcb817d4c\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.889392 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ca8e11eb-2da1-4706-b589-650bcb817d4c-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"ca8e11eb-2da1-4706-b589-650bcb817d4c\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.889429 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca8e11eb-2da1-4706-b589-650bcb817d4c-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"ca8e11eb-2da1-4706-b589-650bcb817d4c\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.889529 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ca8e11eb-2da1-4706-b589-650bcb817d4c-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"ca8e11eb-2da1-4706-b589-650bcb817d4c\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.889620 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/ca8e11eb-2da1-4706-b589-650bcb817d4c-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"ca8e11eb-2da1-4706-b589-650bcb817d4c\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.990986 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ca8e11eb-2da1-4706-b589-650bcb817d4c-config\") pod \"prometheus-metric-storage-0\" (UID: \"ca8e11eb-2da1-4706-b589-650bcb817d4c\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.991052 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/ca8e11eb-2da1-4706-b589-650bcb817d4c-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"ca8e11eb-2da1-4706-b589-650bcb817d4c\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.991092 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ca8e11eb-2da1-4706-b589-650bcb817d4c-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"ca8e11eb-2da1-4706-b589-650bcb817d4c\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.991129 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kn8m2\" (UniqueName: \"kubernetes.io/projected/ca8e11eb-2da1-4706-b589-650bcb817d4c-kube-api-access-kn8m2\") pod \"prometheus-metric-storage-0\" (UID: \"ca8e11eb-2da1-4706-b589-650bcb817d4c\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.991161 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/ca8e11eb-2da1-4706-b589-650bcb817d4c-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"ca8e11eb-2da1-4706-b589-650bcb817d4c\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.991225 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"prometheus-metric-storage-0\" (UID: \"ca8e11eb-2da1-4706-b589-650bcb817d4c\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.991248 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/ca8e11eb-2da1-4706-b589-650bcb817d4c-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"ca8e11eb-2da1-4706-b589-650bcb817d4c\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.991284 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ca8e11eb-2da1-4706-b589-650bcb817d4c-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"ca8e11eb-2da1-4706-b589-650bcb817d4c\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.991315 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca8e11eb-2da1-4706-b589-650bcb817d4c-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"ca8e11eb-2da1-4706-b589-650bcb817d4c\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.991352 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ca8e11eb-2da1-4706-b589-650bcb817d4c-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"ca8e11eb-2da1-4706-b589-650bcb817d4c\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.991421 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/ca8e11eb-2da1-4706-b589-650bcb817d4c-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"ca8e11eb-2da1-4706-b589-650bcb817d4c\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.993114 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/ca8e11eb-2da1-4706-b589-650bcb817d4c-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"ca8e11eb-2da1-4706-b589-650bcb817d4c\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.993309 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"prometheus-metric-storage-0\" (UID: \"ca8e11eb-2da1-4706-b589-650bcb817d4c\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/prometheus-metric-storage-0" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.998841 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ca8e11eb-2da1-4706-b589-650bcb817d4c-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"ca8e11eb-2da1-4706-b589-650bcb817d4c\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.999066 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/ca8e11eb-2da1-4706-b589-650bcb817d4c-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"ca8e11eb-2da1-4706-b589-650bcb817d4c\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.999291 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca8e11eb-2da1-4706-b589-650bcb817d4c-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"ca8e11eb-2da1-4706-b589-650bcb817d4c\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.999379 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ca8e11eb-2da1-4706-b589-650bcb817d4c-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"ca8e11eb-2da1-4706-b589-650bcb817d4c\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.999383 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/ca8e11eb-2da1-4706-b589-650bcb817d4c-config\") pod \"prometheus-metric-storage-0\" (UID: \"ca8e11eb-2da1-4706-b589-650bcb817d4c\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:16:12 crc kubenswrapper[4809]: I1206 06:16:12.999449 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/ca8e11eb-2da1-4706-b589-650bcb817d4c-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"ca8e11eb-2da1-4706-b589-650bcb817d4c\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:16:13 crc kubenswrapper[4809]: I1206 06:16:13.000105 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/ca8e11eb-2da1-4706-b589-650bcb817d4c-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"ca8e11eb-2da1-4706-b589-650bcb817d4c\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:16:13 crc kubenswrapper[4809]: I1206 06:16:13.012688 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ca8e11eb-2da1-4706-b589-650bcb817d4c-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"ca8e11eb-2da1-4706-b589-650bcb817d4c\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:16:13 crc kubenswrapper[4809]: I1206 06:16:13.014614 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kn8m2\" (UniqueName: \"kubernetes.io/projected/ca8e11eb-2da1-4706-b589-650bcb817d4c-kube-api-access-kn8m2\") pod \"prometheus-metric-storage-0\" (UID: \"ca8e11eb-2da1-4706-b589-650bcb817d4c\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:16:13 crc kubenswrapper[4809]: I1206 06:16:13.039915 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"prometheus-metric-storage-0\" (UID: \"ca8e11eb-2da1-4706-b589-650bcb817d4c\") " pod="openstack/prometheus-metric-storage-0" Dec 06 06:16:13 crc kubenswrapper[4809]: I1206 06:16:13.088910 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 06 06:16:13 crc kubenswrapper[4809]: I1206 06:16:13.439264 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6536b25-1cc1-4f74-bf46-613965ddf185" path="/var/lib/kubelet/pods/f6536b25-1cc1-4f74-bf46-613965ddf185/volumes" Dec 06 06:16:13 crc kubenswrapper[4809]: I1206 06:16:13.654599 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-bh98g" event={"ID":"b0c71ada-5bd2-467e-8d93-cbda68cafcb2","Type":"ContainerStarted","Data":"5e3d1b548c31544bd9756e5d158976cbae49fb8a786506f50f03d202850b668e"} Dec 06 06:16:13 crc kubenswrapper[4809]: I1206 06:16:13.657032 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cfd13e5d-ca28-4061-8ca5-2c4566332bf0","Type":"ContainerStarted","Data":"a3627e155fe8df5eccda22bea288f8b2df9671cca80411f2d6e4f1e313195389"} Dec 06 06:16:13 crc kubenswrapper[4809]: I1206 06:16:13.657110 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cfd13e5d-ca28-4061-8ca5-2c4566332bf0","Type":"ContainerStarted","Data":"32d039602ffa3f37f2d13094da1dcd0c4d9a668aaa07261c22dc8d1a0e1f7619"} Dec 06 06:16:13 crc kubenswrapper[4809]: I1206 06:16:13.690559 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-bh98g" podStartSLOduration=4.030146105 podStartE2EDuration="19.690531986s" podCreationTimestamp="2025-12-06 06:15:54 +0000 UTC" firstStartedPulling="2025-12-06 06:15:56.769254115 +0000 UTC m=+1481.658237057" lastFinishedPulling="2025-12-06 06:16:12.429639996 +0000 UTC m=+1497.318622938" observedRunningTime="2025-12-06 06:16:13.673190809 +0000 UTC m=+1498.562173771" watchObservedRunningTime="2025-12-06 06:16:13.690531986 +0000 UTC m=+1498.579514948" Dec 06 06:16:13 crc kubenswrapper[4809]: I1206 06:16:13.765155 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 06 06:16:13 crc kubenswrapper[4809]: I1206 06:16:13.992218 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="f6536b25-1cc1-4f74-bf46-613965ddf185" containerName="prometheus" probeResult="failure" output="Get \"http://10.217.0.134:9090/-/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 06:16:14 crc kubenswrapper[4809]: I1206 06:16:14.668558 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ca8e11eb-2da1-4706-b589-650bcb817d4c","Type":"ContainerStarted","Data":"de096570c93a225e2c9f9269752c0ded1fb0054931d605fdee4a1a807aa88d66"} Dec 06 06:16:15 crc kubenswrapper[4809]: I1206 06:16:15.681058 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cfd13e5d-ca28-4061-8ca5-2c4566332bf0","Type":"ContainerStarted","Data":"13ec7769e8fcc356dd81bf974b8062691bf004009350dee5526f4e35c4d74f7f"} Dec 06 06:16:15 crc kubenswrapper[4809]: I1206 06:16:15.681439 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cfd13e5d-ca28-4061-8ca5-2c4566332bf0","Type":"ContainerStarted","Data":"f7ba17a5b3abcbd0cefd8a313cb13d8bb29cd97d6e3634bd9bb133a9063bd004"} Dec 06 06:16:17 crc kubenswrapper[4809]: I1206 06:16:17.723119 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cfd13e5d-ca28-4061-8ca5-2c4566332bf0","Type":"ContainerStarted","Data":"6ac781e9436c905c857cf5d7d304c7a850c456d26b2119dc35cd847446fc631e"} Dec 06 06:16:17 crc kubenswrapper[4809]: I1206 06:16:17.723735 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cfd13e5d-ca28-4061-8ca5-2c4566332bf0","Type":"ContainerStarted","Data":"f1155415b3e3a4e51e753cc111204da9e64260ff720f3bed1d39fbf8df265f86"} Dec 06 06:16:17 crc kubenswrapper[4809]: I1206 06:16:17.723751 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cfd13e5d-ca28-4061-8ca5-2c4566332bf0","Type":"ContainerStarted","Data":"c299b83626b0f37169813a12ef16c93a68647f7cc03b678e9fe1c4d2eb45b2a4"} Dec 06 06:16:17 crc kubenswrapper[4809]: I1206 06:16:17.723787 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cfd13e5d-ca28-4061-8ca5-2c4566332bf0","Type":"ContainerStarted","Data":"057d5f3b2c192b6306ea99fe33c050638d803f75ea841e657a5120a167842dac"} Dec 06 06:16:17 crc kubenswrapper[4809]: I1206 06:16:17.725231 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ca8e11eb-2da1-4706-b589-650bcb817d4c","Type":"ContainerStarted","Data":"209f5febf4d99fd8dcb763ad4fd6fff03a5214b24e55afc99542f4521334df09"} Dec 06 06:16:17 crc kubenswrapper[4809]: I1206 06:16:17.727459 4809 generic.go:334] "Generic (PLEG): container finished" podID="b0c71ada-5bd2-467e-8d93-cbda68cafcb2" containerID="5e3d1b548c31544bd9756e5d158976cbae49fb8a786506f50f03d202850b668e" exitCode=0 Dec 06 06:16:17 crc kubenswrapper[4809]: I1206 06:16:17.727523 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-bh98g" event={"ID":"b0c71ada-5bd2-467e-8d93-cbda68cafcb2","Type":"ContainerDied","Data":"5e3d1b548c31544bd9756e5d158976cbae49fb8a786506f50f03d202850b668e"} Dec 06 06:16:19 crc kubenswrapper[4809]: I1206 06:16:19.124528 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-bh98g" Dec 06 06:16:19 crc kubenswrapper[4809]: I1206 06:16:19.233210 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0c71ada-5bd2-467e-8d93-cbda68cafcb2-config-data\") pod \"b0c71ada-5bd2-467e-8d93-cbda68cafcb2\" (UID: \"b0c71ada-5bd2-467e-8d93-cbda68cafcb2\") " Dec 06 06:16:19 crc kubenswrapper[4809]: I1206 06:16:19.233453 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0c71ada-5bd2-467e-8d93-cbda68cafcb2-combined-ca-bundle\") pod \"b0c71ada-5bd2-467e-8d93-cbda68cafcb2\" (UID: \"b0c71ada-5bd2-467e-8d93-cbda68cafcb2\") " Dec 06 06:16:19 crc kubenswrapper[4809]: I1206 06:16:19.233488 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-llnjw\" (UniqueName: \"kubernetes.io/projected/b0c71ada-5bd2-467e-8d93-cbda68cafcb2-kube-api-access-llnjw\") pod \"b0c71ada-5bd2-467e-8d93-cbda68cafcb2\" (UID: \"b0c71ada-5bd2-467e-8d93-cbda68cafcb2\") " Dec 06 06:16:19 crc kubenswrapper[4809]: I1206 06:16:19.238414 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0c71ada-5bd2-467e-8d93-cbda68cafcb2-kube-api-access-llnjw" (OuterVolumeSpecName: "kube-api-access-llnjw") pod "b0c71ada-5bd2-467e-8d93-cbda68cafcb2" (UID: "b0c71ada-5bd2-467e-8d93-cbda68cafcb2"). InnerVolumeSpecName "kube-api-access-llnjw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:16:19 crc kubenswrapper[4809]: I1206 06:16:19.262024 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0c71ada-5bd2-467e-8d93-cbda68cafcb2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b0c71ada-5bd2-467e-8d93-cbda68cafcb2" (UID: "b0c71ada-5bd2-467e-8d93-cbda68cafcb2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:16:19 crc kubenswrapper[4809]: I1206 06:16:19.289040 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0c71ada-5bd2-467e-8d93-cbda68cafcb2-config-data" (OuterVolumeSpecName: "config-data") pod "b0c71ada-5bd2-467e-8d93-cbda68cafcb2" (UID: "b0c71ada-5bd2-467e-8d93-cbda68cafcb2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:16:19 crc kubenswrapper[4809]: I1206 06:16:19.335679 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0c71ada-5bd2-467e-8d93-cbda68cafcb2-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:19 crc kubenswrapper[4809]: I1206 06:16:19.335723 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0c71ada-5bd2-467e-8d93-cbda68cafcb2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:19 crc kubenswrapper[4809]: I1206 06:16:19.335736 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-llnjw\" (UniqueName: \"kubernetes.io/projected/b0c71ada-5bd2-467e-8d93-cbda68cafcb2-kube-api-access-llnjw\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:19 crc kubenswrapper[4809]: I1206 06:16:19.748385 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-bh98g" event={"ID":"b0c71ada-5bd2-467e-8d93-cbda68cafcb2","Type":"ContainerDied","Data":"4225bdf2ba24105a7bb38ed868e40ffd53979a45758912e45e18320d4eea950a"} Dec 06 06:16:19 crc kubenswrapper[4809]: I1206 06:16:19.748425 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4225bdf2ba24105a7bb38ed868e40ffd53979a45758912e45e18320d4eea950a" Dec 06 06:16:19 crc kubenswrapper[4809]: I1206 06:16:19.748478 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-bh98g" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.027544 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f877ddd87-w7rt2"] Dec 06 06:16:20 crc kubenswrapper[4809]: E1206 06:16:20.028022 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0c71ada-5bd2-467e-8d93-cbda68cafcb2" containerName="keystone-db-sync" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.028040 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0c71ada-5bd2-467e-8d93-cbda68cafcb2" containerName="keystone-db-sync" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.028287 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0c71ada-5bd2-467e-8d93-cbda68cafcb2" containerName="keystone-db-sync" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.029418 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f877ddd87-w7rt2" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.037344 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f877ddd87-w7rt2"] Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.074012 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-q29b5"] Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.075684 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-q29b5" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.084295 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-2h2rg" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.084475 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.084562 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.084914 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.085094 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.176624 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74-config\") pod \"dnsmasq-dns-f877ddd87-w7rt2\" (UID: \"9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74\") " pod="openstack/dnsmasq-dns-f877ddd87-w7rt2" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.179059 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74-ovsdbserver-sb\") pod \"dnsmasq-dns-f877ddd87-w7rt2\" (UID: \"9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74\") " pod="openstack/dnsmasq-dns-f877ddd87-w7rt2" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.179440 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bkkl\" (UniqueName: \"kubernetes.io/projected/9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74-kube-api-access-4bkkl\") pod \"dnsmasq-dns-f877ddd87-w7rt2\" (UID: \"9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74\") " pod="openstack/dnsmasq-dns-f877ddd87-w7rt2" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.179498 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74-dns-svc\") pod \"dnsmasq-dns-f877ddd87-w7rt2\" (UID: \"9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74\") " pod="openstack/dnsmasq-dns-f877ddd87-w7rt2" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.179693 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74-ovsdbserver-nb\") pod \"dnsmasq-dns-f877ddd87-w7rt2\" (UID: \"9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74\") " pod="openstack/dnsmasq-dns-f877ddd87-w7rt2" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.207108 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-q29b5"] Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.260789 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-ckhg9"] Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.264126 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-ckhg9" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.267570 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-km9wl" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.267783 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.282030 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74-ovsdbserver-sb\") pod \"dnsmasq-dns-f877ddd87-w7rt2\" (UID: \"9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74\") " pod="openstack/dnsmasq-dns-f877ddd87-w7rt2" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.282077 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f2a4615-0ec1-404a-9ab9-848d8b45be2a-scripts\") pod \"keystone-bootstrap-q29b5\" (UID: \"4f2a4615-0ec1-404a-9ab9-848d8b45be2a\") " pod="openstack/keystone-bootstrap-q29b5" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.282124 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvplm\" (UniqueName: \"kubernetes.io/projected/4f2a4615-0ec1-404a-9ab9-848d8b45be2a-kube-api-access-tvplm\") pod \"keystone-bootstrap-q29b5\" (UID: \"4f2a4615-0ec1-404a-9ab9-848d8b45be2a\") " pod="openstack/keystone-bootstrap-q29b5" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.282144 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4f2a4615-0ec1-404a-9ab9-848d8b45be2a-fernet-keys\") pod \"keystone-bootstrap-q29b5\" (UID: \"4f2a4615-0ec1-404a-9ab9-848d8b45be2a\") " pod="openstack/keystone-bootstrap-q29b5" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.282201 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edf1ab66-b483-4172-81fd-dd3f3b9b44aa-combined-ca-bundle\") pod \"heat-db-sync-ckhg9\" (UID: \"edf1ab66-b483-4172-81fd-dd3f3b9b44aa\") " pod="openstack/heat-db-sync-ckhg9" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.282231 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4f2a4615-0ec1-404a-9ab9-848d8b45be2a-credential-keys\") pod \"keystone-bootstrap-q29b5\" (UID: \"4f2a4615-0ec1-404a-9ab9-848d8b45be2a\") " pod="openstack/keystone-bootstrap-q29b5" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.282249 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkz67\" (UniqueName: \"kubernetes.io/projected/edf1ab66-b483-4172-81fd-dd3f3b9b44aa-kube-api-access-zkz67\") pod \"heat-db-sync-ckhg9\" (UID: \"edf1ab66-b483-4172-81fd-dd3f3b9b44aa\") " pod="openstack/heat-db-sync-ckhg9" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.282274 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bkkl\" (UniqueName: \"kubernetes.io/projected/9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74-kube-api-access-4bkkl\") pod \"dnsmasq-dns-f877ddd87-w7rt2\" (UID: \"9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74\") " pod="openstack/dnsmasq-dns-f877ddd87-w7rt2" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.282294 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edf1ab66-b483-4172-81fd-dd3f3b9b44aa-config-data\") pod \"heat-db-sync-ckhg9\" (UID: \"edf1ab66-b483-4172-81fd-dd3f3b9b44aa\") " pod="openstack/heat-db-sync-ckhg9" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.282318 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74-dns-svc\") pod \"dnsmasq-dns-f877ddd87-w7rt2\" (UID: \"9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74\") " pod="openstack/dnsmasq-dns-f877ddd87-w7rt2" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.282357 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f2a4615-0ec1-404a-9ab9-848d8b45be2a-combined-ca-bundle\") pod \"keystone-bootstrap-q29b5\" (UID: \"4f2a4615-0ec1-404a-9ab9-848d8b45be2a\") " pod="openstack/keystone-bootstrap-q29b5" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.282385 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74-ovsdbserver-nb\") pod \"dnsmasq-dns-f877ddd87-w7rt2\" (UID: \"9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74\") " pod="openstack/dnsmasq-dns-f877ddd87-w7rt2" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.282410 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f2a4615-0ec1-404a-9ab9-848d8b45be2a-config-data\") pod \"keystone-bootstrap-q29b5\" (UID: \"4f2a4615-0ec1-404a-9ab9-848d8b45be2a\") " pod="openstack/keystone-bootstrap-q29b5" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.282427 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74-config\") pod \"dnsmasq-dns-f877ddd87-w7rt2\" (UID: \"9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74\") " pod="openstack/dnsmasq-dns-f877ddd87-w7rt2" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.283460 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74-config\") pod \"dnsmasq-dns-f877ddd87-w7rt2\" (UID: \"9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74\") " pod="openstack/dnsmasq-dns-f877ddd87-w7rt2" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.283970 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74-ovsdbserver-sb\") pod \"dnsmasq-dns-f877ddd87-w7rt2\" (UID: \"9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74\") " pod="openstack/dnsmasq-dns-f877ddd87-w7rt2" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.284730 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74-ovsdbserver-nb\") pod \"dnsmasq-dns-f877ddd87-w7rt2\" (UID: \"9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74\") " pod="openstack/dnsmasq-dns-f877ddd87-w7rt2" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.286429 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74-dns-svc\") pod \"dnsmasq-dns-f877ddd87-w7rt2\" (UID: \"9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74\") " pod="openstack/dnsmasq-dns-f877ddd87-w7rt2" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.287449 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-ckhg9"] Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.299209 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-xs2c8"] Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.321354 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-xs2c8"] Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.321484 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-xs2c8" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.346567 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.346752 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.349697 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bkkl\" (UniqueName: \"kubernetes.io/projected/9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74-kube-api-access-4bkkl\") pod \"dnsmasq-dns-f877ddd87-w7rt2\" (UID: \"9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74\") " pod="openstack/dnsmasq-dns-f877ddd87-w7rt2" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.350135 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-hwb92" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.351631 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f877ddd87-w7rt2" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.385800 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f2a4615-0ec1-404a-9ab9-848d8b45be2a-combined-ca-bundle\") pod \"keystone-bootstrap-q29b5\" (UID: \"4f2a4615-0ec1-404a-9ab9-848d8b45be2a\") " pod="openstack/keystone-bootstrap-q29b5" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.385866 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f2a4615-0ec1-404a-9ab9-848d8b45be2a-config-data\") pod \"keystone-bootstrap-q29b5\" (UID: \"4f2a4615-0ec1-404a-9ab9-848d8b45be2a\") " pod="openstack/keystone-bootstrap-q29b5" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.385903 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f2a4615-0ec1-404a-9ab9-848d8b45be2a-scripts\") pod \"keystone-bootstrap-q29b5\" (UID: \"4f2a4615-0ec1-404a-9ab9-848d8b45be2a\") " pod="openstack/keystone-bootstrap-q29b5" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.385952 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvplm\" (UniqueName: \"kubernetes.io/projected/4f2a4615-0ec1-404a-9ab9-848d8b45be2a-kube-api-access-tvplm\") pod \"keystone-bootstrap-q29b5\" (UID: \"4f2a4615-0ec1-404a-9ab9-848d8b45be2a\") " pod="openstack/keystone-bootstrap-q29b5" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.385966 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4f2a4615-0ec1-404a-9ab9-848d8b45be2a-fernet-keys\") pod \"keystone-bootstrap-q29b5\" (UID: \"4f2a4615-0ec1-404a-9ab9-848d8b45be2a\") " pod="openstack/keystone-bootstrap-q29b5" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.386022 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edf1ab66-b483-4172-81fd-dd3f3b9b44aa-combined-ca-bundle\") pod \"heat-db-sync-ckhg9\" (UID: \"edf1ab66-b483-4172-81fd-dd3f3b9b44aa\") " pod="openstack/heat-db-sync-ckhg9" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.386063 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4f2a4615-0ec1-404a-9ab9-848d8b45be2a-credential-keys\") pod \"keystone-bootstrap-q29b5\" (UID: \"4f2a4615-0ec1-404a-9ab9-848d8b45be2a\") " pod="openstack/keystone-bootstrap-q29b5" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.386084 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkz67\" (UniqueName: \"kubernetes.io/projected/edf1ab66-b483-4172-81fd-dd3f3b9b44aa-kube-api-access-zkz67\") pod \"heat-db-sync-ckhg9\" (UID: \"edf1ab66-b483-4172-81fd-dd3f3b9b44aa\") " pod="openstack/heat-db-sync-ckhg9" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.386111 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edf1ab66-b483-4172-81fd-dd3f3b9b44aa-config-data\") pod \"heat-db-sync-ckhg9\" (UID: \"edf1ab66-b483-4172-81fd-dd3f3b9b44aa\") " pod="openstack/heat-db-sync-ckhg9" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.406090 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-hsm9x"] Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.407498 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-hsm9x" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.417146 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edf1ab66-b483-4172-81fd-dd3f3b9b44aa-combined-ca-bundle\") pod \"heat-db-sync-ckhg9\" (UID: \"edf1ab66-b483-4172-81fd-dd3f3b9b44aa\") " pod="openstack/heat-db-sync-ckhg9" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.418723 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f2a4615-0ec1-404a-9ab9-848d8b45be2a-scripts\") pod \"keystone-bootstrap-q29b5\" (UID: \"4f2a4615-0ec1-404a-9ab9-848d8b45be2a\") " pod="openstack/keystone-bootstrap-q29b5" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.419234 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edf1ab66-b483-4172-81fd-dd3f3b9b44aa-config-data\") pod \"heat-db-sync-ckhg9\" (UID: \"edf1ab66-b483-4172-81fd-dd3f3b9b44aa\") " pod="openstack/heat-db-sync-ckhg9" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.421735 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.421922 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-t2ttn" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.422124 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.424722 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4f2a4615-0ec1-404a-9ab9-848d8b45be2a-credential-keys\") pod \"keystone-bootstrap-q29b5\" (UID: \"4f2a4615-0ec1-404a-9ab9-848d8b45be2a\") " pod="openstack/keystone-bootstrap-q29b5" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.429594 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkz67\" (UniqueName: \"kubernetes.io/projected/edf1ab66-b483-4172-81fd-dd3f3b9b44aa-kube-api-access-zkz67\") pod \"heat-db-sync-ckhg9\" (UID: \"edf1ab66-b483-4172-81fd-dd3f3b9b44aa\") " pod="openstack/heat-db-sync-ckhg9" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.430540 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f2a4615-0ec1-404a-9ab9-848d8b45be2a-config-data\") pod \"keystone-bootstrap-q29b5\" (UID: \"4f2a4615-0ec1-404a-9ab9-848d8b45be2a\") " pod="openstack/keystone-bootstrap-q29b5" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.430558 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f2a4615-0ec1-404a-9ab9-848d8b45be2a-combined-ca-bundle\") pod \"keystone-bootstrap-q29b5\" (UID: \"4f2a4615-0ec1-404a-9ab9-848d8b45be2a\") " pod="openstack/keystone-bootstrap-q29b5" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.431758 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4f2a4615-0ec1-404a-9ab9-848d8b45be2a-fernet-keys\") pod \"keystone-bootstrap-q29b5\" (UID: \"4f2a4615-0ec1-404a-9ab9-848d8b45be2a\") " pod="openstack/keystone-bootstrap-q29b5" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.456893 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvplm\" (UniqueName: \"kubernetes.io/projected/4f2a4615-0ec1-404a-9ab9-848d8b45be2a-kube-api-access-tvplm\") pod \"keystone-bootstrap-q29b5\" (UID: \"4f2a4615-0ec1-404a-9ab9-848d8b45be2a\") " pod="openstack/keystone-bootstrap-q29b5" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.466003 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-hsm9x"] Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.489169 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c75cf833-5179-4e18-8f74-3faed72757ed-db-sync-config-data\") pod \"cinder-db-sync-xs2c8\" (UID: \"c75cf833-5179-4e18-8f74-3faed72757ed\") " pod="openstack/cinder-db-sync-xs2c8" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.489254 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c75cf833-5179-4e18-8f74-3faed72757ed-scripts\") pod \"cinder-db-sync-xs2c8\" (UID: \"c75cf833-5179-4e18-8f74-3faed72757ed\") " pod="openstack/cinder-db-sync-xs2c8" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.489279 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxm74\" (UniqueName: \"kubernetes.io/projected/c75cf833-5179-4e18-8f74-3faed72757ed-kube-api-access-wxm74\") pod \"cinder-db-sync-xs2c8\" (UID: \"c75cf833-5179-4e18-8f74-3faed72757ed\") " pod="openstack/cinder-db-sync-xs2c8" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.489301 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c75cf833-5179-4e18-8f74-3faed72757ed-config-data\") pod \"cinder-db-sync-xs2c8\" (UID: \"c75cf833-5179-4e18-8f74-3faed72757ed\") " pod="openstack/cinder-db-sync-xs2c8" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.489384 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c75cf833-5179-4e18-8f74-3faed72757ed-combined-ca-bundle\") pod \"cinder-db-sync-xs2c8\" (UID: \"c75cf833-5179-4e18-8f74-3faed72757ed\") " pod="openstack/cinder-db-sync-xs2c8" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.489409 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c75cf833-5179-4e18-8f74-3faed72757ed-etc-machine-id\") pod \"cinder-db-sync-xs2c8\" (UID: \"c75cf833-5179-4e18-8f74-3faed72757ed\") " pod="openstack/cinder-db-sync-xs2c8" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.507640 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-4gxwk"] Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.509454 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-4gxwk" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.521582 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.521823 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-s2gvb" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.560997 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-4gxwk"] Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.585618 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-ckhg9" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.591189 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxm74\" (UniqueName: \"kubernetes.io/projected/c75cf833-5179-4e18-8f74-3faed72757ed-kube-api-access-wxm74\") pod \"cinder-db-sync-xs2c8\" (UID: \"c75cf833-5179-4e18-8f74-3faed72757ed\") " pod="openstack/cinder-db-sync-xs2c8" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.591250 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c75cf833-5179-4e18-8f74-3faed72757ed-config-data\") pod \"cinder-db-sync-xs2c8\" (UID: \"c75cf833-5179-4e18-8f74-3faed72757ed\") " pod="openstack/cinder-db-sync-xs2c8" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.591305 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08195221-42d1-4ce3-93a6-050e44339013-combined-ca-bundle\") pod \"neutron-db-sync-hsm9x\" (UID: \"08195221-42d1-4ce3-93a6-050e44339013\") " pod="openstack/neutron-db-sync-hsm9x" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.591354 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/08195221-42d1-4ce3-93a6-050e44339013-config\") pod \"neutron-db-sync-hsm9x\" (UID: \"08195221-42d1-4ce3-93a6-050e44339013\") " pod="openstack/neutron-db-sync-hsm9x" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.591425 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c75cf833-5179-4e18-8f74-3faed72757ed-combined-ca-bundle\") pod \"cinder-db-sync-xs2c8\" (UID: \"c75cf833-5179-4e18-8f74-3faed72757ed\") " pod="openstack/cinder-db-sync-xs2c8" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.591444 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c75cf833-5179-4e18-8f74-3faed72757ed-etc-machine-id\") pod \"cinder-db-sync-xs2c8\" (UID: \"c75cf833-5179-4e18-8f74-3faed72757ed\") " pod="openstack/cinder-db-sync-xs2c8" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.591563 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c75cf833-5179-4e18-8f74-3faed72757ed-db-sync-config-data\") pod \"cinder-db-sync-xs2c8\" (UID: \"c75cf833-5179-4e18-8f74-3faed72757ed\") " pod="openstack/cinder-db-sync-xs2c8" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.591607 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64xgk\" (UniqueName: \"kubernetes.io/projected/08195221-42d1-4ce3-93a6-050e44339013-kube-api-access-64xgk\") pod \"neutron-db-sync-hsm9x\" (UID: \"08195221-42d1-4ce3-93a6-050e44339013\") " pod="openstack/neutron-db-sync-hsm9x" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.591661 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c75cf833-5179-4e18-8f74-3faed72757ed-scripts\") pod \"cinder-db-sync-xs2c8\" (UID: \"c75cf833-5179-4e18-8f74-3faed72757ed\") " pod="openstack/cinder-db-sync-xs2c8" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.615426 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c75cf833-5179-4e18-8f74-3faed72757ed-etc-machine-id\") pod \"cinder-db-sync-xs2c8\" (UID: \"c75cf833-5179-4e18-8f74-3faed72757ed\") " pod="openstack/cinder-db-sync-xs2c8" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.617823 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-m98hv"] Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.633153 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-m98hv" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.638769 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.640289 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-vjw2g" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.640628 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.641441 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-m98hv"] Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.649869 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c75cf833-5179-4e18-8f74-3faed72757ed-db-sync-config-data\") pod \"cinder-db-sync-xs2c8\" (UID: \"c75cf833-5179-4e18-8f74-3faed72757ed\") " pod="openstack/cinder-db-sync-xs2c8" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.653355 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c75cf833-5179-4e18-8f74-3faed72757ed-combined-ca-bundle\") pod \"cinder-db-sync-xs2c8\" (UID: \"c75cf833-5179-4e18-8f74-3faed72757ed\") " pod="openstack/cinder-db-sync-xs2c8" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.666518 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c75cf833-5179-4e18-8f74-3faed72757ed-config-data\") pod \"cinder-db-sync-xs2c8\" (UID: \"c75cf833-5179-4e18-8f74-3faed72757ed\") " pod="openstack/cinder-db-sync-xs2c8" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.666667 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c75cf833-5179-4e18-8f74-3faed72757ed-scripts\") pod \"cinder-db-sync-xs2c8\" (UID: \"c75cf833-5179-4e18-8f74-3faed72757ed\") " pod="openstack/cinder-db-sync-xs2c8" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.689697 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxm74\" (UniqueName: \"kubernetes.io/projected/c75cf833-5179-4e18-8f74-3faed72757ed-kube-api-access-wxm74\") pod \"cinder-db-sync-xs2c8\" (UID: \"c75cf833-5179-4e18-8f74-3faed72757ed\") " pod="openstack/cinder-db-sync-xs2c8" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.694387 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f877ddd87-w7rt2"] Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.696259 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08195221-42d1-4ce3-93a6-050e44339013-combined-ca-bundle\") pod \"neutron-db-sync-hsm9x\" (UID: \"08195221-42d1-4ce3-93a6-050e44339013\") " pod="openstack/neutron-db-sync-hsm9x" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.696343 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/08195221-42d1-4ce3-93a6-050e44339013-config\") pod \"neutron-db-sync-hsm9x\" (UID: \"08195221-42d1-4ce3-93a6-050e44339013\") " pod="openstack/neutron-db-sync-hsm9x" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.696443 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/662e24e4-5dd6-4d18-822c-2863017f15af-combined-ca-bundle\") pod \"barbican-db-sync-4gxwk\" (UID: \"662e24e4-5dd6-4d18-822c-2863017f15af\") " pod="openstack/barbican-db-sync-4gxwk" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.696552 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpp7j\" (UniqueName: \"kubernetes.io/projected/662e24e4-5dd6-4d18-822c-2863017f15af-kube-api-access-qpp7j\") pod \"barbican-db-sync-4gxwk\" (UID: \"662e24e4-5dd6-4d18-822c-2863017f15af\") " pod="openstack/barbican-db-sync-4gxwk" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.696621 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/662e24e4-5dd6-4d18-822c-2863017f15af-db-sync-config-data\") pod \"barbican-db-sync-4gxwk\" (UID: \"662e24e4-5dd6-4d18-822c-2863017f15af\") " pod="openstack/barbican-db-sync-4gxwk" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.696671 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64xgk\" (UniqueName: \"kubernetes.io/projected/08195221-42d1-4ce3-93a6-050e44339013-kube-api-access-64xgk\") pod \"neutron-db-sync-hsm9x\" (UID: \"08195221-42d1-4ce3-93a6-050e44339013\") " pod="openstack/neutron-db-sync-hsm9x" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.710190 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-q29b5" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.713625 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08195221-42d1-4ce3-93a6-050e44339013-combined-ca-bundle\") pod \"neutron-db-sync-hsm9x\" (UID: \"08195221-42d1-4ce3-93a6-050e44339013\") " pod="openstack/neutron-db-sync-hsm9x" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.715471 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/08195221-42d1-4ce3-93a6-050e44339013-config\") pod \"neutron-db-sync-hsm9x\" (UID: \"08195221-42d1-4ce3-93a6-050e44339013\") " pod="openstack/neutron-db-sync-hsm9x" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.720440 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64xgk\" (UniqueName: \"kubernetes.io/projected/08195221-42d1-4ce3-93a6-050e44339013-kube-api-access-64xgk\") pod \"neutron-db-sync-hsm9x\" (UID: \"08195221-42d1-4ce3-93a6-050e44339013\") " pod="openstack/neutron-db-sync-hsm9x" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.800141 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/662e24e4-5dd6-4d18-822c-2863017f15af-combined-ca-bundle\") pod \"barbican-db-sync-4gxwk\" (UID: \"662e24e4-5dd6-4d18-822c-2863017f15af\") " pod="openstack/barbican-db-sync-4gxwk" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.800206 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8336af68-51f9-4f3a-8c92-61dd1aba9067-logs\") pod \"placement-db-sync-m98hv\" (UID: \"8336af68-51f9-4f3a-8c92-61dd1aba9067\") " pod="openstack/placement-db-sync-m98hv" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.800257 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpp7j\" (UniqueName: \"kubernetes.io/projected/662e24e4-5dd6-4d18-822c-2863017f15af-kube-api-access-qpp7j\") pod \"barbican-db-sync-4gxwk\" (UID: \"662e24e4-5dd6-4d18-822c-2863017f15af\") " pod="openstack/barbican-db-sync-4gxwk" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.800325 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/662e24e4-5dd6-4d18-822c-2863017f15af-db-sync-config-data\") pod \"barbican-db-sync-4gxwk\" (UID: \"662e24e4-5dd6-4d18-822c-2863017f15af\") " pod="openstack/barbican-db-sync-4gxwk" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.800432 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8336af68-51f9-4f3a-8c92-61dd1aba9067-config-data\") pod \"placement-db-sync-m98hv\" (UID: \"8336af68-51f9-4f3a-8c92-61dd1aba9067\") " pod="openstack/placement-db-sync-m98hv" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.800567 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8336af68-51f9-4f3a-8c92-61dd1aba9067-scripts\") pod \"placement-db-sync-m98hv\" (UID: \"8336af68-51f9-4f3a-8c92-61dd1aba9067\") " pod="openstack/placement-db-sync-m98hv" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.800596 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8336af68-51f9-4f3a-8c92-61dd1aba9067-combined-ca-bundle\") pod \"placement-db-sync-m98hv\" (UID: \"8336af68-51f9-4f3a-8c92-61dd1aba9067\") " pod="openstack/placement-db-sync-m98hv" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.800814 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r96sr\" (UniqueName: \"kubernetes.io/projected/8336af68-51f9-4f3a-8c92-61dd1aba9067-kube-api-access-r96sr\") pod \"placement-db-sync-m98hv\" (UID: \"8336af68-51f9-4f3a-8c92-61dd1aba9067\") " pod="openstack/placement-db-sync-m98hv" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.808217 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/662e24e4-5dd6-4d18-822c-2863017f15af-db-sync-config-data\") pod \"barbican-db-sync-4gxwk\" (UID: \"662e24e4-5dd6-4d18-822c-2863017f15af\") " pod="openstack/barbican-db-sync-4gxwk" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.824927 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpp7j\" (UniqueName: \"kubernetes.io/projected/662e24e4-5dd6-4d18-822c-2863017f15af-kube-api-access-qpp7j\") pod \"barbican-db-sync-4gxwk\" (UID: \"662e24e4-5dd6-4d18-822c-2863017f15af\") " pod="openstack/barbican-db-sync-4gxwk" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.833582 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/662e24e4-5dd6-4d18-822c-2863017f15af-combined-ca-bundle\") pod \"barbican-db-sync-4gxwk\" (UID: \"662e24e4-5dd6-4d18-822c-2863017f15af\") " pod="openstack/barbican-db-sync-4gxwk" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.833665 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-68dcc9cf6f-vjzzr"] Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.835898 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68dcc9cf6f-vjzzr" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.865888 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68dcc9cf6f-vjzzr"] Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.902430 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8336af68-51f9-4f3a-8c92-61dd1aba9067-scripts\") pod \"placement-db-sync-m98hv\" (UID: \"8336af68-51f9-4f3a-8c92-61dd1aba9067\") " pod="openstack/placement-db-sync-m98hv" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.902478 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8336af68-51f9-4f3a-8c92-61dd1aba9067-combined-ca-bundle\") pod \"placement-db-sync-m98hv\" (UID: \"8336af68-51f9-4f3a-8c92-61dd1aba9067\") " pod="openstack/placement-db-sync-m98hv" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.902510 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r96sr\" (UniqueName: \"kubernetes.io/projected/8336af68-51f9-4f3a-8c92-61dd1aba9067-kube-api-access-r96sr\") pod \"placement-db-sync-m98hv\" (UID: \"8336af68-51f9-4f3a-8c92-61dd1aba9067\") " pod="openstack/placement-db-sync-m98hv" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.902570 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8336af68-51f9-4f3a-8c92-61dd1aba9067-logs\") pod \"placement-db-sync-m98hv\" (UID: \"8336af68-51f9-4f3a-8c92-61dd1aba9067\") " pod="openstack/placement-db-sync-m98hv" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.902658 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8336af68-51f9-4f3a-8c92-61dd1aba9067-config-data\") pod \"placement-db-sync-m98hv\" (UID: \"8336af68-51f9-4f3a-8c92-61dd1aba9067\") " pod="openstack/placement-db-sync-m98hv" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.907061 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8336af68-51f9-4f3a-8c92-61dd1aba9067-logs\") pod \"placement-db-sync-m98hv\" (UID: \"8336af68-51f9-4f3a-8c92-61dd1aba9067\") " pod="openstack/placement-db-sync-m98hv" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.907129 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.909362 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8336af68-51f9-4f3a-8c92-61dd1aba9067-scripts\") pod \"placement-db-sync-m98hv\" (UID: \"8336af68-51f9-4f3a-8c92-61dd1aba9067\") " pod="openstack/placement-db-sync-m98hv" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.909397 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8336af68-51f9-4f3a-8c92-61dd1aba9067-combined-ca-bundle\") pod \"placement-db-sync-m98hv\" (UID: \"8336af68-51f9-4f3a-8c92-61dd1aba9067\") " pod="openstack/placement-db-sync-m98hv" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.910761 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.913753 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.913975 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.927661 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.933542 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8336af68-51f9-4f3a-8c92-61dd1aba9067-config-data\") pod \"placement-db-sync-m98hv\" (UID: \"8336af68-51f9-4f3a-8c92-61dd1aba9067\") " pod="openstack/placement-db-sync-m98hv" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.933810 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r96sr\" (UniqueName: \"kubernetes.io/projected/8336af68-51f9-4f3a-8c92-61dd1aba9067-kube-api-access-r96sr\") pod \"placement-db-sync-m98hv\" (UID: \"8336af68-51f9-4f3a-8c92-61dd1aba9067\") " pod="openstack/placement-db-sync-m98hv" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.964133 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-xs2c8" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.964714 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-hsm9x" Dec 06 06:16:20 crc kubenswrapper[4809]: I1206 06:16:20.998969 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-4gxwk" Dec 06 06:16:21 crc kubenswrapper[4809]: I1206 06:16:21.011763 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93e2d67f-9481-4ca3-83e7-4f7c0673b9d4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"93e2d67f-9481-4ca3-83e7-4f7c0673b9d4\") " pod="openstack/ceilometer-0" Dec 06 06:16:21 crc kubenswrapper[4809]: I1206 06:16:21.012482 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/93e2d67f-9481-4ca3-83e7-4f7c0673b9d4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"93e2d67f-9481-4ca3-83e7-4f7c0673b9d4\") " pod="openstack/ceilometer-0" Dec 06 06:16:21 crc kubenswrapper[4809]: I1206 06:16:21.012525 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93e2d67f-9481-4ca3-83e7-4f7c0673b9d4-config-data\") pod \"ceilometer-0\" (UID: \"93e2d67f-9481-4ca3-83e7-4f7c0673b9d4\") " pod="openstack/ceilometer-0" Dec 06 06:16:21 crc kubenswrapper[4809]: I1206 06:16:21.012572 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zc4q\" (UniqueName: \"kubernetes.io/projected/93e2d67f-9481-4ca3-83e7-4f7c0673b9d4-kube-api-access-6zc4q\") pod \"ceilometer-0\" (UID: \"93e2d67f-9481-4ca3-83e7-4f7c0673b9d4\") " pod="openstack/ceilometer-0" Dec 06 06:16:21 crc kubenswrapper[4809]: I1206 06:16:21.013044 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93e2d67f-9481-4ca3-83e7-4f7c0673b9d4-scripts\") pod \"ceilometer-0\" (UID: \"93e2d67f-9481-4ca3-83e7-4f7c0673b9d4\") " pod="openstack/ceilometer-0" Dec 06 06:16:21 crc kubenswrapper[4809]: I1206 06:16:21.013134 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/46f0989d-3592-45c2-b046-5d33d6fd97b4-dns-svc\") pod \"dnsmasq-dns-68dcc9cf6f-vjzzr\" (UID: \"46f0989d-3592-45c2-b046-5d33d6fd97b4\") " pod="openstack/dnsmasq-dns-68dcc9cf6f-vjzzr" Dec 06 06:16:21 crc kubenswrapper[4809]: I1206 06:16:21.013197 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46f0989d-3592-45c2-b046-5d33d6fd97b4-config\") pod \"dnsmasq-dns-68dcc9cf6f-vjzzr\" (UID: \"46f0989d-3592-45c2-b046-5d33d6fd97b4\") " pod="openstack/dnsmasq-dns-68dcc9cf6f-vjzzr" Dec 06 06:16:21 crc kubenswrapper[4809]: I1206 06:16:21.013269 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ft7cs\" (UniqueName: \"kubernetes.io/projected/46f0989d-3592-45c2-b046-5d33d6fd97b4-kube-api-access-ft7cs\") pod \"dnsmasq-dns-68dcc9cf6f-vjzzr\" (UID: \"46f0989d-3592-45c2-b046-5d33d6fd97b4\") " pod="openstack/dnsmasq-dns-68dcc9cf6f-vjzzr" Dec 06 06:16:21 crc kubenswrapper[4809]: I1206 06:16:21.013310 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/93e2d67f-9481-4ca3-83e7-4f7c0673b9d4-log-httpd\") pod \"ceilometer-0\" (UID: \"93e2d67f-9481-4ca3-83e7-4f7c0673b9d4\") " pod="openstack/ceilometer-0" Dec 06 06:16:21 crc kubenswrapper[4809]: I1206 06:16:21.013390 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/46f0989d-3592-45c2-b046-5d33d6fd97b4-ovsdbserver-nb\") pod \"dnsmasq-dns-68dcc9cf6f-vjzzr\" (UID: \"46f0989d-3592-45c2-b046-5d33d6fd97b4\") " pod="openstack/dnsmasq-dns-68dcc9cf6f-vjzzr" Dec 06 06:16:21 crc kubenswrapper[4809]: I1206 06:16:21.013522 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/46f0989d-3592-45c2-b046-5d33d6fd97b4-ovsdbserver-sb\") pod \"dnsmasq-dns-68dcc9cf6f-vjzzr\" (UID: \"46f0989d-3592-45c2-b046-5d33d6fd97b4\") " pod="openstack/dnsmasq-dns-68dcc9cf6f-vjzzr" Dec 06 06:16:21 crc kubenswrapper[4809]: I1206 06:16:21.013559 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/93e2d67f-9481-4ca3-83e7-4f7c0673b9d4-run-httpd\") pod \"ceilometer-0\" (UID: \"93e2d67f-9481-4ca3-83e7-4f7c0673b9d4\") " pod="openstack/ceilometer-0" Dec 06 06:16:21 crc kubenswrapper[4809]: I1206 06:16:21.018909 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-m98hv" Dec 06 06:16:21 crc kubenswrapper[4809]: I1206 06:16:21.115472 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ft7cs\" (UniqueName: \"kubernetes.io/projected/46f0989d-3592-45c2-b046-5d33d6fd97b4-kube-api-access-ft7cs\") pod \"dnsmasq-dns-68dcc9cf6f-vjzzr\" (UID: \"46f0989d-3592-45c2-b046-5d33d6fd97b4\") " pod="openstack/dnsmasq-dns-68dcc9cf6f-vjzzr" Dec 06 06:16:21 crc kubenswrapper[4809]: I1206 06:16:21.115533 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/93e2d67f-9481-4ca3-83e7-4f7c0673b9d4-log-httpd\") pod \"ceilometer-0\" (UID: \"93e2d67f-9481-4ca3-83e7-4f7c0673b9d4\") " pod="openstack/ceilometer-0" Dec 06 06:16:21 crc kubenswrapper[4809]: I1206 06:16:21.115576 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/46f0989d-3592-45c2-b046-5d33d6fd97b4-ovsdbserver-nb\") pod \"dnsmasq-dns-68dcc9cf6f-vjzzr\" (UID: \"46f0989d-3592-45c2-b046-5d33d6fd97b4\") " pod="openstack/dnsmasq-dns-68dcc9cf6f-vjzzr" Dec 06 06:16:21 crc kubenswrapper[4809]: I1206 06:16:21.115628 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/46f0989d-3592-45c2-b046-5d33d6fd97b4-ovsdbserver-sb\") pod \"dnsmasq-dns-68dcc9cf6f-vjzzr\" (UID: \"46f0989d-3592-45c2-b046-5d33d6fd97b4\") " pod="openstack/dnsmasq-dns-68dcc9cf6f-vjzzr" Dec 06 06:16:21 crc kubenswrapper[4809]: I1206 06:16:21.115666 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/93e2d67f-9481-4ca3-83e7-4f7c0673b9d4-run-httpd\") pod \"ceilometer-0\" (UID: \"93e2d67f-9481-4ca3-83e7-4f7c0673b9d4\") " pod="openstack/ceilometer-0" Dec 06 06:16:21 crc kubenswrapper[4809]: I1206 06:16:21.115738 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93e2d67f-9481-4ca3-83e7-4f7c0673b9d4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"93e2d67f-9481-4ca3-83e7-4f7c0673b9d4\") " pod="openstack/ceilometer-0" Dec 06 06:16:21 crc kubenswrapper[4809]: I1206 06:16:21.115764 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/93e2d67f-9481-4ca3-83e7-4f7c0673b9d4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"93e2d67f-9481-4ca3-83e7-4f7c0673b9d4\") " pod="openstack/ceilometer-0" Dec 06 06:16:21 crc kubenswrapper[4809]: I1206 06:16:21.115782 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93e2d67f-9481-4ca3-83e7-4f7c0673b9d4-config-data\") pod \"ceilometer-0\" (UID: \"93e2d67f-9481-4ca3-83e7-4f7c0673b9d4\") " pod="openstack/ceilometer-0" Dec 06 06:16:21 crc kubenswrapper[4809]: I1206 06:16:21.115813 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zc4q\" (UniqueName: \"kubernetes.io/projected/93e2d67f-9481-4ca3-83e7-4f7c0673b9d4-kube-api-access-6zc4q\") pod \"ceilometer-0\" (UID: \"93e2d67f-9481-4ca3-83e7-4f7c0673b9d4\") " pod="openstack/ceilometer-0" Dec 06 06:16:21 crc kubenswrapper[4809]: I1206 06:16:21.115849 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93e2d67f-9481-4ca3-83e7-4f7c0673b9d4-scripts\") pod \"ceilometer-0\" (UID: \"93e2d67f-9481-4ca3-83e7-4f7c0673b9d4\") " pod="openstack/ceilometer-0" Dec 06 06:16:21 crc kubenswrapper[4809]: I1206 06:16:21.115886 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/46f0989d-3592-45c2-b046-5d33d6fd97b4-dns-svc\") pod \"dnsmasq-dns-68dcc9cf6f-vjzzr\" (UID: \"46f0989d-3592-45c2-b046-5d33d6fd97b4\") " pod="openstack/dnsmasq-dns-68dcc9cf6f-vjzzr" Dec 06 06:16:21 crc kubenswrapper[4809]: I1206 06:16:21.115911 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46f0989d-3592-45c2-b046-5d33d6fd97b4-config\") pod \"dnsmasq-dns-68dcc9cf6f-vjzzr\" (UID: \"46f0989d-3592-45c2-b046-5d33d6fd97b4\") " pod="openstack/dnsmasq-dns-68dcc9cf6f-vjzzr" Dec 06 06:16:21 crc kubenswrapper[4809]: I1206 06:16:21.117233 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/46f0989d-3592-45c2-b046-5d33d6fd97b4-ovsdbserver-nb\") pod \"dnsmasq-dns-68dcc9cf6f-vjzzr\" (UID: \"46f0989d-3592-45c2-b046-5d33d6fd97b4\") " pod="openstack/dnsmasq-dns-68dcc9cf6f-vjzzr" Dec 06 06:16:21 crc kubenswrapper[4809]: I1206 06:16:21.117226 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/46f0989d-3592-45c2-b046-5d33d6fd97b4-ovsdbserver-sb\") pod \"dnsmasq-dns-68dcc9cf6f-vjzzr\" (UID: \"46f0989d-3592-45c2-b046-5d33d6fd97b4\") " pod="openstack/dnsmasq-dns-68dcc9cf6f-vjzzr" Dec 06 06:16:21 crc kubenswrapper[4809]: I1206 06:16:21.117606 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/93e2d67f-9481-4ca3-83e7-4f7c0673b9d4-run-httpd\") pod \"ceilometer-0\" (UID: \"93e2d67f-9481-4ca3-83e7-4f7c0673b9d4\") " pod="openstack/ceilometer-0" Dec 06 06:16:21 crc kubenswrapper[4809]: I1206 06:16:21.117716 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/93e2d67f-9481-4ca3-83e7-4f7c0673b9d4-log-httpd\") pod \"ceilometer-0\" (UID: \"93e2d67f-9481-4ca3-83e7-4f7c0673b9d4\") " pod="openstack/ceilometer-0" Dec 06 06:16:21 crc kubenswrapper[4809]: I1206 06:16:21.118437 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46f0989d-3592-45c2-b046-5d33d6fd97b4-config\") pod \"dnsmasq-dns-68dcc9cf6f-vjzzr\" (UID: \"46f0989d-3592-45c2-b046-5d33d6fd97b4\") " pod="openstack/dnsmasq-dns-68dcc9cf6f-vjzzr" Dec 06 06:16:21 crc kubenswrapper[4809]: I1206 06:16:21.120121 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/46f0989d-3592-45c2-b046-5d33d6fd97b4-dns-svc\") pod \"dnsmasq-dns-68dcc9cf6f-vjzzr\" (UID: \"46f0989d-3592-45c2-b046-5d33d6fd97b4\") " pod="openstack/dnsmasq-dns-68dcc9cf6f-vjzzr" Dec 06 06:16:21 crc kubenswrapper[4809]: I1206 06:16:21.121249 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93e2d67f-9481-4ca3-83e7-4f7c0673b9d4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"93e2d67f-9481-4ca3-83e7-4f7c0673b9d4\") " pod="openstack/ceilometer-0" Dec 06 06:16:21 crc kubenswrapper[4809]: I1206 06:16:21.123327 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93e2d67f-9481-4ca3-83e7-4f7c0673b9d4-scripts\") pod \"ceilometer-0\" (UID: \"93e2d67f-9481-4ca3-83e7-4f7c0673b9d4\") " pod="openstack/ceilometer-0" Dec 06 06:16:21 crc kubenswrapper[4809]: I1206 06:16:21.123758 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93e2d67f-9481-4ca3-83e7-4f7c0673b9d4-config-data\") pod \"ceilometer-0\" (UID: \"93e2d67f-9481-4ca3-83e7-4f7c0673b9d4\") " pod="openstack/ceilometer-0" Dec 06 06:16:21 crc kubenswrapper[4809]: I1206 06:16:21.130709 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/93e2d67f-9481-4ca3-83e7-4f7c0673b9d4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"93e2d67f-9481-4ca3-83e7-4f7c0673b9d4\") " pod="openstack/ceilometer-0" Dec 06 06:16:21 crc kubenswrapper[4809]: I1206 06:16:21.137272 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zc4q\" (UniqueName: \"kubernetes.io/projected/93e2d67f-9481-4ca3-83e7-4f7c0673b9d4-kube-api-access-6zc4q\") pod \"ceilometer-0\" (UID: \"93e2d67f-9481-4ca3-83e7-4f7c0673b9d4\") " pod="openstack/ceilometer-0" Dec 06 06:16:21 crc kubenswrapper[4809]: I1206 06:16:21.155460 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ft7cs\" (UniqueName: \"kubernetes.io/projected/46f0989d-3592-45c2-b046-5d33d6fd97b4-kube-api-access-ft7cs\") pod \"dnsmasq-dns-68dcc9cf6f-vjzzr\" (UID: \"46f0989d-3592-45c2-b046-5d33d6fd97b4\") " pod="openstack/dnsmasq-dns-68dcc9cf6f-vjzzr" Dec 06 06:16:21 crc kubenswrapper[4809]: I1206 06:16:21.277549 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68dcc9cf6f-vjzzr" Dec 06 06:16:21 crc kubenswrapper[4809]: I1206 06:16:21.300157 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:16:21 crc kubenswrapper[4809]: W1206 06:16:21.417714 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podedf1ab66_b483_4172_81fd_dd3f3b9b44aa.slice/crio-a9627522d80b87d58fb1947fb4b4e91e8a19b1bd560edd020ccf2a4d8fd042e4 WatchSource:0}: Error finding container a9627522d80b87d58fb1947fb4b4e91e8a19b1bd560edd020ccf2a4d8fd042e4: Status 404 returned error can't find the container with id a9627522d80b87d58fb1947fb4b4e91e8a19b1bd560edd020ccf2a4d8fd042e4 Dec 06 06:16:21 crc kubenswrapper[4809]: I1206 06:16:21.434585 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-ckhg9"] Dec 06 06:16:21 crc kubenswrapper[4809]: I1206 06:16:21.538520 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f877ddd87-w7rt2"] Dec 06 06:16:21 crc kubenswrapper[4809]: I1206 06:16:21.750474 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-q29b5"] Dec 06 06:16:21 crc kubenswrapper[4809]: I1206 06:16:21.796893 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f877ddd87-w7rt2" event={"ID":"9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74","Type":"ContainerStarted","Data":"10208dae573ac40edc9fe4502cb297c73cf3c3da260aaf6d4b0150fa1f754055"} Dec 06 06:16:21 crc kubenswrapper[4809]: I1206 06:16:21.798368 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-q29b5" event={"ID":"4f2a4615-0ec1-404a-9ab9-848d8b45be2a","Type":"ContainerStarted","Data":"9578d78e9f50c9e3d82545c18f2a31b880e5ef5f4725256ed4e6533ae91fb0d4"} Dec 06 06:16:21 crc kubenswrapper[4809]: I1206 06:16:21.803069 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-ckhg9" event={"ID":"edf1ab66-b483-4172-81fd-dd3f3b9b44aa","Type":"ContainerStarted","Data":"a9627522d80b87d58fb1947fb4b4e91e8a19b1bd560edd020ccf2a4d8fd042e4"} Dec 06 06:16:22 crc kubenswrapper[4809]: I1206 06:16:22.094968 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-4gxwk"] Dec 06 06:16:22 crc kubenswrapper[4809]: W1206 06:16:22.126090 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08195221_42d1_4ce3_93a6_050e44339013.slice/crio-6e1fb3895287a3164ef0b7f918151a1d6b9bd8377505b3589e7842cd1dbc2702 WatchSource:0}: Error finding container 6e1fb3895287a3164ef0b7f918151a1d6b9bd8377505b3589e7842cd1dbc2702: Status 404 returned error can't find the container with id 6e1fb3895287a3164ef0b7f918151a1d6b9bd8377505b3589e7842cd1dbc2702 Dec 06 06:16:22 crc kubenswrapper[4809]: I1206 06:16:22.128339 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-hsm9x"] Dec 06 06:16:22 crc kubenswrapper[4809]: I1206 06:16:22.397244 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68dcc9cf6f-vjzzr"] Dec 06 06:16:22 crc kubenswrapper[4809]: W1206 06:16:22.410584 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod46f0989d_3592_45c2_b046_5d33d6fd97b4.slice/crio-f952a937b415d1cd00aea947eddd935597ec0f67c8c691066e2a9ac40105151c WatchSource:0}: Error finding container f952a937b415d1cd00aea947eddd935597ec0f67c8c691066e2a9ac40105151c: Status 404 returned error can't find the container with id f952a937b415d1cd00aea947eddd935597ec0f67c8c691066e2a9ac40105151c Dec 06 06:16:22 crc kubenswrapper[4809]: I1206 06:16:22.420999 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-xs2c8"] Dec 06 06:16:22 crc kubenswrapper[4809]: W1206 06:16:22.423098 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc75cf833_5179_4e18_8f74_3faed72757ed.slice/crio-cd2c57afa753ad81e6088dc8cff8286bd25fc0a3b10e34b89956257ae4809bc3 WatchSource:0}: Error finding container cd2c57afa753ad81e6088dc8cff8286bd25fc0a3b10e34b89956257ae4809bc3: Status 404 returned error can't find the container with id cd2c57afa753ad81e6088dc8cff8286bd25fc0a3b10e34b89956257ae4809bc3 Dec 06 06:16:22 crc kubenswrapper[4809]: I1206 06:16:22.476060 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:16:22 crc kubenswrapper[4809]: W1206 06:16:22.476760 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93e2d67f_9481_4ca3_83e7_4f7c0673b9d4.slice/crio-063ec88f2d6bfed89c694189278a247c83add1e67252bd65335a8c84c9afd9f2 WatchSource:0}: Error finding container 063ec88f2d6bfed89c694189278a247c83add1e67252bd65335a8c84c9afd9f2: Status 404 returned error can't find the container with id 063ec88f2d6bfed89c694189278a247c83add1e67252bd65335a8c84c9afd9f2 Dec 06 06:16:22 crc kubenswrapper[4809]: I1206 06:16:22.725452 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-m98hv"] Dec 06 06:16:22 crc kubenswrapper[4809]: I1206 06:16:22.814208 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-hsm9x" event={"ID":"08195221-42d1-4ce3-93a6-050e44339013","Type":"ContainerStarted","Data":"6e1fb3895287a3164ef0b7f918151a1d6b9bd8377505b3589e7842cd1dbc2702"} Dec 06 06:16:22 crc kubenswrapper[4809]: I1206 06:16:22.815534 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-m98hv" event={"ID":"8336af68-51f9-4f3a-8c92-61dd1aba9067","Type":"ContainerStarted","Data":"e670b82e7f278746dda04ed70357e80a34f69e6381934a87d255e3f054730829"} Dec 06 06:16:22 crc kubenswrapper[4809]: I1206 06:16:22.816872 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-xs2c8" event={"ID":"c75cf833-5179-4e18-8f74-3faed72757ed","Type":"ContainerStarted","Data":"cd2c57afa753ad81e6088dc8cff8286bd25fc0a3b10e34b89956257ae4809bc3"} Dec 06 06:16:22 crc kubenswrapper[4809]: I1206 06:16:22.817957 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-4gxwk" event={"ID":"662e24e4-5dd6-4d18-822c-2863017f15af","Type":"ContainerStarted","Data":"a5bb6b2166abf5b42fe39fdb1928631a4d7d9d596113bc95699f0ce321b401c9"} Dec 06 06:16:22 crc kubenswrapper[4809]: I1206 06:16:22.822053 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cfd13e5d-ca28-4061-8ca5-2c4566332bf0","Type":"ContainerStarted","Data":"7b063892e7c6bb73167eac13521837f6799140c8519c5d2adcd082427846b893"} Dec 06 06:16:22 crc kubenswrapper[4809]: I1206 06:16:22.824202 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"93e2d67f-9481-4ca3-83e7-4f7c0673b9d4","Type":"ContainerStarted","Data":"063ec88f2d6bfed89c694189278a247c83add1e67252bd65335a8c84c9afd9f2"} Dec 06 06:16:22 crc kubenswrapper[4809]: I1206 06:16:22.826157 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68dcc9cf6f-vjzzr" event={"ID":"46f0989d-3592-45c2-b046-5d33d6fd97b4","Type":"ContainerStarted","Data":"f952a937b415d1cd00aea947eddd935597ec0f67c8c691066e2a9ac40105151c"} Dec 06 06:16:22 crc kubenswrapper[4809]: I1206 06:16:22.942723 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:16:23 crc kubenswrapper[4809]: I1206 06:16:23.847466 4809 generic.go:334] "Generic (PLEG): container finished" podID="46f0989d-3592-45c2-b046-5d33d6fd97b4" containerID="f1127a18075a3c31522ec64330d3d64798b254aeca584a9632c89a7ab71af4cb" exitCode=0 Dec 06 06:16:23 crc kubenswrapper[4809]: I1206 06:16:23.847686 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68dcc9cf6f-vjzzr" event={"ID":"46f0989d-3592-45c2-b046-5d33d6fd97b4","Type":"ContainerDied","Data":"f1127a18075a3c31522ec64330d3d64798b254aeca584a9632c89a7ab71af4cb"} Dec 06 06:16:23 crc kubenswrapper[4809]: I1206 06:16:23.855412 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-hsm9x" event={"ID":"08195221-42d1-4ce3-93a6-050e44339013","Type":"ContainerStarted","Data":"7c24eacb26efb396e941d2db0e9a7e55a9a496af0f47192b9d1f6a49c739e882"} Dec 06 06:16:23 crc kubenswrapper[4809]: I1206 06:16:23.859691 4809 generic.go:334] "Generic (PLEG): container finished" podID="9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74" containerID="197da59d8485397fdcbe97d5cae31e99f72a51b3b068dd30ec59be80ac3a1948" exitCode=0 Dec 06 06:16:23 crc kubenswrapper[4809]: I1206 06:16:23.859979 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f877ddd87-w7rt2" event={"ID":"9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74","Type":"ContainerDied","Data":"197da59d8485397fdcbe97d5cae31e99f72a51b3b068dd30ec59be80ac3a1948"} Dec 06 06:16:23 crc kubenswrapper[4809]: I1206 06:16:23.868090 4809 generic.go:334] "Generic (PLEG): container finished" podID="ca8e11eb-2da1-4706-b589-650bcb817d4c" containerID="209f5febf4d99fd8dcb763ad4fd6fff03a5214b24e55afc99542f4521334df09" exitCode=0 Dec 06 06:16:23 crc kubenswrapper[4809]: I1206 06:16:23.868162 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ca8e11eb-2da1-4706-b589-650bcb817d4c","Type":"ContainerDied","Data":"209f5febf4d99fd8dcb763ad4fd6fff03a5214b24e55afc99542f4521334df09"} Dec 06 06:16:23 crc kubenswrapper[4809]: I1206 06:16:23.881128 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-q29b5" event={"ID":"4f2a4615-0ec1-404a-9ab9-848d8b45be2a","Type":"ContainerStarted","Data":"d7f6127f423b0d477bde7856693ff15b1c9c16bf8a4fec6fd9c275df64f72120"} Dec 06 06:16:23 crc kubenswrapper[4809]: I1206 06:16:23.921526 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-hsm9x" podStartSLOduration=3.921497864 podStartE2EDuration="3.921497864s" podCreationTimestamp="2025-12-06 06:16:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:16:23.89984355 +0000 UTC m=+1508.788826492" watchObservedRunningTime="2025-12-06 06:16:23.921497864 +0000 UTC m=+1508.810480806" Dec 06 06:16:24 crc kubenswrapper[4809]: I1206 06:16:24.098223 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-q29b5" podStartSLOduration=4.098196391 podStartE2EDuration="4.098196391s" podCreationTimestamp="2025-12-06 06:16:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:16:24.022433407 +0000 UTC m=+1508.911416359" watchObservedRunningTime="2025-12-06 06:16:24.098196391 +0000 UTC m=+1508.987179343" Dec 06 06:16:24 crc kubenswrapper[4809]: I1206 06:16:24.611446 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f877ddd87-w7rt2" Dec 06 06:16:24 crc kubenswrapper[4809]: I1206 06:16:24.748714 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4bkkl\" (UniqueName: \"kubernetes.io/projected/9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74-kube-api-access-4bkkl\") pod \"9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74\" (UID: \"9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74\") " Dec 06 06:16:24 crc kubenswrapper[4809]: I1206 06:16:24.748831 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74-ovsdbserver-nb\") pod \"9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74\" (UID: \"9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74\") " Dec 06 06:16:24 crc kubenswrapper[4809]: I1206 06:16:24.748877 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74-config\") pod \"9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74\" (UID: \"9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74\") " Dec 06 06:16:24 crc kubenswrapper[4809]: I1206 06:16:24.749138 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74-ovsdbserver-sb\") pod \"9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74\" (UID: \"9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74\") " Dec 06 06:16:24 crc kubenswrapper[4809]: I1206 06:16:24.749186 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74-dns-svc\") pod \"9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74\" (UID: \"9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74\") " Dec 06 06:16:24 crc kubenswrapper[4809]: I1206 06:16:24.769289 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74-kube-api-access-4bkkl" (OuterVolumeSpecName: "kube-api-access-4bkkl") pod "9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74" (UID: "9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74"). InnerVolumeSpecName "kube-api-access-4bkkl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:16:24 crc kubenswrapper[4809]: I1206 06:16:24.812578 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74" (UID: "9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:16:24 crc kubenswrapper[4809]: I1206 06:16:24.821060 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74" (UID: "9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:16:24 crc kubenswrapper[4809]: I1206 06:16:24.831059 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74" (UID: "9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:16:24 crc kubenswrapper[4809]: I1206 06:16:24.852275 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:24 crc kubenswrapper[4809]: I1206 06:16:24.852312 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:24 crc kubenswrapper[4809]: I1206 06:16:24.852324 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4bkkl\" (UniqueName: \"kubernetes.io/projected/9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74-kube-api-access-4bkkl\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:24 crc kubenswrapper[4809]: I1206 06:16:24.852335 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:24 crc kubenswrapper[4809]: I1206 06:16:24.864905 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74-config" (OuterVolumeSpecName: "config") pod "9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74" (UID: "9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:16:24 crc kubenswrapper[4809]: I1206 06:16:24.922235 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cfd13e5d-ca28-4061-8ca5-2c4566332bf0","Type":"ContainerStarted","Data":"4e96ab39f8481c618b4527e59d81dda14b0f2faa874f25839efb8e8b8b2fe4f8"} Dec 06 06:16:24 crc kubenswrapper[4809]: I1206 06:16:24.925230 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f877ddd87-w7rt2" event={"ID":"9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74","Type":"ContainerDied","Data":"10208dae573ac40edc9fe4502cb297c73cf3c3da260aaf6d4b0150fa1f754055"} Dec 06 06:16:24 crc kubenswrapper[4809]: I1206 06:16:24.925287 4809 scope.go:117] "RemoveContainer" containerID="197da59d8485397fdcbe97d5cae31e99f72a51b3b068dd30ec59be80ac3a1948" Dec 06 06:16:24 crc kubenswrapper[4809]: I1206 06:16:24.925452 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f877ddd87-w7rt2" Dec 06 06:16:24 crc kubenswrapper[4809]: I1206 06:16:24.944558 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ca8e11eb-2da1-4706-b589-650bcb817d4c","Type":"ContainerStarted","Data":"26d0ce78802b68c255a613ffb9196adf8b899f26444687f4cc0b2b78232512ab"} Dec 06 06:16:24 crc kubenswrapper[4809]: I1206 06:16:24.948721 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68dcc9cf6f-vjzzr" event={"ID":"46f0989d-3592-45c2-b046-5d33d6fd97b4","Type":"ContainerStarted","Data":"33eab5c70cd4af9a2c988a431a752ead40b0627ed370c368fbbdc0f17aef3e78"} Dec 06 06:16:24 crc kubenswrapper[4809]: I1206 06:16:24.948892 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-68dcc9cf6f-vjzzr" Dec 06 06:16:24 crc kubenswrapper[4809]: I1206 06:16:24.954843 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:25 crc kubenswrapper[4809]: I1206 06:16:25.001869 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f877ddd87-w7rt2"] Dec 06 06:16:25 crc kubenswrapper[4809]: I1206 06:16:25.032922 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f877ddd87-w7rt2"] Dec 06 06:16:25 crc kubenswrapper[4809]: I1206 06:16:25.057593 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-68dcc9cf6f-vjzzr" podStartSLOduration=5.057573078 podStartE2EDuration="5.057573078s" podCreationTimestamp="2025-12-06 06:16:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:16:25.005663237 +0000 UTC m=+1509.894646179" watchObservedRunningTime="2025-12-06 06:16:25.057573078 +0000 UTC m=+1509.946556020" Dec 06 06:16:25 crc kubenswrapper[4809]: E1206 06:16:25.130201 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9b18bcd3_2177_4c2d_aace_f8f5b1ef0e74.slice/crio-10208dae573ac40edc9fe4502cb297c73cf3c3da260aaf6d4b0150fa1f754055\": RecentStats: unable to find data in memory cache]" Dec 06 06:16:25 crc kubenswrapper[4809]: I1206 06:16:25.474788 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74" path="/var/lib/kubelet/pods/9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74/volumes" Dec 06 06:16:25 crc kubenswrapper[4809]: I1206 06:16:25.994623 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cfd13e5d-ca28-4061-8ca5-2c4566332bf0","Type":"ContainerStarted","Data":"0b316681dab835f05cd3a86c6af03ceb8ee0fcd06109af37959aac5908bfb1ad"} Dec 06 06:16:25 crc kubenswrapper[4809]: I1206 06:16:25.994667 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cfd13e5d-ca28-4061-8ca5-2c4566332bf0","Type":"ContainerStarted","Data":"09a749240258c90e84bf4badf6c34120274f6055665c362635668d15de2595c8"} Dec 06 06:16:27 crc kubenswrapper[4809]: I1206 06:16:27.017692 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cfd13e5d-ca28-4061-8ca5-2c4566332bf0","Type":"ContainerStarted","Data":"540f9ce834e8426f6e9d100dbc97f415270e1e3d602a27978bbd6846f415f43d"} Dec 06 06:16:27 crc kubenswrapper[4809]: I1206 06:16:27.018603 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cfd13e5d-ca28-4061-8ca5-2c4566332bf0","Type":"ContainerStarted","Data":"d1be5afd1dc39e81f62e6ef73679916553cbf1936812623d4bdb3693ed384146"} Dec 06 06:16:28 crc kubenswrapper[4809]: I1206 06:16:28.039970 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-mc9z8" event={"ID":"413a30cf-4b57-44d7-b9bf-3d400bf4b897","Type":"ContainerStarted","Data":"5fa4ff9a4f10e7d52d4499104cc17b8069a880b80c0127d3a607c15b16f88cf1"} Dec 06 06:16:28 crc kubenswrapper[4809]: I1206 06:16:28.065437 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-mc9z8" podStartSLOduration=3.062613924 podStartE2EDuration="40.065415624s" podCreationTimestamp="2025-12-06 06:15:48 +0000 UTC" firstStartedPulling="2025-12-06 06:15:49.289561794 +0000 UTC m=+1474.178544736" lastFinishedPulling="2025-12-06 06:16:26.292363494 +0000 UTC m=+1511.181346436" observedRunningTime="2025-12-06 06:16:28.059421903 +0000 UTC m=+1512.948404845" watchObservedRunningTime="2025-12-06 06:16:28.065415624 +0000 UTC m=+1512.954398566" Dec 06 06:16:28 crc kubenswrapper[4809]: I1206 06:16:28.086864 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cfd13e5d-ca28-4061-8ca5-2c4566332bf0","Type":"ContainerStarted","Data":"6a2b23417fb249cc3c804c274c4742bb568dc140fb4a9bd054e538be75e631d2"} Dec 06 06:16:28 crc kubenswrapper[4809]: I1206 06:16:28.099203 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ca8e11eb-2da1-4706-b589-650bcb817d4c","Type":"ContainerStarted","Data":"f2566c9a9763dbfff5ef90ecbd639ac425a6d5c4f11ab8323f32857353f8e644"} Dec 06 06:16:28 crc kubenswrapper[4809]: I1206 06:16:28.143796 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=46.932883957 podStartE2EDuration="1m9.143761138s" podCreationTimestamp="2025-12-06 06:15:19 +0000 UTC" firstStartedPulling="2025-12-06 06:15:56.622517048 +0000 UTC m=+1481.511499990" lastFinishedPulling="2025-12-06 06:16:18.833394229 +0000 UTC m=+1503.722377171" observedRunningTime="2025-12-06 06:16:28.128440194 +0000 UTC m=+1513.017423136" watchObservedRunningTime="2025-12-06 06:16:28.143761138 +0000 UTC m=+1513.032744110" Dec 06 06:16:28 crc kubenswrapper[4809]: I1206 06:16:28.490228 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68dcc9cf6f-vjzzr"] Dec 06 06:16:28 crc kubenswrapper[4809]: I1206 06:16:28.490510 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-68dcc9cf6f-vjzzr" podUID="46f0989d-3592-45c2-b046-5d33d6fd97b4" containerName="dnsmasq-dns" containerID="cri-o://33eab5c70cd4af9a2c988a431a752ead40b0627ed370c368fbbdc0f17aef3e78" gracePeriod=10 Dec 06 06:16:28 crc kubenswrapper[4809]: I1206 06:16:28.538415 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-b75nx"] Dec 06 06:16:28 crc kubenswrapper[4809]: E1206 06:16:28.539372 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74" containerName="init" Dec 06 06:16:28 crc kubenswrapper[4809]: I1206 06:16:28.539399 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74" containerName="init" Dec 06 06:16:28 crc kubenswrapper[4809]: I1206 06:16:28.539691 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b18bcd3-2177-4c2d-aace-f8f5b1ef0e74" containerName="init" Dec 06 06:16:28 crc kubenswrapper[4809]: I1206 06:16:28.541203 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-b75nx" Dec 06 06:16:28 crc kubenswrapper[4809]: I1206 06:16:28.543862 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 06 06:16:28 crc kubenswrapper[4809]: I1206 06:16:28.550911 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-b75nx"] Dec 06 06:16:28 crc kubenswrapper[4809]: I1206 06:16:28.583398 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40abc93a-e45f-44cb-8fb1-f14961fe3e74-config\") pod \"dnsmasq-dns-58dd9ff6bc-b75nx\" (UID: \"40abc93a-e45f-44cb-8fb1-f14961fe3e74\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-b75nx" Dec 06 06:16:28 crc kubenswrapper[4809]: I1206 06:16:28.583450 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/40abc93a-e45f-44cb-8fb1-f14961fe3e74-dns-svc\") pod \"dnsmasq-dns-58dd9ff6bc-b75nx\" (UID: \"40abc93a-e45f-44cb-8fb1-f14961fe3e74\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-b75nx" Dec 06 06:16:28 crc kubenswrapper[4809]: I1206 06:16:28.583468 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/40abc93a-e45f-44cb-8fb1-f14961fe3e74-ovsdbserver-nb\") pod \"dnsmasq-dns-58dd9ff6bc-b75nx\" (UID: \"40abc93a-e45f-44cb-8fb1-f14961fe3e74\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-b75nx" Dec 06 06:16:28 crc kubenswrapper[4809]: I1206 06:16:28.583551 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/40abc93a-e45f-44cb-8fb1-f14961fe3e74-ovsdbserver-sb\") pod \"dnsmasq-dns-58dd9ff6bc-b75nx\" (UID: \"40abc93a-e45f-44cb-8fb1-f14961fe3e74\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-b75nx" Dec 06 06:16:28 crc kubenswrapper[4809]: I1206 06:16:28.583607 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wngr7\" (UniqueName: \"kubernetes.io/projected/40abc93a-e45f-44cb-8fb1-f14961fe3e74-kube-api-access-wngr7\") pod \"dnsmasq-dns-58dd9ff6bc-b75nx\" (UID: \"40abc93a-e45f-44cb-8fb1-f14961fe3e74\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-b75nx" Dec 06 06:16:28 crc kubenswrapper[4809]: I1206 06:16:28.583709 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/40abc93a-e45f-44cb-8fb1-f14961fe3e74-dns-swift-storage-0\") pod \"dnsmasq-dns-58dd9ff6bc-b75nx\" (UID: \"40abc93a-e45f-44cb-8fb1-f14961fe3e74\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-b75nx" Dec 06 06:16:28 crc kubenswrapper[4809]: I1206 06:16:28.697126 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/40abc93a-e45f-44cb-8fb1-f14961fe3e74-ovsdbserver-sb\") pod \"dnsmasq-dns-58dd9ff6bc-b75nx\" (UID: \"40abc93a-e45f-44cb-8fb1-f14961fe3e74\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-b75nx" Dec 06 06:16:28 crc kubenswrapper[4809]: I1206 06:16:28.697277 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wngr7\" (UniqueName: \"kubernetes.io/projected/40abc93a-e45f-44cb-8fb1-f14961fe3e74-kube-api-access-wngr7\") pod \"dnsmasq-dns-58dd9ff6bc-b75nx\" (UID: \"40abc93a-e45f-44cb-8fb1-f14961fe3e74\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-b75nx" Dec 06 06:16:28 crc kubenswrapper[4809]: I1206 06:16:28.697401 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/40abc93a-e45f-44cb-8fb1-f14961fe3e74-dns-swift-storage-0\") pod \"dnsmasq-dns-58dd9ff6bc-b75nx\" (UID: \"40abc93a-e45f-44cb-8fb1-f14961fe3e74\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-b75nx" Dec 06 06:16:28 crc kubenswrapper[4809]: I1206 06:16:28.697455 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40abc93a-e45f-44cb-8fb1-f14961fe3e74-config\") pod \"dnsmasq-dns-58dd9ff6bc-b75nx\" (UID: \"40abc93a-e45f-44cb-8fb1-f14961fe3e74\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-b75nx" Dec 06 06:16:28 crc kubenswrapper[4809]: I1206 06:16:28.697534 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/40abc93a-e45f-44cb-8fb1-f14961fe3e74-dns-svc\") pod \"dnsmasq-dns-58dd9ff6bc-b75nx\" (UID: \"40abc93a-e45f-44cb-8fb1-f14961fe3e74\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-b75nx" Dec 06 06:16:28 crc kubenswrapper[4809]: I1206 06:16:28.697557 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/40abc93a-e45f-44cb-8fb1-f14961fe3e74-ovsdbserver-nb\") pod \"dnsmasq-dns-58dd9ff6bc-b75nx\" (UID: \"40abc93a-e45f-44cb-8fb1-f14961fe3e74\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-b75nx" Dec 06 06:16:28 crc kubenswrapper[4809]: I1206 06:16:28.699547 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/40abc93a-e45f-44cb-8fb1-f14961fe3e74-ovsdbserver-nb\") pod \"dnsmasq-dns-58dd9ff6bc-b75nx\" (UID: \"40abc93a-e45f-44cb-8fb1-f14961fe3e74\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-b75nx" Dec 06 06:16:28 crc kubenswrapper[4809]: I1206 06:16:28.700306 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/40abc93a-e45f-44cb-8fb1-f14961fe3e74-ovsdbserver-sb\") pod \"dnsmasq-dns-58dd9ff6bc-b75nx\" (UID: \"40abc93a-e45f-44cb-8fb1-f14961fe3e74\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-b75nx" Dec 06 06:16:28 crc kubenswrapper[4809]: I1206 06:16:28.714397 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40abc93a-e45f-44cb-8fb1-f14961fe3e74-config\") pod \"dnsmasq-dns-58dd9ff6bc-b75nx\" (UID: \"40abc93a-e45f-44cb-8fb1-f14961fe3e74\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-b75nx" Dec 06 06:16:28 crc kubenswrapper[4809]: I1206 06:16:28.714566 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/40abc93a-e45f-44cb-8fb1-f14961fe3e74-dns-svc\") pod \"dnsmasq-dns-58dd9ff6bc-b75nx\" (UID: \"40abc93a-e45f-44cb-8fb1-f14961fe3e74\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-b75nx" Dec 06 06:16:28 crc kubenswrapper[4809]: I1206 06:16:28.716205 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/40abc93a-e45f-44cb-8fb1-f14961fe3e74-dns-swift-storage-0\") pod \"dnsmasq-dns-58dd9ff6bc-b75nx\" (UID: \"40abc93a-e45f-44cb-8fb1-f14961fe3e74\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-b75nx" Dec 06 06:16:28 crc kubenswrapper[4809]: I1206 06:16:28.743276 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wngr7\" (UniqueName: \"kubernetes.io/projected/40abc93a-e45f-44cb-8fb1-f14961fe3e74-kube-api-access-wngr7\") pod \"dnsmasq-dns-58dd9ff6bc-b75nx\" (UID: \"40abc93a-e45f-44cb-8fb1-f14961fe3e74\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-b75nx" Dec 06 06:16:28 crc kubenswrapper[4809]: I1206 06:16:28.929044 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-b75nx" Dec 06 06:16:29 crc kubenswrapper[4809]: I1206 06:16:29.127275 4809 generic.go:334] "Generic (PLEG): container finished" podID="4f2a4615-0ec1-404a-9ab9-848d8b45be2a" containerID="d7f6127f423b0d477bde7856693ff15b1c9c16bf8a4fec6fd9c275df64f72120" exitCode=0 Dec 06 06:16:29 crc kubenswrapper[4809]: I1206 06:16:29.127367 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-q29b5" event={"ID":"4f2a4615-0ec1-404a-9ab9-848d8b45be2a","Type":"ContainerDied","Data":"d7f6127f423b0d477bde7856693ff15b1c9c16bf8a4fec6fd9c275df64f72120"} Dec 06 06:16:29 crc kubenswrapper[4809]: I1206 06:16:29.138830 4809 generic.go:334] "Generic (PLEG): container finished" podID="46f0989d-3592-45c2-b046-5d33d6fd97b4" containerID="33eab5c70cd4af9a2c988a431a752ead40b0627ed370c368fbbdc0f17aef3e78" exitCode=0 Dec 06 06:16:29 crc kubenswrapper[4809]: I1206 06:16:29.140089 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68dcc9cf6f-vjzzr" event={"ID":"46f0989d-3592-45c2-b046-5d33d6fd97b4","Type":"ContainerDied","Data":"33eab5c70cd4af9a2c988a431a752ead40b0627ed370c368fbbdc0f17aef3e78"} Dec 06 06:16:31 crc kubenswrapper[4809]: I1206 06:16:31.279540 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-68dcc9cf6f-vjzzr" podUID="46f0989d-3592-45c2-b046-5d33d6fd97b4" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.179:5353: connect: connection refused" Dec 06 06:16:41 crc kubenswrapper[4809]: I1206 06:16:41.279071 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-68dcc9cf6f-vjzzr" podUID="46f0989d-3592-45c2-b046-5d33d6fd97b4" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.179:5353: i/o timeout" Dec 06 06:16:41 crc kubenswrapper[4809]: I1206 06:16:41.312923 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-q29b5" event={"ID":"4f2a4615-0ec1-404a-9ab9-848d8b45be2a","Type":"ContainerDied","Data":"9578d78e9f50c9e3d82545c18f2a31b880e5ef5f4725256ed4e6533ae91fb0d4"} Dec 06 06:16:41 crc kubenswrapper[4809]: I1206 06:16:41.312996 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9578d78e9f50c9e3d82545c18f2a31b880e5ef5f4725256ed4e6533ae91fb0d4" Dec 06 06:16:41 crc kubenswrapper[4809]: I1206 06:16:41.396168 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-q29b5" Dec 06 06:16:41 crc kubenswrapper[4809]: I1206 06:16:41.526139 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f2a4615-0ec1-404a-9ab9-848d8b45be2a-scripts\") pod \"4f2a4615-0ec1-404a-9ab9-848d8b45be2a\" (UID: \"4f2a4615-0ec1-404a-9ab9-848d8b45be2a\") " Dec 06 06:16:41 crc kubenswrapper[4809]: I1206 06:16:41.526254 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f2a4615-0ec1-404a-9ab9-848d8b45be2a-combined-ca-bundle\") pod \"4f2a4615-0ec1-404a-9ab9-848d8b45be2a\" (UID: \"4f2a4615-0ec1-404a-9ab9-848d8b45be2a\") " Dec 06 06:16:41 crc kubenswrapper[4809]: I1206 06:16:41.526313 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4f2a4615-0ec1-404a-9ab9-848d8b45be2a-fernet-keys\") pod \"4f2a4615-0ec1-404a-9ab9-848d8b45be2a\" (UID: \"4f2a4615-0ec1-404a-9ab9-848d8b45be2a\") " Dec 06 06:16:41 crc kubenswrapper[4809]: I1206 06:16:41.526348 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tvplm\" (UniqueName: \"kubernetes.io/projected/4f2a4615-0ec1-404a-9ab9-848d8b45be2a-kube-api-access-tvplm\") pod \"4f2a4615-0ec1-404a-9ab9-848d8b45be2a\" (UID: \"4f2a4615-0ec1-404a-9ab9-848d8b45be2a\") " Dec 06 06:16:41 crc kubenswrapper[4809]: I1206 06:16:41.526523 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f2a4615-0ec1-404a-9ab9-848d8b45be2a-config-data\") pod \"4f2a4615-0ec1-404a-9ab9-848d8b45be2a\" (UID: \"4f2a4615-0ec1-404a-9ab9-848d8b45be2a\") " Dec 06 06:16:41 crc kubenswrapper[4809]: I1206 06:16:41.526569 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4f2a4615-0ec1-404a-9ab9-848d8b45be2a-credential-keys\") pod \"4f2a4615-0ec1-404a-9ab9-848d8b45be2a\" (UID: \"4f2a4615-0ec1-404a-9ab9-848d8b45be2a\") " Dec 06 06:16:41 crc kubenswrapper[4809]: I1206 06:16:41.537062 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f2a4615-0ec1-404a-9ab9-848d8b45be2a-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "4f2a4615-0ec1-404a-9ab9-848d8b45be2a" (UID: "4f2a4615-0ec1-404a-9ab9-848d8b45be2a"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:16:41 crc kubenswrapper[4809]: I1206 06:16:41.537132 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f2a4615-0ec1-404a-9ab9-848d8b45be2a-scripts" (OuterVolumeSpecName: "scripts") pod "4f2a4615-0ec1-404a-9ab9-848d8b45be2a" (UID: "4f2a4615-0ec1-404a-9ab9-848d8b45be2a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:16:41 crc kubenswrapper[4809]: I1206 06:16:41.537218 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f2a4615-0ec1-404a-9ab9-848d8b45be2a-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "4f2a4615-0ec1-404a-9ab9-848d8b45be2a" (UID: "4f2a4615-0ec1-404a-9ab9-848d8b45be2a"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:16:41 crc kubenswrapper[4809]: I1206 06:16:41.542306 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f2a4615-0ec1-404a-9ab9-848d8b45be2a-kube-api-access-tvplm" (OuterVolumeSpecName: "kube-api-access-tvplm") pod "4f2a4615-0ec1-404a-9ab9-848d8b45be2a" (UID: "4f2a4615-0ec1-404a-9ab9-848d8b45be2a"). InnerVolumeSpecName "kube-api-access-tvplm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:16:41 crc kubenswrapper[4809]: I1206 06:16:41.558387 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f2a4615-0ec1-404a-9ab9-848d8b45be2a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4f2a4615-0ec1-404a-9ab9-848d8b45be2a" (UID: "4f2a4615-0ec1-404a-9ab9-848d8b45be2a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:16:41 crc kubenswrapper[4809]: I1206 06:16:41.562697 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f2a4615-0ec1-404a-9ab9-848d8b45be2a-config-data" (OuterVolumeSpecName: "config-data") pod "4f2a4615-0ec1-404a-9ab9-848d8b45be2a" (UID: "4f2a4615-0ec1-404a-9ab9-848d8b45be2a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:16:41 crc kubenswrapper[4809]: I1206 06:16:41.629177 4809 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4f2a4615-0ec1-404a-9ab9-848d8b45be2a-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:41 crc kubenswrapper[4809]: I1206 06:16:41.629220 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tvplm\" (UniqueName: \"kubernetes.io/projected/4f2a4615-0ec1-404a-9ab9-848d8b45be2a-kube-api-access-tvplm\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:41 crc kubenswrapper[4809]: I1206 06:16:41.629234 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f2a4615-0ec1-404a-9ab9-848d8b45be2a-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:41 crc kubenswrapper[4809]: I1206 06:16:41.629244 4809 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4f2a4615-0ec1-404a-9ab9-848d8b45be2a-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:41 crc kubenswrapper[4809]: I1206 06:16:41.629255 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f2a4615-0ec1-404a-9ab9-848d8b45be2a-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:41 crc kubenswrapper[4809]: I1206 06:16:41.629265 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f2a4615-0ec1-404a-9ab9-848d8b45be2a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:41 crc kubenswrapper[4809]: E1206 06:16:41.936925 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Dec 06 06:16:41 crc kubenswrapper[4809]: E1206 06:16:41.937142 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qpp7j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-4gxwk_openstack(662e24e4-5dd6-4d18-822c-2863017f15af): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 06:16:41 crc kubenswrapper[4809]: E1206 06:16:41.938366 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-4gxwk" podUID="662e24e4-5dd6-4d18-822c-2863017f15af" Dec 06 06:16:42 crc kubenswrapper[4809]: I1206 06:16:42.340463 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-q29b5" Dec 06 06:16:42 crc kubenswrapper[4809]: E1206 06:16:42.360661 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-4gxwk" podUID="662e24e4-5dd6-4d18-822c-2863017f15af" Dec 06 06:16:42 crc kubenswrapper[4809]: I1206 06:16:42.489001 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-q29b5"] Dec 06 06:16:42 crc kubenswrapper[4809]: I1206 06:16:42.504889 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-q29b5"] Dec 06 06:16:42 crc kubenswrapper[4809]: I1206 06:16:42.577542 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-tc5gt"] Dec 06 06:16:42 crc kubenswrapper[4809]: E1206 06:16:42.578278 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f2a4615-0ec1-404a-9ab9-848d8b45be2a" containerName="keystone-bootstrap" Dec 06 06:16:42 crc kubenswrapper[4809]: I1206 06:16:42.578306 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f2a4615-0ec1-404a-9ab9-848d8b45be2a" containerName="keystone-bootstrap" Dec 06 06:16:42 crc kubenswrapper[4809]: I1206 06:16:42.578575 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f2a4615-0ec1-404a-9ab9-848d8b45be2a" containerName="keystone-bootstrap" Dec 06 06:16:42 crc kubenswrapper[4809]: I1206 06:16:42.579613 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-tc5gt" Dec 06 06:16:42 crc kubenswrapper[4809]: I1206 06:16:42.582721 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 06 06:16:42 crc kubenswrapper[4809]: I1206 06:16:42.585238 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 06 06:16:42 crc kubenswrapper[4809]: I1206 06:16:42.585681 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 06 06:16:42 crc kubenswrapper[4809]: I1206 06:16:42.585875 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 06 06:16:42 crc kubenswrapper[4809]: I1206 06:16:42.586108 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-2h2rg" Dec 06 06:16:42 crc kubenswrapper[4809]: I1206 06:16:42.587172 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-tc5gt"] Dec 06 06:16:42 crc kubenswrapper[4809]: I1206 06:16:42.656389 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c805cd51-5aaa-4c78-b98e-29a5a6ae44a9-credential-keys\") pod \"keystone-bootstrap-tc5gt\" (UID: \"c805cd51-5aaa-4c78-b98e-29a5a6ae44a9\") " pod="openstack/keystone-bootstrap-tc5gt" Dec 06 06:16:42 crc kubenswrapper[4809]: I1206 06:16:42.656862 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c805cd51-5aaa-4c78-b98e-29a5a6ae44a9-config-data\") pod \"keystone-bootstrap-tc5gt\" (UID: \"c805cd51-5aaa-4c78-b98e-29a5a6ae44a9\") " pod="openstack/keystone-bootstrap-tc5gt" Dec 06 06:16:42 crc kubenswrapper[4809]: I1206 06:16:42.657225 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c805cd51-5aaa-4c78-b98e-29a5a6ae44a9-fernet-keys\") pod \"keystone-bootstrap-tc5gt\" (UID: \"c805cd51-5aaa-4c78-b98e-29a5a6ae44a9\") " pod="openstack/keystone-bootstrap-tc5gt" Dec 06 06:16:42 crc kubenswrapper[4809]: I1206 06:16:42.657685 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c805cd51-5aaa-4c78-b98e-29a5a6ae44a9-scripts\") pod \"keystone-bootstrap-tc5gt\" (UID: \"c805cd51-5aaa-4c78-b98e-29a5a6ae44a9\") " pod="openstack/keystone-bootstrap-tc5gt" Dec 06 06:16:42 crc kubenswrapper[4809]: I1206 06:16:42.657846 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c805cd51-5aaa-4c78-b98e-29a5a6ae44a9-combined-ca-bundle\") pod \"keystone-bootstrap-tc5gt\" (UID: \"c805cd51-5aaa-4c78-b98e-29a5a6ae44a9\") " pod="openstack/keystone-bootstrap-tc5gt" Dec 06 06:16:42 crc kubenswrapper[4809]: I1206 06:16:42.658036 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpsfs\" (UniqueName: \"kubernetes.io/projected/c805cd51-5aaa-4c78-b98e-29a5a6ae44a9-kube-api-access-vpsfs\") pod \"keystone-bootstrap-tc5gt\" (UID: \"c805cd51-5aaa-4c78-b98e-29a5a6ae44a9\") " pod="openstack/keystone-bootstrap-tc5gt" Dec 06 06:16:42 crc kubenswrapper[4809]: I1206 06:16:42.761122 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c805cd51-5aaa-4c78-b98e-29a5a6ae44a9-fernet-keys\") pod \"keystone-bootstrap-tc5gt\" (UID: \"c805cd51-5aaa-4c78-b98e-29a5a6ae44a9\") " pod="openstack/keystone-bootstrap-tc5gt" Dec 06 06:16:42 crc kubenswrapper[4809]: I1206 06:16:42.761272 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c805cd51-5aaa-4c78-b98e-29a5a6ae44a9-scripts\") pod \"keystone-bootstrap-tc5gt\" (UID: \"c805cd51-5aaa-4c78-b98e-29a5a6ae44a9\") " pod="openstack/keystone-bootstrap-tc5gt" Dec 06 06:16:42 crc kubenswrapper[4809]: I1206 06:16:42.761307 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c805cd51-5aaa-4c78-b98e-29a5a6ae44a9-combined-ca-bundle\") pod \"keystone-bootstrap-tc5gt\" (UID: \"c805cd51-5aaa-4c78-b98e-29a5a6ae44a9\") " pod="openstack/keystone-bootstrap-tc5gt" Dec 06 06:16:42 crc kubenswrapper[4809]: I1206 06:16:42.761330 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpsfs\" (UniqueName: \"kubernetes.io/projected/c805cd51-5aaa-4c78-b98e-29a5a6ae44a9-kube-api-access-vpsfs\") pod \"keystone-bootstrap-tc5gt\" (UID: \"c805cd51-5aaa-4c78-b98e-29a5a6ae44a9\") " pod="openstack/keystone-bootstrap-tc5gt" Dec 06 06:16:42 crc kubenswrapper[4809]: I1206 06:16:42.761357 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c805cd51-5aaa-4c78-b98e-29a5a6ae44a9-credential-keys\") pod \"keystone-bootstrap-tc5gt\" (UID: \"c805cd51-5aaa-4c78-b98e-29a5a6ae44a9\") " pod="openstack/keystone-bootstrap-tc5gt" Dec 06 06:16:42 crc kubenswrapper[4809]: I1206 06:16:42.761447 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c805cd51-5aaa-4c78-b98e-29a5a6ae44a9-config-data\") pod \"keystone-bootstrap-tc5gt\" (UID: \"c805cd51-5aaa-4c78-b98e-29a5a6ae44a9\") " pod="openstack/keystone-bootstrap-tc5gt" Dec 06 06:16:42 crc kubenswrapper[4809]: I1206 06:16:42.767355 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c805cd51-5aaa-4c78-b98e-29a5a6ae44a9-fernet-keys\") pod \"keystone-bootstrap-tc5gt\" (UID: \"c805cd51-5aaa-4c78-b98e-29a5a6ae44a9\") " pod="openstack/keystone-bootstrap-tc5gt" Dec 06 06:16:42 crc kubenswrapper[4809]: I1206 06:16:42.768177 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c805cd51-5aaa-4c78-b98e-29a5a6ae44a9-combined-ca-bundle\") pod \"keystone-bootstrap-tc5gt\" (UID: \"c805cd51-5aaa-4c78-b98e-29a5a6ae44a9\") " pod="openstack/keystone-bootstrap-tc5gt" Dec 06 06:16:42 crc kubenswrapper[4809]: I1206 06:16:42.768157 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c805cd51-5aaa-4c78-b98e-29a5a6ae44a9-scripts\") pod \"keystone-bootstrap-tc5gt\" (UID: \"c805cd51-5aaa-4c78-b98e-29a5a6ae44a9\") " pod="openstack/keystone-bootstrap-tc5gt" Dec 06 06:16:42 crc kubenswrapper[4809]: I1206 06:16:42.768383 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c805cd51-5aaa-4c78-b98e-29a5a6ae44a9-credential-keys\") pod \"keystone-bootstrap-tc5gt\" (UID: \"c805cd51-5aaa-4c78-b98e-29a5a6ae44a9\") " pod="openstack/keystone-bootstrap-tc5gt" Dec 06 06:16:42 crc kubenswrapper[4809]: I1206 06:16:42.770801 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c805cd51-5aaa-4c78-b98e-29a5a6ae44a9-config-data\") pod \"keystone-bootstrap-tc5gt\" (UID: \"c805cd51-5aaa-4c78-b98e-29a5a6ae44a9\") " pod="openstack/keystone-bootstrap-tc5gt" Dec 06 06:16:42 crc kubenswrapper[4809]: I1206 06:16:42.786520 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpsfs\" (UniqueName: \"kubernetes.io/projected/c805cd51-5aaa-4c78-b98e-29a5a6ae44a9-kube-api-access-vpsfs\") pod \"keystone-bootstrap-tc5gt\" (UID: \"c805cd51-5aaa-4c78-b98e-29a5a6ae44a9\") " pod="openstack/keystone-bootstrap-tc5gt" Dec 06 06:16:42 crc kubenswrapper[4809]: I1206 06:16:42.899112 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-tc5gt" Dec 06 06:16:43 crc kubenswrapper[4809]: I1206 06:16:43.400524 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f2a4615-0ec1-404a-9ab9-848d8b45be2a" path="/var/lib/kubelet/pods/4f2a4615-0ec1-404a-9ab9-848d8b45be2a/volumes" Dec 06 06:16:43 crc kubenswrapper[4809]: E1206 06:16:43.808391 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Dec 06 06:16:43 crc kubenswrapper[4809]: E1206 06:16:43.808831 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-r96sr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-m98hv_openstack(8336af68-51f9-4f3a-8c92-61dd1aba9067): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 06:16:43 crc kubenswrapper[4809]: E1206 06:16:43.809981 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-m98hv" podUID="8336af68-51f9-4f3a-8c92-61dd1aba9067" Dec 06 06:16:43 crc kubenswrapper[4809]: I1206 06:16:43.924249 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68dcc9cf6f-vjzzr" Dec 06 06:16:43 crc kubenswrapper[4809]: I1206 06:16:43.991606 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ft7cs\" (UniqueName: \"kubernetes.io/projected/46f0989d-3592-45c2-b046-5d33d6fd97b4-kube-api-access-ft7cs\") pod \"46f0989d-3592-45c2-b046-5d33d6fd97b4\" (UID: \"46f0989d-3592-45c2-b046-5d33d6fd97b4\") " Dec 06 06:16:43 crc kubenswrapper[4809]: I1206 06:16:43.991659 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/46f0989d-3592-45c2-b046-5d33d6fd97b4-ovsdbserver-sb\") pod \"46f0989d-3592-45c2-b046-5d33d6fd97b4\" (UID: \"46f0989d-3592-45c2-b046-5d33d6fd97b4\") " Dec 06 06:16:43 crc kubenswrapper[4809]: I1206 06:16:43.991727 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46f0989d-3592-45c2-b046-5d33d6fd97b4-config\") pod \"46f0989d-3592-45c2-b046-5d33d6fd97b4\" (UID: \"46f0989d-3592-45c2-b046-5d33d6fd97b4\") " Dec 06 06:16:43 crc kubenswrapper[4809]: I1206 06:16:43.991788 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/46f0989d-3592-45c2-b046-5d33d6fd97b4-dns-svc\") pod \"46f0989d-3592-45c2-b046-5d33d6fd97b4\" (UID: \"46f0989d-3592-45c2-b046-5d33d6fd97b4\") " Dec 06 06:16:43 crc kubenswrapper[4809]: I1206 06:16:43.991892 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/46f0989d-3592-45c2-b046-5d33d6fd97b4-ovsdbserver-nb\") pod \"46f0989d-3592-45c2-b046-5d33d6fd97b4\" (UID: \"46f0989d-3592-45c2-b046-5d33d6fd97b4\") " Dec 06 06:16:43 crc kubenswrapper[4809]: I1206 06:16:43.996670 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46f0989d-3592-45c2-b046-5d33d6fd97b4-kube-api-access-ft7cs" (OuterVolumeSpecName: "kube-api-access-ft7cs") pod "46f0989d-3592-45c2-b046-5d33d6fd97b4" (UID: "46f0989d-3592-45c2-b046-5d33d6fd97b4"). InnerVolumeSpecName "kube-api-access-ft7cs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:16:44 crc kubenswrapper[4809]: I1206 06:16:44.055237 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46f0989d-3592-45c2-b046-5d33d6fd97b4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "46f0989d-3592-45c2-b046-5d33d6fd97b4" (UID: "46f0989d-3592-45c2-b046-5d33d6fd97b4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:16:44 crc kubenswrapper[4809]: I1206 06:16:44.060401 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46f0989d-3592-45c2-b046-5d33d6fd97b4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "46f0989d-3592-45c2-b046-5d33d6fd97b4" (UID: "46f0989d-3592-45c2-b046-5d33d6fd97b4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:16:44 crc kubenswrapper[4809]: I1206 06:16:44.064786 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46f0989d-3592-45c2-b046-5d33d6fd97b4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "46f0989d-3592-45c2-b046-5d33d6fd97b4" (UID: "46f0989d-3592-45c2-b046-5d33d6fd97b4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:16:44 crc kubenswrapper[4809]: I1206 06:16:44.065599 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46f0989d-3592-45c2-b046-5d33d6fd97b4-config" (OuterVolumeSpecName: "config") pod "46f0989d-3592-45c2-b046-5d33d6fd97b4" (UID: "46f0989d-3592-45c2-b046-5d33d6fd97b4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:16:44 crc kubenswrapper[4809]: I1206 06:16:44.095044 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/46f0989d-3592-45c2-b046-5d33d6fd97b4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:44 crc kubenswrapper[4809]: I1206 06:16:44.095080 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ft7cs\" (UniqueName: \"kubernetes.io/projected/46f0989d-3592-45c2-b046-5d33d6fd97b4-kube-api-access-ft7cs\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:44 crc kubenswrapper[4809]: I1206 06:16:44.095092 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/46f0989d-3592-45c2-b046-5d33d6fd97b4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:44 crc kubenswrapper[4809]: I1206 06:16:44.095102 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46f0989d-3592-45c2-b046-5d33d6fd97b4-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:44 crc kubenswrapper[4809]: I1206 06:16:44.095113 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/46f0989d-3592-45c2-b046-5d33d6fd97b4-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:44 crc kubenswrapper[4809]: E1206 06:16:44.286814 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified" Dec 06 06:16:44 crc kubenswrapper[4809]: E1206 06:16:44.286995 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zkz67,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-ckhg9_openstack(edf1ab66-b483-4172-81fd-dd3f3b9b44aa): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 06:16:44 crc kubenswrapper[4809]: E1206 06:16:44.288202 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/heat-db-sync-ckhg9" podUID="edf1ab66-b483-4172-81fd-dd3f3b9b44aa" Dec 06 06:16:44 crc kubenswrapper[4809]: I1206 06:16:44.373127 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68dcc9cf6f-vjzzr" event={"ID":"46f0989d-3592-45c2-b046-5d33d6fd97b4","Type":"ContainerDied","Data":"f952a937b415d1cd00aea947eddd935597ec0f67c8c691066e2a9ac40105151c"} Dec 06 06:16:44 crc kubenswrapper[4809]: I1206 06:16:44.373460 4809 scope.go:117] "RemoveContainer" containerID="33eab5c70cd4af9a2c988a431a752ead40b0627ed370c368fbbdc0f17aef3e78" Dec 06 06:16:44 crc kubenswrapper[4809]: I1206 06:16:44.373589 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68dcc9cf6f-vjzzr" Dec 06 06:16:44 crc kubenswrapper[4809]: E1206 06:16:44.376120 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-m98hv" podUID="8336af68-51f9-4f3a-8c92-61dd1aba9067" Dec 06 06:16:44 crc kubenswrapper[4809]: E1206 06:16:44.378979 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified\\\"\"" pod="openstack/heat-db-sync-ckhg9" podUID="edf1ab66-b483-4172-81fd-dd3f3b9b44aa" Dec 06 06:16:44 crc kubenswrapper[4809]: I1206 06:16:44.436850 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68dcc9cf6f-vjzzr"] Dec 06 06:16:44 crc kubenswrapper[4809]: I1206 06:16:44.451836 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-68dcc9cf6f-vjzzr"] Dec 06 06:16:45 crc kubenswrapper[4809]: I1206 06:16:45.429800 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46f0989d-3592-45c2-b046-5d33d6fd97b4" path="/var/lib/kubelet/pods/46f0989d-3592-45c2-b046-5d33d6fd97b4/volumes" Dec 06 06:16:46 crc kubenswrapper[4809]: I1206 06:16:46.280889 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-68dcc9cf6f-vjzzr" podUID="46f0989d-3592-45c2-b046-5d33d6fd97b4" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.179:5353: i/o timeout" Dec 06 06:16:48 crc kubenswrapper[4809]: I1206 06:16:48.427951 4809 generic.go:334] "Generic (PLEG): container finished" podID="413a30cf-4b57-44d7-b9bf-3d400bf4b897" containerID="5fa4ff9a4f10e7d52d4499104cc17b8069a880b80c0127d3a607c15b16f88cf1" exitCode=0 Dec 06 06:16:48 crc kubenswrapper[4809]: I1206 06:16:48.428032 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-mc9z8" event={"ID":"413a30cf-4b57-44d7-b9bf-3d400bf4b897","Type":"ContainerDied","Data":"5fa4ff9a4f10e7d52d4499104cc17b8069a880b80c0127d3a607c15b16f88cf1"} Dec 06 06:16:52 crc kubenswrapper[4809]: I1206 06:16:52.596052 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-mc9z8" Dec 06 06:16:52 crc kubenswrapper[4809]: I1206 06:16:52.683164 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/413a30cf-4b57-44d7-b9bf-3d400bf4b897-db-sync-config-data\") pod \"413a30cf-4b57-44d7-b9bf-3d400bf4b897\" (UID: \"413a30cf-4b57-44d7-b9bf-3d400bf4b897\") " Dec 06 06:16:52 crc kubenswrapper[4809]: I1206 06:16:52.683344 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-75lxt\" (UniqueName: \"kubernetes.io/projected/413a30cf-4b57-44d7-b9bf-3d400bf4b897-kube-api-access-75lxt\") pod \"413a30cf-4b57-44d7-b9bf-3d400bf4b897\" (UID: \"413a30cf-4b57-44d7-b9bf-3d400bf4b897\") " Dec 06 06:16:52 crc kubenswrapper[4809]: I1206 06:16:52.683381 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/413a30cf-4b57-44d7-b9bf-3d400bf4b897-config-data\") pod \"413a30cf-4b57-44d7-b9bf-3d400bf4b897\" (UID: \"413a30cf-4b57-44d7-b9bf-3d400bf4b897\") " Dec 06 06:16:52 crc kubenswrapper[4809]: I1206 06:16:52.683402 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/413a30cf-4b57-44d7-b9bf-3d400bf4b897-combined-ca-bundle\") pod \"413a30cf-4b57-44d7-b9bf-3d400bf4b897\" (UID: \"413a30cf-4b57-44d7-b9bf-3d400bf4b897\") " Dec 06 06:16:52 crc kubenswrapper[4809]: I1206 06:16:52.699153 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/413a30cf-4b57-44d7-b9bf-3d400bf4b897-kube-api-access-75lxt" (OuterVolumeSpecName: "kube-api-access-75lxt") pod "413a30cf-4b57-44d7-b9bf-3d400bf4b897" (UID: "413a30cf-4b57-44d7-b9bf-3d400bf4b897"). InnerVolumeSpecName "kube-api-access-75lxt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:16:52 crc kubenswrapper[4809]: I1206 06:16:52.731958 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/413a30cf-4b57-44d7-b9bf-3d400bf4b897-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "413a30cf-4b57-44d7-b9bf-3d400bf4b897" (UID: "413a30cf-4b57-44d7-b9bf-3d400bf4b897"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:16:52 crc kubenswrapper[4809]: I1206 06:16:52.735455 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/413a30cf-4b57-44d7-b9bf-3d400bf4b897-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "413a30cf-4b57-44d7-b9bf-3d400bf4b897" (UID: "413a30cf-4b57-44d7-b9bf-3d400bf4b897"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:16:52 crc kubenswrapper[4809]: I1206 06:16:52.771958 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/413a30cf-4b57-44d7-b9bf-3d400bf4b897-config-data" (OuterVolumeSpecName: "config-data") pod "413a30cf-4b57-44d7-b9bf-3d400bf4b897" (UID: "413a30cf-4b57-44d7-b9bf-3d400bf4b897"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:16:52 crc kubenswrapper[4809]: I1206 06:16:52.785617 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-75lxt\" (UniqueName: \"kubernetes.io/projected/413a30cf-4b57-44d7-b9bf-3d400bf4b897-kube-api-access-75lxt\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:52 crc kubenswrapper[4809]: I1206 06:16:52.785880 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/413a30cf-4b57-44d7-b9bf-3d400bf4b897-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:52 crc kubenswrapper[4809]: I1206 06:16:52.785972 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/413a30cf-4b57-44d7-b9bf-3d400bf4b897-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:52 crc kubenswrapper[4809]: I1206 06:16:52.786055 4809 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/413a30cf-4b57-44d7-b9bf-3d400bf4b897-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:53 crc kubenswrapper[4809]: I1206 06:16:53.480957 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-mc9z8" event={"ID":"413a30cf-4b57-44d7-b9bf-3d400bf4b897","Type":"ContainerDied","Data":"b9cf3c39badf7c33ac4e9c51dbb256bdb00b7f12f045eb32a6b04feef00fdd38"} Dec 06 06:16:53 crc kubenswrapper[4809]: I1206 06:16:53.481307 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b9cf3c39badf7c33ac4e9c51dbb256bdb00b7f12f045eb32a6b04feef00fdd38" Dec 06 06:16:53 crc kubenswrapper[4809]: I1206 06:16:53.481102 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-mc9z8" Dec 06 06:16:54 crc kubenswrapper[4809]: I1206 06:16:54.344512 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-b75nx"] Dec 06 06:16:54 crc kubenswrapper[4809]: I1206 06:16:54.377990 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-wxlvd"] Dec 06 06:16:54 crc kubenswrapper[4809]: E1206 06:16:54.378657 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="413a30cf-4b57-44d7-b9bf-3d400bf4b897" containerName="glance-db-sync" Dec 06 06:16:54 crc kubenswrapper[4809]: I1206 06:16:54.378811 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="413a30cf-4b57-44d7-b9bf-3d400bf4b897" containerName="glance-db-sync" Dec 06 06:16:54 crc kubenswrapper[4809]: E1206 06:16:54.378951 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46f0989d-3592-45c2-b046-5d33d6fd97b4" containerName="init" Dec 06 06:16:54 crc kubenswrapper[4809]: I1206 06:16:54.379045 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="46f0989d-3592-45c2-b046-5d33d6fd97b4" containerName="init" Dec 06 06:16:54 crc kubenswrapper[4809]: E1206 06:16:54.379160 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46f0989d-3592-45c2-b046-5d33d6fd97b4" containerName="dnsmasq-dns" Dec 06 06:16:54 crc kubenswrapper[4809]: I1206 06:16:54.379225 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="46f0989d-3592-45c2-b046-5d33d6fd97b4" containerName="dnsmasq-dns" Dec 06 06:16:54 crc kubenswrapper[4809]: I1206 06:16:54.379484 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="413a30cf-4b57-44d7-b9bf-3d400bf4b897" containerName="glance-db-sync" Dec 06 06:16:54 crc kubenswrapper[4809]: I1206 06:16:54.379546 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="46f0989d-3592-45c2-b046-5d33d6fd97b4" containerName="dnsmasq-dns" Dec 06 06:16:54 crc kubenswrapper[4809]: I1206 06:16:54.380726 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-wxlvd" Dec 06 06:16:54 crc kubenswrapper[4809]: I1206 06:16:54.408364 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-wxlvd"] Dec 06 06:16:54 crc kubenswrapper[4809]: I1206 06:16:54.449407 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b6e30c4b-8103-4399-b434-58bedae43fc5-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-wxlvd\" (UID: \"b6e30c4b-8103-4399-b434-58bedae43fc5\") " pod="openstack/dnsmasq-dns-785d8bcb8c-wxlvd" Dec 06 06:16:54 crc kubenswrapper[4809]: I1206 06:16:54.449856 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b6e30c4b-8103-4399-b434-58bedae43fc5-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-wxlvd\" (UID: \"b6e30c4b-8103-4399-b434-58bedae43fc5\") " pod="openstack/dnsmasq-dns-785d8bcb8c-wxlvd" Dec 06 06:16:54 crc kubenswrapper[4809]: I1206 06:16:54.449878 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-698pg\" (UniqueName: \"kubernetes.io/projected/b6e30c4b-8103-4399-b434-58bedae43fc5-kube-api-access-698pg\") pod \"dnsmasq-dns-785d8bcb8c-wxlvd\" (UID: \"b6e30c4b-8103-4399-b434-58bedae43fc5\") " pod="openstack/dnsmasq-dns-785d8bcb8c-wxlvd" Dec 06 06:16:54 crc kubenswrapper[4809]: I1206 06:16:54.449924 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b6e30c4b-8103-4399-b434-58bedae43fc5-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-wxlvd\" (UID: \"b6e30c4b-8103-4399-b434-58bedae43fc5\") " pod="openstack/dnsmasq-dns-785d8bcb8c-wxlvd" Dec 06 06:16:54 crc kubenswrapper[4809]: I1206 06:16:54.449993 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b6e30c4b-8103-4399-b434-58bedae43fc5-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-wxlvd\" (UID: \"b6e30c4b-8103-4399-b434-58bedae43fc5\") " pod="openstack/dnsmasq-dns-785d8bcb8c-wxlvd" Dec 06 06:16:54 crc kubenswrapper[4809]: I1206 06:16:54.450026 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6e30c4b-8103-4399-b434-58bedae43fc5-config\") pod \"dnsmasq-dns-785d8bcb8c-wxlvd\" (UID: \"b6e30c4b-8103-4399-b434-58bedae43fc5\") " pod="openstack/dnsmasq-dns-785d8bcb8c-wxlvd" Dec 06 06:16:54 crc kubenswrapper[4809]: I1206 06:16:54.552029 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b6e30c4b-8103-4399-b434-58bedae43fc5-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-wxlvd\" (UID: \"b6e30c4b-8103-4399-b434-58bedae43fc5\") " pod="openstack/dnsmasq-dns-785d8bcb8c-wxlvd" Dec 06 06:16:54 crc kubenswrapper[4809]: I1206 06:16:54.552454 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6e30c4b-8103-4399-b434-58bedae43fc5-config\") pod \"dnsmasq-dns-785d8bcb8c-wxlvd\" (UID: \"b6e30c4b-8103-4399-b434-58bedae43fc5\") " pod="openstack/dnsmasq-dns-785d8bcb8c-wxlvd" Dec 06 06:16:54 crc kubenswrapper[4809]: I1206 06:16:54.552496 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b6e30c4b-8103-4399-b434-58bedae43fc5-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-wxlvd\" (UID: \"b6e30c4b-8103-4399-b434-58bedae43fc5\") " pod="openstack/dnsmasq-dns-785d8bcb8c-wxlvd" Dec 06 06:16:54 crc kubenswrapper[4809]: I1206 06:16:54.552662 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b6e30c4b-8103-4399-b434-58bedae43fc5-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-wxlvd\" (UID: \"b6e30c4b-8103-4399-b434-58bedae43fc5\") " pod="openstack/dnsmasq-dns-785d8bcb8c-wxlvd" Dec 06 06:16:54 crc kubenswrapper[4809]: I1206 06:16:54.552687 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-698pg\" (UniqueName: \"kubernetes.io/projected/b6e30c4b-8103-4399-b434-58bedae43fc5-kube-api-access-698pg\") pod \"dnsmasq-dns-785d8bcb8c-wxlvd\" (UID: \"b6e30c4b-8103-4399-b434-58bedae43fc5\") " pod="openstack/dnsmasq-dns-785d8bcb8c-wxlvd" Dec 06 06:16:54 crc kubenswrapper[4809]: I1206 06:16:54.552738 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b6e30c4b-8103-4399-b434-58bedae43fc5-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-wxlvd\" (UID: \"b6e30c4b-8103-4399-b434-58bedae43fc5\") " pod="openstack/dnsmasq-dns-785d8bcb8c-wxlvd" Dec 06 06:16:54 crc kubenswrapper[4809]: I1206 06:16:54.553857 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b6e30c4b-8103-4399-b434-58bedae43fc5-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-wxlvd\" (UID: \"b6e30c4b-8103-4399-b434-58bedae43fc5\") " pod="openstack/dnsmasq-dns-785d8bcb8c-wxlvd" Dec 06 06:16:54 crc kubenswrapper[4809]: I1206 06:16:54.554878 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b6e30c4b-8103-4399-b434-58bedae43fc5-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-wxlvd\" (UID: \"b6e30c4b-8103-4399-b434-58bedae43fc5\") " pod="openstack/dnsmasq-dns-785d8bcb8c-wxlvd" Dec 06 06:16:54 crc kubenswrapper[4809]: I1206 06:16:54.556447 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b6e30c4b-8103-4399-b434-58bedae43fc5-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-wxlvd\" (UID: \"b6e30c4b-8103-4399-b434-58bedae43fc5\") " pod="openstack/dnsmasq-dns-785d8bcb8c-wxlvd" Dec 06 06:16:54 crc kubenswrapper[4809]: I1206 06:16:54.557382 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6e30c4b-8103-4399-b434-58bedae43fc5-config\") pod \"dnsmasq-dns-785d8bcb8c-wxlvd\" (UID: \"b6e30c4b-8103-4399-b434-58bedae43fc5\") " pod="openstack/dnsmasq-dns-785d8bcb8c-wxlvd" Dec 06 06:16:54 crc kubenswrapper[4809]: I1206 06:16:54.557529 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b6e30c4b-8103-4399-b434-58bedae43fc5-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-wxlvd\" (UID: \"b6e30c4b-8103-4399-b434-58bedae43fc5\") " pod="openstack/dnsmasq-dns-785d8bcb8c-wxlvd" Dec 06 06:16:54 crc kubenswrapper[4809]: I1206 06:16:54.579662 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-698pg\" (UniqueName: \"kubernetes.io/projected/b6e30c4b-8103-4399-b434-58bedae43fc5-kube-api-access-698pg\") pod \"dnsmasq-dns-785d8bcb8c-wxlvd\" (UID: \"b6e30c4b-8103-4399-b434-58bedae43fc5\") " pod="openstack/dnsmasq-dns-785d8bcb8c-wxlvd" Dec 06 06:16:54 crc kubenswrapper[4809]: I1206 06:16:54.732885 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-wxlvd" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.486698 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.488852 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.504003 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.504531 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.504695 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-rzj8j" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.522352 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.560457 4809 generic.go:334] "Generic (PLEG): container finished" podID="08195221-42d1-4ce3-93a6-050e44339013" containerID="7c24eacb26efb396e941d2db0e9a7e55a9a496af0f47192b9d1f6a49c739e882" exitCode=0 Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.560494 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-hsm9x" event={"ID":"08195221-42d1-4ce3-93a6-050e44339013","Type":"ContainerDied","Data":"7c24eacb26efb396e941d2db0e9a7e55a9a496af0f47192b9d1f6a49c739e882"} Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.564777 4809 scope.go:117] "RemoveContainer" containerID="f1127a18075a3c31522ec64330d3d64798b254aeca584a9632c89a7ab71af4cb" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.577070 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df904bdc-a504-4deb-8dd2-d2acf7626200-config-data\") pod \"glance-default-external-api-0\" (UID: \"df904bdc-a504-4deb-8dd2-d2acf7626200\") " pod="openstack/glance-default-external-api-0" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.577381 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df904bdc-a504-4deb-8dd2-d2acf7626200-logs\") pod \"glance-default-external-api-0\" (UID: \"df904bdc-a504-4deb-8dd2-d2acf7626200\") " pod="openstack/glance-default-external-api-0" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.577593 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df904bdc-a504-4deb-8dd2-d2acf7626200-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"df904bdc-a504-4deb-8dd2-d2acf7626200\") " pod="openstack/glance-default-external-api-0" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.577806 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g42m7\" (UniqueName: \"kubernetes.io/projected/df904bdc-a504-4deb-8dd2-d2acf7626200-kube-api-access-g42m7\") pod \"glance-default-external-api-0\" (UID: \"df904bdc-a504-4deb-8dd2-d2acf7626200\") " pod="openstack/glance-default-external-api-0" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.578059 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"df904bdc-a504-4deb-8dd2-d2acf7626200\") " pod="openstack/glance-default-external-api-0" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.578287 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/df904bdc-a504-4deb-8dd2-d2acf7626200-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"df904bdc-a504-4deb-8dd2-d2acf7626200\") " pod="openstack/glance-default-external-api-0" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.578392 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df904bdc-a504-4deb-8dd2-d2acf7626200-scripts\") pod \"glance-default-external-api-0\" (UID: \"df904bdc-a504-4deb-8dd2-d2acf7626200\") " pod="openstack/glance-default-external-api-0" Dec 06 06:16:55 crc kubenswrapper[4809]: E1206 06:16:55.585579 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 06 06:16:55 crc kubenswrapper[4809]: E1206 06:16:55.586507 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wxm74,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-xs2c8_openstack(c75cf833-5179-4e18-8f74-3faed72757ed): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 06:16:55 crc kubenswrapper[4809]: E1206 06:16:55.587968 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-xs2c8" podUID="c75cf833-5179-4e18-8f74-3faed72757ed" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.645830 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.648448 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.651386 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.681797 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df904bdc-a504-4deb-8dd2-d2acf7626200-config-data\") pod \"glance-default-external-api-0\" (UID: \"df904bdc-a504-4deb-8dd2-d2acf7626200\") " pod="openstack/glance-default-external-api-0" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.681953 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df904bdc-a504-4deb-8dd2-d2acf7626200-logs\") pod \"glance-default-external-api-0\" (UID: \"df904bdc-a504-4deb-8dd2-d2acf7626200\") " pod="openstack/glance-default-external-api-0" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.682040 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df904bdc-a504-4deb-8dd2-d2acf7626200-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"df904bdc-a504-4deb-8dd2-d2acf7626200\") " pod="openstack/glance-default-external-api-0" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.682139 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g42m7\" (UniqueName: \"kubernetes.io/projected/df904bdc-a504-4deb-8dd2-d2acf7626200-kube-api-access-g42m7\") pod \"glance-default-external-api-0\" (UID: \"df904bdc-a504-4deb-8dd2-d2acf7626200\") " pod="openstack/glance-default-external-api-0" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.682220 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"df904bdc-a504-4deb-8dd2-d2acf7626200\") " pod="openstack/glance-default-external-api-0" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.682357 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/df904bdc-a504-4deb-8dd2-d2acf7626200-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"df904bdc-a504-4deb-8dd2-d2acf7626200\") " pod="openstack/glance-default-external-api-0" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.682389 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df904bdc-a504-4deb-8dd2-d2acf7626200-scripts\") pod \"glance-default-external-api-0\" (UID: \"df904bdc-a504-4deb-8dd2-d2acf7626200\") " pod="openstack/glance-default-external-api-0" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.683073 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.683099 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df904bdc-a504-4deb-8dd2-d2acf7626200-logs\") pod \"glance-default-external-api-0\" (UID: \"df904bdc-a504-4deb-8dd2-d2acf7626200\") " pod="openstack/glance-default-external-api-0" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.683319 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"df904bdc-a504-4deb-8dd2-d2acf7626200\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-external-api-0" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.684041 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/df904bdc-a504-4deb-8dd2-d2acf7626200-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"df904bdc-a504-4deb-8dd2-d2acf7626200\") " pod="openstack/glance-default-external-api-0" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.687028 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df904bdc-a504-4deb-8dd2-d2acf7626200-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"df904bdc-a504-4deb-8dd2-d2acf7626200\") " pod="openstack/glance-default-external-api-0" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.687800 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df904bdc-a504-4deb-8dd2-d2acf7626200-scripts\") pod \"glance-default-external-api-0\" (UID: \"df904bdc-a504-4deb-8dd2-d2acf7626200\") " pod="openstack/glance-default-external-api-0" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.699558 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df904bdc-a504-4deb-8dd2-d2acf7626200-config-data\") pod \"glance-default-external-api-0\" (UID: \"df904bdc-a504-4deb-8dd2-d2acf7626200\") " pod="openstack/glance-default-external-api-0" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.738908 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g42m7\" (UniqueName: \"kubernetes.io/projected/df904bdc-a504-4deb-8dd2-d2acf7626200-kube-api-access-g42m7\") pod \"glance-default-external-api-0\" (UID: \"df904bdc-a504-4deb-8dd2-d2acf7626200\") " pod="openstack/glance-default-external-api-0" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.790790 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b6f4174-473c-4e12-bcbb-7dd6ab7633ee-logs\") pod \"glance-default-internal-api-0\" (UID: \"5b6f4174-473c-4e12-bcbb-7dd6ab7633ee\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.790841 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6gh5\" (UniqueName: \"kubernetes.io/projected/5b6f4174-473c-4e12-bcbb-7dd6ab7633ee-kube-api-access-v6gh5\") pod \"glance-default-internal-api-0\" (UID: \"5b6f4174-473c-4e12-bcbb-7dd6ab7633ee\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.792858 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b6f4174-473c-4e12-bcbb-7dd6ab7633ee-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5b6f4174-473c-4e12-bcbb-7dd6ab7633ee\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.793013 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"5b6f4174-473c-4e12-bcbb-7dd6ab7633ee\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.793048 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b6f4174-473c-4e12-bcbb-7dd6ab7633ee-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5b6f4174-473c-4e12-bcbb-7dd6ab7633ee\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.793234 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b6f4174-473c-4e12-bcbb-7dd6ab7633ee-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5b6f4174-473c-4e12-bcbb-7dd6ab7633ee\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.793418 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5b6f4174-473c-4e12-bcbb-7dd6ab7633ee-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5b6f4174-473c-4e12-bcbb-7dd6ab7633ee\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.803745 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"df904bdc-a504-4deb-8dd2-d2acf7626200\") " pod="openstack/glance-default-external-api-0" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.895125 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b6f4174-473c-4e12-bcbb-7dd6ab7633ee-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5b6f4174-473c-4e12-bcbb-7dd6ab7633ee\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.897435 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"5b6f4174-473c-4e12-bcbb-7dd6ab7633ee\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.901391 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"5b6f4174-473c-4e12-bcbb-7dd6ab7633ee\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-internal-api-0" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.902417 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b6f4174-473c-4e12-bcbb-7dd6ab7633ee-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5b6f4174-473c-4e12-bcbb-7dd6ab7633ee\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.921151 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b6f4174-473c-4e12-bcbb-7dd6ab7633ee-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5b6f4174-473c-4e12-bcbb-7dd6ab7633ee\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.928234 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b6f4174-473c-4e12-bcbb-7dd6ab7633ee-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5b6f4174-473c-4e12-bcbb-7dd6ab7633ee\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.932381 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5b6f4174-473c-4e12-bcbb-7dd6ab7633ee-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5b6f4174-473c-4e12-bcbb-7dd6ab7633ee\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.932763 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b6f4174-473c-4e12-bcbb-7dd6ab7633ee-logs\") pod \"glance-default-internal-api-0\" (UID: \"5b6f4174-473c-4e12-bcbb-7dd6ab7633ee\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.932872 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5b6f4174-473c-4e12-bcbb-7dd6ab7633ee-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5b6f4174-473c-4e12-bcbb-7dd6ab7633ee\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.933093 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b6f4174-473c-4e12-bcbb-7dd6ab7633ee-logs\") pod \"glance-default-internal-api-0\" (UID: \"5b6f4174-473c-4e12-bcbb-7dd6ab7633ee\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.933098 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6gh5\" (UniqueName: \"kubernetes.io/projected/5b6f4174-473c-4e12-bcbb-7dd6ab7633ee-kube-api-access-v6gh5\") pod \"glance-default-internal-api-0\" (UID: \"5b6f4174-473c-4e12-bcbb-7dd6ab7633ee\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.933867 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b6f4174-473c-4e12-bcbb-7dd6ab7633ee-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5b6f4174-473c-4e12-bcbb-7dd6ab7633ee\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.934900 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b6f4174-473c-4e12-bcbb-7dd6ab7633ee-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5b6f4174-473c-4e12-bcbb-7dd6ab7633ee\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.955802 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6gh5\" (UniqueName: \"kubernetes.io/projected/5b6f4174-473c-4e12-bcbb-7dd6ab7633ee-kube-api-access-v6gh5\") pod \"glance-default-internal-api-0\" (UID: \"5b6f4174-473c-4e12-bcbb-7dd6ab7633ee\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.979669 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"5b6f4174-473c-4e12-bcbb-7dd6ab7633ee\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.986867 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 06:16:55 crc kubenswrapper[4809]: I1206 06:16:55.991615 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 06:16:56 crc kubenswrapper[4809]: I1206 06:16:56.171620 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-b75nx"] Dec 06 06:16:56 crc kubenswrapper[4809]: I1206 06:16:56.311692 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-wxlvd"] Dec 06 06:16:56 crc kubenswrapper[4809]: W1206 06:16:56.485103 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc805cd51_5aaa_4c78_b98e_29a5a6ae44a9.slice/crio-a49b8a15dc5a86f3fddb084e2c1f54527c729dc39ff07863f28dbb445672ecec WatchSource:0}: Error finding container a49b8a15dc5a86f3fddb084e2c1f54527c729dc39ff07863f28dbb445672ecec: Status 404 returned error can't find the container with id a49b8a15dc5a86f3fddb084e2c1f54527c729dc39ff07863f28dbb445672ecec Dec 06 06:16:56 crc kubenswrapper[4809]: I1206 06:16:56.503051 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-tc5gt"] Dec 06 06:16:56 crc kubenswrapper[4809]: I1206 06:16:56.669235 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-tc5gt" event={"ID":"c805cd51-5aaa-4c78-b98e-29a5a6ae44a9","Type":"ContainerStarted","Data":"a49b8a15dc5a86f3fddb084e2c1f54527c729dc39ff07863f28dbb445672ecec"} Dec 06 06:16:56 crc kubenswrapper[4809]: I1206 06:16:56.673103 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-wxlvd" event={"ID":"b6e30c4b-8103-4399-b434-58bedae43fc5","Type":"ContainerStarted","Data":"a3d0de81ce3713c8508abf317d9721e0b720ace598ca136ca7e0f7781bcc148a"} Dec 06 06:16:56 crc kubenswrapper[4809]: I1206 06:16:56.682415 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"93e2d67f-9481-4ca3-83e7-4f7c0673b9d4","Type":"ContainerStarted","Data":"081257dd2f4b335328e16317962b5d8e7e6fec0da5505d2793bea03badf12dd1"} Dec 06 06:16:56 crc kubenswrapper[4809]: I1206 06:16:56.706892 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ca8e11eb-2da1-4706-b589-650bcb817d4c","Type":"ContainerStarted","Data":"1615c86b9ba078c80109dcd28e5c0d84d16120125094137b7e03581672e707be"} Dec 06 06:16:56 crc kubenswrapper[4809]: I1206 06:16:56.734306 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-b75nx" event={"ID":"40abc93a-e45f-44cb-8fb1-f14961fe3e74","Type":"ContainerStarted","Data":"4e3d67c3955338a59473286e274b6b5de4272bef4db2488748449eb617c9af15"} Dec 06 06:16:56 crc kubenswrapper[4809]: E1206 06:16:56.740897 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-xs2c8" podUID="c75cf833-5179-4e18-8f74-3faed72757ed" Dec 06 06:16:56 crc kubenswrapper[4809]: I1206 06:16:56.775886 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=44.775863272 podStartE2EDuration="44.775863272s" podCreationTimestamp="2025-12-06 06:16:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:16:56.746734225 +0000 UTC m=+1541.635717167" watchObservedRunningTime="2025-12-06 06:16:56.775863272 +0000 UTC m=+1541.664846214" Dec 06 06:16:56 crc kubenswrapper[4809]: I1206 06:16:56.849416 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 06:16:56 crc kubenswrapper[4809]: W1206 06:16:56.904741 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddf904bdc_a504_4deb_8dd2_d2acf7626200.slice/crio-b26c1ce80fa736673bf48ee790d716702359b3240f0ea5cef6a82299ac840a37 WatchSource:0}: Error finding container b26c1ce80fa736673bf48ee790d716702359b3240f0ea5cef6a82299ac840a37: Status 404 returned error can't find the container with id b26c1ce80fa736673bf48ee790d716702359b3240f0ea5cef6a82299ac840a37 Dec 06 06:16:56 crc kubenswrapper[4809]: I1206 06:16:56.918778 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 06:16:56 crc kubenswrapper[4809]: W1206 06:16:56.933695 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b6f4174_473c_4e12_bcbb_7dd6ab7633ee.slice/crio-5dda841c122fa4d761f1a9b61cd62e6392c4aa39121faab6151c5d97cfdd2eb2 WatchSource:0}: Error finding container 5dda841c122fa4d761f1a9b61cd62e6392c4aa39121faab6151c5d97cfdd2eb2: Status 404 returned error can't find the container with id 5dda841c122fa4d761f1a9b61cd62e6392c4aa39121faab6151c5d97cfdd2eb2 Dec 06 06:16:57 crc kubenswrapper[4809]: I1206 06:16:57.228486 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-hsm9x" Dec 06 06:16:57 crc kubenswrapper[4809]: I1206 06:16:57.285069 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08195221-42d1-4ce3-93a6-050e44339013-combined-ca-bundle\") pod \"08195221-42d1-4ce3-93a6-050e44339013\" (UID: \"08195221-42d1-4ce3-93a6-050e44339013\") " Dec 06 06:16:57 crc kubenswrapper[4809]: I1206 06:16:57.287257 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/08195221-42d1-4ce3-93a6-050e44339013-config\") pod \"08195221-42d1-4ce3-93a6-050e44339013\" (UID: \"08195221-42d1-4ce3-93a6-050e44339013\") " Dec 06 06:16:57 crc kubenswrapper[4809]: I1206 06:16:57.287501 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-64xgk\" (UniqueName: \"kubernetes.io/projected/08195221-42d1-4ce3-93a6-050e44339013-kube-api-access-64xgk\") pod \"08195221-42d1-4ce3-93a6-050e44339013\" (UID: \"08195221-42d1-4ce3-93a6-050e44339013\") " Dec 06 06:16:57 crc kubenswrapper[4809]: I1206 06:16:57.296351 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08195221-42d1-4ce3-93a6-050e44339013-kube-api-access-64xgk" (OuterVolumeSpecName: "kube-api-access-64xgk") pod "08195221-42d1-4ce3-93a6-050e44339013" (UID: "08195221-42d1-4ce3-93a6-050e44339013"). InnerVolumeSpecName "kube-api-access-64xgk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:16:57 crc kubenswrapper[4809]: I1206 06:16:57.324295 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08195221-42d1-4ce3-93a6-050e44339013-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "08195221-42d1-4ce3-93a6-050e44339013" (UID: "08195221-42d1-4ce3-93a6-050e44339013"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:16:57 crc kubenswrapper[4809]: I1206 06:16:57.363636 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08195221-42d1-4ce3-93a6-050e44339013-config" (OuterVolumeSpecName: "config") pod "08195221-42d1-4ce3-93a6-050e44339013" (UID: "08195221-42d1-4ce3-93a6-050e44339013"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:16:57 crc kubenswrapper[4809]: I1206 06:16:57.389821 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-64xgk\" (UniqueName: \"kubernetes.io/projected/08195221-42d1-4ce3-93a6-050e44339013-kube-api-access-64xgk\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:57 crc kubenswrapper[4809]: I1206 06:16:57.389863 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08195221-42d1-4ce3-93a6-050e44339013-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:57 crc kubenswrapper[4809]: I1206 06:16:57.389876 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/08195221-42d1-4ce3-93a6-050e44339013-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:57 crc kubenswrapper[4809]: I1206 06:16:57.781951 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-wxlvd"] Dec 06 06:16:57 crc kubenswrapper[4809]: I1206 06:16:57.852428 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-4gxwk" event={"ID":"662e24e4-5dd6-4d18-822c-2863017f15af","Type":"ContainerStarted","Data":"12a159c9d2e59d6ab6c567a3ac994fcd15176cf11ff241ba517b9ef132bec4bc"} Dec 06 06:16:57 crc kubenswrapper[4809]: I1206 06:16:57.853534 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-58svx"] Dec 06 06:16:57 crc kubenswrapper[4809]: E1206 06:16:57.853949 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08195221-42d1-4ce3-93a6-050e44339013" containerName="neutron-db-sync" Dec 06 06:16:57 crc kubenswrapper[4809]: I1206 06:16:57.853962 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="08195221-42d1-4ce3-93a6-050e44339013" containerName="neutron-db-sync" Dec 06 06:16:57 crc kubenswrapper[4809]: I1206 06:16:57.854169 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="08195221-42d1-4ce3-93a6-050e44339013" containerName="neutron-db-sync" Dec 06 06:16:57 crc kubenswrapper[4809]: I1206 06:16:57.855291 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-58svx" Dec 06 06:16:57 crc kubenswrapper[4809]: I1206 06:16:57.858712 4809 generic.go:334] "Generic (PLEG): container finished" podID="40abc93a-e45f-44cb-8fb1-f14961fe3e74" containerID="74db0c60427064c172da346ab23eb30a4c9dae29ee5e6f33e04e8dee7cfde712" exitCode=0 Dec 06 06:16:57 crc kubenswrapper[4809]: I1206 06:16:57.858769 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-b75nx" event={"ID":"40abc93a-e45f-44cb-8fb1-f14961fe3e74","Type":"ContainerDied","Data":"74db0c60427064c172da346ab23eb30a4c9dae29ee5e6f33e04e8dee7cfde712"} Dec 06 06:16:57 crc kubenswrapper[4809]: I1206 06:16:57.885731 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-tc5gt" event={"ID":"c805cd51-5aaa-4c78-b98e-29a5a6ae44a9","Type":"ContainerStarted","Data":"096c45017712c6fe423f137af1e72ea25ccc70231ab4d70c49373ba0b823944c"} Dec 06 06:16:57 crc kubenswrapper[4809]: I1206 06:16:57.895160 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-58svx"] Dec 06 06:16:57 crc kubenswrapper[4809]: I1206 06:16:57.907108 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-4gxwk" podStartSLOduration=2.8927560960000003 podStartE2EDuration="37.907090274s" podCreationTimestamp="2025-12-06 06:16:20 +0000 UTC" firstStartedPulling="2025-12-06 06:16:22.094117268 +0000 UTC m=+1506.983100210" lastFinishedPulling="2025-12-06 06:16:57.108451446 +0000 UTC m=+1541.997434388" observedRunningTime="2025-12-06 06:16:57.874229308 +0000 UTC m=+1542.763212250" watchObservedRunningTime="2025-12-06 06:16:57.907090274 +0000 UTC m=+1542.796073216" Dec 06 06:16:57 crc kubenswrapper[4809]: I1206 06:16:57.916039 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"df904bdc-a504-4deb-8dd2-d2acf7626200","Type":"ContainerStarted","Data":"b26c1ce80fa736673bf48ee790d716702359b3240f0ea5cef6a82299ac840a37"} Dec 06 06:16:57 crc kubenswrapper[4809]: I1206 06:16:57.925564 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5b6f4174-473c-4e12-bcbb-7dd6ab7633ee","Type":"ContainerStarted","Data":"5dda841c122fa4d761f1a9b61cd62e6392c4aa39121faab6151c5d97cfdd2eb2"} Dec 06 06:16:57 crc kubenswrapper[4809]: I1206 06:16:57.929253 4809 generic.go:334] "Generic (PLEG): container finished" podID="b6e30c4b-8103-4399-b434-58bedae43fc5" containerID="87f24a96e341bc224827b16c9293236499ae758fa60cd7ad890fa2e69a045e17" exitCode=0 Dec 06 06:16:57 crc kubenswrapper[4809]: I1206 06:16:57.929561 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-wxlvd" event={"ID":"b6e30c4b-8103-4399-b434-58bedae43fc5","Type":"ContainerDied","Data":"87f24a96e341bc224827b16c9293236499ae758fa60cd7ad890fa2e69a045e17"} Dec 06 06:16:57 crc kubenswrapper[4809]: I1206 06:16:57.947594 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-hsm9x" Dec 06 06:16:57 crc kubenswrapper[4809]: I1206 06:16:57.949199 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-hsm9x" event={"ID":"08195221-42d1-4ce3-93a6-050e44339013","Type":"ContainerDied","Data":"6e1fb3895287a3164ef0b7f918151a1d6b9bd8377505b3589e7842cd1dbc2702"} Dec 06 06:16:57 crc kubenswrapper[4809]: I1206 06:16:57.951581 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e1fb3895287a3164ef0b7f918151a1d6b9bd8377505b3589e7842cd1dbc2702" Dec 06 06:16:57 crc kubenswrapper[4809]: I1206 06:16:57.953374 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecb6c22c-402b-4bdb-a791-29509985785a-config\") pod \"dnsmasq-dns-55f844cf75-58svx\" (UID: \"ecb6c22c-402b-4bdb-a791-29509985785a\") " pod="openstack/dnsmasq-dns-55f844cf75-58svx" Dec 06 06:16:57 crc kubenswrapper[4809]: I1206 06:16:57.953569 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8qj7\" (UniqueName: \"kubernetes.io/projected/ecb6c22c-402b-4bdb-a791-29509985785a-kube-api-access-t8qj7\") pod \"dnsmasq-dns-55f844cf75-58svx\" (UID: \"ecb6c22c-402b-4bdb-a791-29509985785a\") " pod="openstack/dnsmasq-dns-55f844cf75-58svx" Dec 06 06:16:57 crc kubenswrapper[4809]: I1206 06:16:57.953613 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ecb6c22c-402b-4bdb-a791-29509985785a-dns-svc\") pod \"dnsmasq-dns-55f844cf75-58svx\" (UID: \"ecb6c22c-402b-4bdb-a791-29509985785a\") " pod="openstack/dnsmasq-dns-55f844cf75-58svx" Dec 06 06:16:57 crc kubenswrapper[4809]: I1206 06:16:57.953646 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ecb6c22c-402b-4bdb-a791-29509985785a-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-58svx\" (UID: \"ecb6c22c-402b-4bdb-a791-29509985785a\") " pod="openstack/dnsmasq-dns-55f844cf75-58svx" Dec 06 06:16:57 crc kubenswrapper[4809]: I1206 06:16:57.953688 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ecb6c22c-402b-4bdb-a791-29509985785a-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-58svx\" (UID: \"ecb6c22c-402b-4bdb-a791-29509985785a\") " pod="openstack/dnsmasq-dns-55f844cf75-58svx" Dec 06 06:16:57 crc kubenswrapper[4809]: I1206 06:16:57.953707 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ecb6c22c-402b-4bdb-a791-29509985785a-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-58svx\" (UID: \"ecb6c22c-402b-4bdb-a791-29509985785a\") " pod="openstack/dnsmasq-dns-55f844cf75-58svx" Dec 06 06:16:57 crc kubenswrapper[4809]: I1206 06:16:57.957252 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-tc5gt" podStartSLOduration=15.957237597 podStartE2EDuration="15.957237597s" podCreationTimestamp="2025-12-06 06:16:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:16:57.944564626 +0000 UTC m=+1542.833547658" watchObservedRunningTime="2025-12-06 06:16:57.957237597 +0000 UTC m=+1542.846220529" Dec 06 06:16:58 crc kubenswrapper[4809]: I1206 06:16:58.035746 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7899fdcd68-m4gx6"] Dec 06 06:16:58 crc kubenswrapper[4809]: I1206 06:16:58.038495 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7899fdcd68-m4gx6" Dec 06 06:16:58 crc kubenswrapper[4809]: I1206 06:16:58.043740 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 06 06:16:58 crc kubenswrapper[4809]: I1206 06:16:58.043991 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 06 06:16:58 crc kubenswrapper[4809]: I1206 06:16:58.044105 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 06 06:16:58 crc kubenswrapper[4809]: I1206 06:16:58.044582 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-t2ttn" Dec 06 06:16:58 crc kubenswrapper[4809]: I1206 06:16:58.058500 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecb6c22c-402b-4bdb-a791-29509985785a-config\") pod \"dnsmasq-dns-55f844cf75-58svx\" (UID: \"ecb6c22c-402b-4bdb-a791-29509985785a\") " pod="openstack/dnsmasq-dns-55f844cf75-58svx" Dec 06 06:16:58 crc kubenswrapper[4809]: I1206 06:16:58.058907 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8qj7\" (UniqueName: \"kubernetes.io/projected/ecb6c22c-402b-4bdb-a791-29509985785a-kube-api-access-t8qj7\") pod \"dnsmasq-dns-55f844cf75-58svx\" (UID: \"ecb6c22c-402b-4bdb-a791-29509985785a\") " pod="openstack/dnsmasq-dns-55f844cf75-58svx" Dec 06 06:16:58 crc kubenswrapper[4809]: I1206 06:16:58.058980 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ecb6c22c-402b-4bdb-a791-29509985785a-dns-svc\") pod \"dnsmasq-dns-55f844cf75-58svx\" (UID: \"ecb6c22c-402b-4bdb-a791-29509985785a\") " pod="openstack/dnsmasq-dns-55f844cf75-58svx" Dec 06 06:16:58 crc kubenswrapper[4809]: I1206 06:16:58.059039 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ecb6c22c-402b-4bdb-a791-29509985785a-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-58svx\" (UID: \"ecb6c22c-402b-4bdb-a791-29509985785a\") " pod="openstack/dnsmasq-dns-55f844cf75-58svx" Dec 06 06:16:58 crc kubenswrapper[4809]: I1206 06:16:58.059195 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ecb6c22c-402b-4bdb-a791-29509985785a-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-58svx\" (UID: \"ecb6c22c-402b-4bdb-a791-29509985785a\") " pod="openstack/dnsmasq-dns-55f844cf75-58svx" Dec 06 06:16:58 crc kubenswrapper[4809]: I1206 06:16:58.059225 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ecb6c22c-402b-4bdb-a791-29509985785a-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-58svx\" (UID: \"ecb6c22c-402b-4bdb-a791-29509985785a\") " pod="openstack/dnsmasq-dns-55f844cf75-58svx" Dec 06 06:16:58 crc kubenswrapper[4809]: I1206 06:16:58.060125 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ecb6c22c-402b-4bdb-a791-29509985785a-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-58svx\" (UID: \"ecb6c22c-402b-4bdb-a791-29509985785a\") " pod="openstack/dnsmasq-dns-55f844cf75-58svx" Dec 06 06:16:58 crc kubenswrapper[4809]: I1206 06:16:58.061314 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ecb6c22c-402b-4bdb-a791-29509985785a-dns-svc\") pod \"dnsmasq-dns-55f844cf75-58svx\" (UID: \"ecb6c22c-402b-4bdb-a791-29509985785a\") " pod="openstack/dnsmasq-dns-55f844cf75-58svx" Dec 06 06:16:58 crc kubenswrapper[4809]: I1206 06:16:58.061786 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ecb6c22c-402b-4bdb-a791-29509985785a-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-58svx\" (UID: \"ecb6c22c-402b-4bdb-a791-29509985785a\") " pod="openstack/dnsmasq-dns-55f844cf75-58svx" Dec 06 06:16:58 crc kubenswrapper[4809]: I1206 06:16:58.064030 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ecb6c22c-402b-4bdb-a791-29509985785a-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-58svx\" (UID: \"ecb6c22c-402b-4bdb-a791-29509985785a\") " pod="openstack/dnsmasq-dns-55f844cf75-58svx" Dec 06 06:16:58 crc kubenswrapper[4809]: I1206 06:16:58.076886 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecb6c22c-402b-4bdb-a791-29509985785a-config\") pod \"dnsmasq-dns-55f844cf75-58svx\" (UID: \"ecb6c22c-402b-4bdb-a791-29509985785a\") " pod="openstack/dnsmasq-dns-55f844cf75-58svx" Dec 06 06:16:58 crc kubenswrapper[4809]: I1206 06:16:58.093137 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 06 06:16:58 crc kubenswrapper[4809]: I1206 06:16:58.093627 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 06 06:16:58 crc kubenswrapper[4809]: I1206 06:16:58.109269 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8qj7\" (UniqueName: \"kubernetes.io/projected/ecb6c22c-402b-4bdb-a791-29509985785a-kube-api-access-t8qj7\") pod \"dnsmasq-dns-55f844cf75-58svx\" (UID: \"ecb6c22c-402b-4bdb-a791-29509985785a\") " pod="openstack/dnsmasq-dns-55f844cf75-58svx" Dec 06 06:16:58 crc kubenswrapper[4809]: I1206 06:16:58.118923 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7899fdcd68-m4gx6"] Dec 06 06:16:58 crc kubenswrapper[4809]: I1206 06:16:58.122224 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 06 06:16:58 crc kubenswrapper[4809]: I1206 06:16:58.164256 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4bba6eef-372a-4b9a-a27c-76f834e3519b-httpd-config\") pod \"neutron-7899fdcd68-m4gx6\" (UID: \"4bba6eef-372a-4b9a-a27c-76f834e3519b\") " pod="openstack/neutron-7899fdcd68-m4gx6" Dec 06 06:16:58 crc kubenswrapper[4809]: I1206 06:16:58.164316 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-np4vn\" (UniqueName: \"kubernetes.io/projected/4bba6eef-372a-4b9a-a27c-76f834e3519b-kube-api-access-np4vn\") pod \"neutron-7899fdcd68-m4gx6\" (UID: \"4bba6eef-372a-4b9a-a27c-76f834e3519b\") " pod="openstack/neutron-7899fdcd68-m4gx6" Dec 06 06:16:58 crc kubenswrapper[4809]: I1206 06:16:58.164418 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4bba6eef-372a-4b9a-a27c-76f834e3519b-config\") pod \"neutron-7899fdcd68-m4gx6\" (UID: \"4bba6eef-372a-4b9a-a27c-76f834e3519b\") " pod="openstack/neutron-7899fdcd68-m4gx6" Dec 06 06:16:58 crc kubenswrapper[4809]: I1206 06:16:58.164502 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bba6eef-372a-4b9a-a27c-76f834e3519b-combined-ca-bundle\") pod \"neutron-7899fdcd68-m4gx6\" (UID: \"4bba6eef-372a-4b9a-a27c-76f834e3519b\") " pod="openstack/neutron-7899fdcd68-m4gx6" Dec 06 06:16:58 crc kubenswrapper[4809]: I1206 06:16:58.164654 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bba6eef-372a-4b9a-a27c-76f834e3519b-ovndb-tls-certs\") pod \"neutron-7899fdcd68-m4gx6\" (UID: \"4bba6eef-372a-4b9a-a27c-76f834e3519b\") " pod="openstack/neutron-7899fdcd68-m4gx6" Dec 06 06:16:58 crc kubenswrapper[4809]: I1206 06:16:58.236015 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-58svx" Dec 06 06:16:58 crc kubenswrapper[4809]: I1206 06:16:58.266301 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4bba6eef-372a-4b9a-a27c-76f834e3519b-config\") pod \"neutron-7899fdcd68-m4gx6\" (UID: \"4bba6eef-372a-4b9a-a27c-76f834e3519b\") " pod="openstack/neutron-7899fdcd68-m4gx6" Dec 06 06:16:58 crc kubenswrapper[4809]: I1206 06:16:58.266616 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bba6eef-372a-4b9a-a27c-76f834e3519b-combined-ca-bundle\") pod \"neutron-7899fdcd68-m4gx6\" (UID: \"4bba6eef-372a-4b9a-a27c-76f834e3519b\") " pod="openstack/neutron-7899fdcd68-m4gx6" Dec 06 06:16:58 crc kubenswrapper[4809]: I1206 06:16:58.266698 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bba6eef-372a-4b9a-a27c-76f834e3519b-ovndb-tls-certs\") pod \"neutron-7899fdcd68-m4gx6\" (UID: \"4bba6eef-372a-4b9a-a27c-76f834e3519b\") " pod="openstack/neutron-7899fdcd68-m4gx6" Dec 06 06:16:58 crc kubenswrapper[4809]: I1206 06:16:58.266766 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4bba6eef-372a-4b9a-a27c-76f834e3519b-httpd-config\") pod \"neutron-7899fdcd68-m4gx6\" (UID: \"4bba6eef-372a-4b9a-a27c-76f834e3519b\") " pod="openstack/neutron-7899fdcd68-m4gx6" Dec 06 06:16:58 crc kubenswrapper[4809]: I1206 06:16:58.266783 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-np4vn\" (UniqueName: \"kubernetes.io/projected/4bba6eef-372a-4b9a-a27c-76f834e3519b-kube-api-access-np4vn\") pod \"neutron-7899fdcd68-m4gx6\" (UID: \"4bba6eef-372a-4b9a-a27c-76f834e3519b\") " pod="openstack/neutron-7899fdcd68-m4gx6" Dec 06 06:16:58 crc kubenswrapper[4809]: I1206 06:16:58.276526 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bba6eef-372a-4b9a-a27c-76f834e3519b-combined-ca-bundle\") pod \"neutron-7899fdcd68-m4gx6\" (UID: \"4bba6eef-372a-4b9a-a27c-76f834e3519b\") " pod="openstack/neutron-7899fdcd68-m4gx6" Dec 06 06:16:58 crc kubenswrapper[4809]: I1206 06:16:58.278147 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/4bba6eef-372a-4b9a-a27c-76f834e3519b-config\") pod \"neutron-7899fdcd68-m4gx6\" (UID: \"4bba6eef-372a-4b9a-a27c-76f834e3519b\") " pod="openstack/neutron-7899fdcd68-m4gx6" Dec 06 06:16:58 crc kubenswrapper[4809]: I1206 06:16:58.278616 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4bba6eef-372a-4b9a-a27c-76f834e3519b-httpd-config\") pod \"neutron-7899fdcd68-m4gx6\" (UID: \"4bba6eef-372a-4b9a-a27c-76f834e3519b\") " pod="openstack/neutron-7899fdcd68-m4gx6" Dec 06 06:16:58 crc kubenswrapper[4809]: I1206 06:16:58.295628 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bba6eef-372a-4b9a-a27c-76f834e3519b-ovndb-tls-certs\") pod \"neutron-7899fdcd68-m4gx6\" (UID: \"4bba6eef-372a-4b9a-a27c-76f834e3519b\") " pod="openstack/neutron-7899fdcd68-m4gx6" Dec 06 06:16:58 crc kubenswrapper[4809]: I1206 06:16:58.315578 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-np4vn\" (UniqueName: \"kubernetes.io/projected/4bba6eef-372a-4b9a-a27c-76f834e3519b-kube-api-access-np4vn\") pod \"neutron-7899fdcd68-m4gx6\" (UID: \"4bba6eef-372a-4b9a-a27c-76f834e3519b\") " pod="openstack/neutron-7899fdcd68-m4gx6" Dec 06 06:16:58 crc kubenswrapper[4809]: I1206 06:16:58.377664 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7899fdcd68-m4gx6" Dec 06 06:16:58 crc kubenswrapper[4809]: I1206 06:16:58.767409 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 06:16:58 crc kubenswrapper[4809]: I1206 06:16:58.918214 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 06:16:58 crc kubenswrapper[4809]: I1206 06:16:58.985910 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"df904bdc-a504-4deb-8dd2-d2acf7626200","Type":"ContainerStarted","Data":"58243ecc3fcf5a0158d7575c80f30f61d95ee72566dafb41c1a55d282ff97ec8"} Dec 06 06:16:59 crc kubenswrapper[4809]: I1206 06:16:59.035575 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5b6f4174-473c-4e12-bcbb-7dd6ab7633ee","Type":"ContainerStarted","Data":"0e7b4374efdd9b613874cceea8597496abeca32c82b77e975e2722898bc35d40"} Dec 06 06:16:59 crc kubenswrapper[4809]: I1206 06:16:59.040717 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 06 06:17:01 crc kubenswrapper[4809]: I1206 06:17:01.449767 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-988458f5c-5zk66"] Dec 06 06:17:01 crc kubenswrapper[4809]: I1206 06:17:01.469508 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-988458f5c-5zk66" Dec 06 06:17:01 crc kubenswrapper[4809]: I1206 06:17:01.469666 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-988458f5c-5zk66"] Dec 06 06:17:01 crc kubenswrapper[4809]: I1206 06:17:01.472980 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 06 06:17:01 crc kubenswrapper[4809]: I1206 06:17:01.488293 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 06 06:17:01 crc kubenswrapper[4809]: I1206 06:17:01.548334 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6796bbd5-687e-4925-9ca1-700187c93ab8-ovndb-tls-certs\") pod \"neutron-988458f5c-5zk66\" (UID: \"6796bbd5-687e-4925-9ca1-700187c93ab8\") " pod="openstack/neutron-988458f5c-5zk66" Dec 06 06:17:01 crc kubenswrapper[4809]: I1206 06:17:01.548399 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6796bbd5-687e-4925-9ca1-700187c93ab8-public-tls-certs\") pod \"neutron-988458f5c-5zk66\" (UID: \"6796bbd5-687e-4925-9ca1-700187c93ab8\") " pod="openstack/neutron-988458f5c-5zk66" Dec 06 06:17:01 crc kubenswrapper[4809]: I1206 06:17:01.548431 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6796bbd5-687e-4925-9ca1-700187c93ab8-httpd-config\") pod \"neutron-988458f5c-5zk66\" (UID: \"6796bbd5-687e-4925-9ca1-700187c93ab8\") " pod="openstack/neutron-988458f5c-5zk66" Dec 06 06:17:01 crc kubenswrapper[4809]: I1206 06:17:01.548520 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6796bbd5-687e-4925-9ca1-700187c93ab8-combined-ca-bundle\") pod \"neutron-988458f5c-5zk66\" (UID: \"6796bbd5-687e-4925-9ca1-700187c93ab8\") " pod="openstack/neutron-988458f5c-5zk66" Dec 06 06:17:01 crc kubenswrapper[4809]: I1206 06:17:01.548550 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grp6p\" (UniqueName: \"kubernetes.io/projected/6796bbd5-687e-4925-9ca1-700187c93ab8-kube-api-access-grp6p\") pod \"neutron-988458f5c-5zk66\" (UID: \"6796bbd5-687e-4925-9ca1-700187c93ab8\") " pod="openstack/neutron-988458f5c-5zk66" Dec 06 06:17:01 crc kubenswrapper[4809]: I1206 06:17:01.548576 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6796bbd5-687e-4925-9ca1-700187c93ab8-internal-tls-certs\") pod \"neutron-988458f5c-5zk66\" (UID: \"6796bbd5-687e-4925-9ca1-700187c93ab8\") " pod="openstack/neutron-988458f5c-5zk66" Dec 06 06:17:01 crc kubenswrapper[4809]: I1206 06:17:01.548667 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6796bbd5-687e-4925-9ca1-700187c93ab8-config\") pod \"neutron-988458f5c-5zk66\" (UID: \"6796bbd5-687e-4925-9ca1-700187c93ab8\") " pod="openstack/neutron-988458f5c-5zk66" Dec 06 06:17:01 crc kubenswrapper[4809]: I1206 06:17:01.650025 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6796bbd5-687e-4925-9ca1-700187c93ab8-config\") pod \"neutron-988458f5c-5zk66\" (UID: \"6796bbd5-687e-4925-9ca1-700187c93ab8\") " pod="openstack/neutron-988458f5c-5zk66" Dec 06 06:17:01 crc kubenswrapper[4809]: I1206 06:17:01.650128 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6796bbd5-687e-4925-9ca1-700187c93ab8-ovndb-tls-certs\") pod \"neutron-988458f5c-5zk66\" (UID: \"6796bbd5-687e-4925-9ca1-700187c93ab8\") " pod="openstack/neutron-988458f5c-5zk66" Dec 06 06:17:01 crc kubenswrapper[4809]: I1206 06:17:01.650157 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6796bbd5-687e-4925-9ca1-700187c93ab8-public-tls-certs\") pod \"neutron-988458f5c-5zk66\" (UID: \"6796bbd5-687e-4925-9ca1-700187c93ab8\") " pod="openstack/neutron-988458f5c-5zk66" Dec 06 06:17:01 crc kubenswrapper[4809]: I1206 06:17:01.650174 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6796bbd5-687e-4925-9ca1-700187c93ab8-httpd-config\") pod \"neutron-988458f5c-5zk66\" (UID: \"6796bbd5-687e-4925-9ca1-700187c93ab8\") " pod="openstack/neutron-988458f5c-5zk66" Dec 06 06:17:01 crc kubenswrapper[4809]: I1206 06:17:01.650233 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6796bbd5-687e-4925-9ca1-700187c93ab8-combined-ca-bundle\") pod \"neutron-988458f5c-5zk66\" (UID: \"6796bbd5-687e-4925-9ca1-700187c93ab8\") " pod="openstack/neutron-988458f5c-5zk66" Dec 06 06:17:01 crc kubenswrapper[4809]: I1206 06:17:01.650249 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grp6p\" (UniqueName: \"kubernetes.io/projected/6796bbd5-687e-4925-9ca1-700187c93ab8-kube-api-access-grp6p\") pod \"neutron-988458f5c-5zk66\" (UID: \"6796bbd5-687e-4925-9ca1-700187c93ab8\") " pod="openstack/neutron-988458f5c-5zk66" Dec 06 06:17:01 crc kubenswrapper[4809]: I1206 06:17:01.650266 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6796bbd5-687e-4925-9ca1-700187c93ab8-internal-tls-certs\") pod \"neutron-988458f5c-5zk66\" (UID: \"6796bbd5-687e-4925-9ca1-700187c93ab8\") " pod="openstack/neutron-988458f5c-5zk66" Dec 06 06:17:01 crc kubenswrapper[4809]: I1206 06:17:01.657600 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6796bbd5-687e-4925-9ca1-700187c93ab8-combined-ca-bundle\") pod \"neutron-988458f5c-5zk66\" (UID: \"6796bbd5-687e-4925-9ca1-700187c93ab8\") " pod="openstack/neutron-988458f5c-5zk66" Dec 06 06:17:01 crc kubenswrapper[4809]: I1206 06:17:01.659802 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6796bbd5-687e-4925-9ca1-700187c93ab8-ovndb-tls-certs\") pod \"neutron-988458f5c-5zk66\" (UID: \"6796bbd5-687e-4925-9ca1-700187c93ab8\") " pod="openstack/neutron-988458f5c-5zk66" Dec 06 06:17:01 crc kubenswrapper[4809]: I1206 06:17:01.663628 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6796bbd5-687e-4925-9ca1-700187c93ab8-public-tls-certs\") pod \"neutron-988458f5c-5zk66\" (UID: \"6796bbd5-687e-4925-9ca1-700187c93ab8\") " pod="openstack/neutron-988458f5c-5zk66" Dec 06 06:17:01 crc kubenswrapper[4809]: I1206 06:17:01.670212 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6796bbd5-687e-4925-9ca1-700187c93ab8-httpd-config\") pod \"neutron-988458f5c-5zk66\" (UID: \"6796bbd5-687e-4925-9ca1-700187c93ab8\") " pod="openstack/neutron-988458f5c-5zk66" Dec 06 06:17:01 crc kubenswrapper[4809]: I1206 06:17:01.670483 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/6796bbd5-687e-4925-9ca1-700187c93ab8-config\") pod \"neutron-988458f5c-5zk66\" (UID: \"6796bbd5-687e-4925-9ca1-700187c93ab8\") " pod="openstack/neutron-988458f5c-5zk66" Dec 06 06:17:01 crc kubenswrapper[4809]: I1206 06:17:01.674817 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6796bbd5-687e-4925-9ca1-700187c93ab8-internal-tls-certs\") pod \"neutron-988458f5c-5zk66\" (UID: \"6796bbd5-687e-4925-9ca1-700187c93ab8\") " pod="openstack/neutron-988458f5c-5zk66" Dec 06 06:17:01 crc kubenswrapper[4809]: I1206 06:17:01.681708 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grp6p\" (UniqueName: \"kubernetes.io/projected/6796bbd5-687e-4925-9ca1-700187c93ab8-kube-api-access-grp6p\") pod \"neutron-988458f5c-5zk66\" (UID: \"6796bbd5-687e-4925-9ca1-700187c93ab8\") " pod="openstack/neutron-988458f5c-5zk66" Dec 06 06:17:01 crc kubenswrapper[4809]: I1206 06:17:01.824852 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-988458f5c-5zk66" Dec 06 06:17:02 crc kubenswrapper[4809]: I1206 06:17:02.066691 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-b75nx" event={"ID":"40abc93a-e45f-44cb-8fb1-f14961fe3e74","Type":"ContainerDied","Data":"4e3d67c3955338a59473286e274b6b5de4272bef4db2488748449eb617c9af15"} Dec 06 06:17:02 crc kubenswrapper[4809]: I1206 06:17:02.066735 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e3d67c3955338a59473286e274b6b5de4272bef4db2488748449eb617c9af15" Dec 06 06:17:02 crc kubenswrapper[4809]: I1206 06:17:02.200993 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-b75nx" Dec 06 06:17:02 crc kubenswrapper[4809]: I1206 06:17:02.263561 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/40abc93a-e45f-44cb-8fb1-f14961fe3e74-ovsdbserver-sb\") pod \"40abc93a-e45f-44cb-8fb1-f14961fe3e74\" (UID: \"40abc93a-e45f-44cb-8fb1-f14961fe3e74\") " Dec 06 06:17:02 crc kubenswrapper[4809]: I1206 06:17:02.263720 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/40abc93a-e45f-44cb-8fb1-f14961fe3e74-dns-svc\") pod \"40abc93a-e45f-44cb-8fb1-f14961fe3e74\" (UID: \"40abc93a-e45f-44cb-8fb1-f14961fe3e74\") " Dec 06 06:17:02 crc kubenswrapper[4809]: I1206 06:17:02.263839 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40abc93a-e45f-44cb-8fb1-f14961fe3e74-config\") pod \"40abc93a-e45f-44cb-8fb1-f14961fe3e74\" (UID: \"40abc93a-e45f-44cb-8fb1-f14961fe3e74\") " Dec 06 06:17:02 crc kubenswrapper[4809]: I1206 06:17:02.263909 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/40abc93a-e45f-44cb-8fb1-f14961fe3e74-ovsdbserver-nb\") pod \"40abc93a-e45f-44cb-8fb1-f14961fe3e74\" (UID: \"40abc93a-e45f-44cb-8fb1-f14961fe3e74\") " Dec 06 06:17:02 crc kubenswrapper[4809]: I1206 06:17:02.263969 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wngr7\" (UniqueName: \"kubernetes.io/projected/40abc93a-e45f-44cb-8fb1-f14961fe3e74-kube-api-access-wngr7\") pod \"40abc93a-e45f-44cb-8fb1-f14961fe3e74\" (UID: \"40abc93a-e45f-44cb-8fb1-f14961fe3e74\") " Dec 06 06:17:02 crc kubenswrapper[4809]: I1206 06:17:02.264094 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/40abc93a-e45f-44cb-8fb1-f14961fe3e74-dns-swift-storage-0\") pod \"40abc93a-e45f-44cb-8fb1-f14961fe3e74\" (UID: \"40abc93a-e45f-44cb-8fb1-f14961fe3e74\") " Dec 06 06:17:02 crc kubenswrapper[4809]: I1206 06:17:02.273787 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40abc93a-e45f-44cb-8fb1-f14961fe3e74-kube-api-access-wngr7" (OuterVolumeSpecName: "kube-api-access-wngr7") pod "40abc93a-e45f-44cb-8fb1-f14961fe3e74" (UID: "40abc93a-e45f-44cb-8fb1-f14961fe3e74"). InnerVolumeSpecName "kube-api-access-wngr7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:17:02 crc kubenswrapper[4809]: I1206 06:17:02.294493 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40abc93a-e45f-44cb-8fb1-f14961fe3e74-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "40abc93a-e45f-44cb-8fb1-f14961fe3e74" (UID: "40abc93a-e45f-44cb-8fb1-f14961fe3e74"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:17:02 crc kubenswrapper[4809]: I1206 06:17:02.295866 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40abc93a-e45f-44cb-8fb1-f14961fe3e74-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "40abc93a-e45f-44cb-8fb1-f14961fe3e74" (UID: "40abc93a-e45f-44cb-8fb1-f14961fe3e74"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:17:02 crc kubenswrapper[4809]: I1206 06:17:02.298382 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40abc93a-e45f-44cb-8fb1-f14961fe3e74-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "40abc93a-e45f-44cb-8fb1-f14961fe3e74" (UID: "40abc93a-e45f-44cb-8fb1-f14961fe3e74"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:17:02 crc kubenswrapper[4809]: I1206 06:17:02.310831 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40abc93a-e45f-44cb-8fb1-f14961fe3e74-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "40abc93a-e45f-44cb-8fb1-f14961fe3e74" (UID: "40abc93a-e45f-44cb-8fb1-f14961fe3e74"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:17:02 crc kubenswrapper[4809]: I1206 06:17:02.325115 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40abc93a-e45f-44cb-8fb1-f14961fe3e74-config" (OuterVolumeSpecName: "config") pod "40abc93a-e45f-44cb-8fb1-f14961fe3e74" (UID: "40abc93a-e45f-44cb-8fb1-f14961fe3e74"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:17:02 crc kubenswrapper[4809]: I1206 06:17:02.367017 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/40abc93a-e45f-44cb-8fb1-f14961fe3e74-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:02 crc kubenswrapper[4809]: I1206 06:17:02.367056 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wngr7\" (UniqueName: \"kubernetes.io/projected/40abc93a-e45f-44cb-8fb1-f14961fe3e74-kube-api-access-wngr7\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:02 crc kubenswrapper[4809]: I1206 06:17:02.367066 4809 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/40abc93a-e45f-44cb-8fb1-f14961fe3e74-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:02 crc kubenswrapper[4809]: I1206 06:17:02.367077 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/40abc93a-e45f-44cb-8fb1-f14961fe3e74-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:02 crc kubenswrapper[4809]: I1206 06:17:02.367086 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/40abc93a-e45f-44cb-8fb1-f14961fe3e74-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:02 crc kubenswrapper[4809]: I1206 06:17:02.367095 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40abc93a-e45f-44cb-8fb1-f14961fe3e74-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:02 crc kubenswrapper[4809]: I1206 06:17:02.623008 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-988458f5c-5zk66"] Dec 06 06:17:02 crc kubenswrapper[4809]: I1206 06:17:02.706199 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-58svx"] Dec 06 06:17:03 crc kubenswrapper[4809]: I1206 06:17:03.076942 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-b75nx" Dec 06 06:17:03 crc kubenswrapper[4809]: I1206 06:17:03.157283 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-b75nx"] Dec 06 06:17:03 crc kubenswrapper[4809]: I1206 06:17:03.170010 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-b75nx"] Dec 06 06:17:03 crc kubenswrapper[4809]: I1206 06:17:03.401992 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40abc93a-e45f-44cb-8fb1-f14961fe3e74" path="/var/lib/kubelet/pods/40abc93a-e45f-44cb-8fb1-f14961fe3e74/volumes" Dec 06 06:17:03 crc kubenswrapper[4809]: I1206 06:17:03.458249 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7899fdcd68-m4gx6"] Dec 06 06:17:03 crc kubenswrapper[4809]: W1206 06:17:03.482595 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podecb6c22c_402b_4bdb_a791_29509985785a.slice/crio-3ef6aa1add3472dbb20bb435c15329561d4fb74293136cdaae01957e26374cec WatchSource:0}: Error finding container 3ef6aa1add3472dbb20bb435c15329561d4fb74293136cdaae01957e26374cec: Status 404 returned error can't find the container with id 3ef6aa1add3472dbb20bb435c15329561d4fb74293136cdaae01957e26374cec Dec 06 06:17:03 crc kubenswrapper[4809]: W1206 06:17:03.504111 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4bba6eef_372a_4b9a_a27c_76f834e3519b.slice/crio-8ede14d1ccf475536a392841b50ecdfa8c618338be8ca8321222153be7222ae0 WatchSource:0}: Error finding container 8ede14d1ccf475536a392841b50ecdfa8c618338be8ca8321222153be7222ae0: Status 404 returned error can't find the container with id 8ede14d1ccf475536a392841b50ecdfa8c618338be8ca8321222153be7222ae0 Dec 06 06:17:04 crc kubenswrapper[4809]: I1206 06:17:04.104072 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7899fdcd68-m4gx6" event={"ID":"4bba6eef-372a-4b9a-a27c-76f834e3519b","Type":"ContainerStarted","Data":"8ede14d1ccf475536a392841b50ecdfa8c618338be8ca8321222153be7222ae0"} Dec 06 06:17:04 crc kubenswrapper[4809]: I1206 06:17:04.106039 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-988458f5c-5zk66" event={"ID":"6796bbd5-687e-4925-9ca1-700187c93ab8","Type":"ContainerStarted","Data":"cd47081608dcdcf518e66a2dabb54668f2b15b46d65fe7740dd70f0aef25da98"} Dec 06 06:17:04 crc kubenswrapper[4809]: I1206 06:17:04.109201 4809 generic.go:334] "Generic (PLEG): container finished" podID="c805cd51-5aaa-4c78-b98e-29a5a6ae44a9" containerID="096c45017712c6fe423f137af1e72ea25ccc70231ab4d70c49373ba0b823944c" exitCode=0 Dec 06 06:17:04 crc kubenswrapper[4809]: I1206 06:17:04.109260 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-tc5gt" event={"ID":"c805cd51-5aaa-4c78-b98e-29a5a6ae44a9","Type":"ContainerDied","Data":"096c45017712c6fe423f137af1e72ea25ccc70231ab4d70c49373ba0b823944c"} Dec 06 06:17:04 crc kubenswrapper[4809]: I1206 06:17:04.112208 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-58svx" event={"ID":"ecb6c22c-402b-4bdb-a791-29509985785a","Type":"ContainerStarted","Data":"3ef6aa1add3472dbb20bb435c15329561d4fb74293136cdaae01957e26374cec"} Dec 06 06:17:05 crc kubenswrapper[4809]: I1206 06:17:05.126094 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"93e2d67f-9481-4ca3-83e7-4f7c0673b9d4","Type":"ContainerStarted","Data":"ce4c3b3f255d533aa317c246f5f07360cdecb16cf7ab4271977b7f91345a258f"} Dec 06 06:17:05 crc kubenswrapper[4809]: I1206 06:17:05.128912 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7899fdcd68-m4gx6" event={"ID":"4bba6eef-372a-4b9a-a27c-76f834e3519b","Type":"ContainerStarted","Data":"ec39c808be73c216788b8c2a0005c75163a22a80031d12660535a4f3142004ed"} Dec 06 06:17:05 crc kubenswrapper[4809]: I1206 06:17:05.129169 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7899fdcd68-m4gx6" event={"ID":"4bba6eef-372a-4b9a-a27c-76f834e3519b","Type":"ContainerStarted","Data":"79601c67c345fc5970db8138981011bc46985c80eaa23e7012e78c7dc020e526"} Dec 06 06:17:05 crc kubenswrapper[4809]: I1206 06:17:05.129344 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7899fdcd68-m4gx6" Dec 06 06:17:05 crc kubenswrapper[4809]: I1206 06:17:05.136780 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-988458f5c-5zk66" event={"ID":"6796bbd5-687e-4925-9ca1-700187c93ab8","Type":"ContainerStarted","Data":"2ef635299f803f062af78d889043d5026a826e5473a4d7e3f4f034c024239105"} Dec 06 06:17:05 crc kubenswrapper[4809]: I1206 06:17:05.136834 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-988458f5c-5zk66" event={"ID":"6796bbd5-687e-4925-9ca1-700187c93ab8","Type":"ContainerStarted","Data":"a1a772008b5315945c6d4eb144f336e989a9efcd4163c0f44945b7c21d288b73"} Dec 06 06:17:05 crc kubenswrapper[4809]: I1206 06:17:05.138267 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-988458f5c-5zk66" Dec 06 06:17:05 crc kubenswrapper[4809]: I1206 06:17:05.143890 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5b6f4174-473c-4e12-bcbb-7dd6ab7633ee","Type":"ContainerStarted","Data":"5ab1c005b596383180356c364a3e4f5a45a30307e69f7e04a113bbc5382f1c40"} Dec 06 06:17:05 crc kubenswrapper[4809]: I1206 06:17:05.144155 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5b6f4174-473c-4e12-bcbb-7dd6ab7633ee" containerName="glance-log" containerID="cri-o://0e7b4374efdd9b613874cceea8597496abeca32c82b77e975e2722898bc35d40" gracePeriod=30 Dec 06 06:17:05 crc kubenswrapper[4809]: I1206 06:17:05.144466 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5b6f4174-473c-4e12-bcbb-7dd6ab7633ee" containerName="glance-httpd" containerID="cri-o://5ab1c005b596383180356c364a3e4f5a45a30307e69f7e04a113bbc5382f1c40" gracePeriod=30 Dec 06 06:17:05 crc kubenswrapper[4809]: I1206 06:17:05.166414 4809 generic.go:334] "Generic (PLEG): container finished" podID="ecb6c22c-402b-4bdb-a791-29509985785a" containerID="29529fe529b79a26d45b811f45d236fc54331abc835dbe69763e64c6f471e0e8" exitCode=0 Dec 06 06:17:05 crc kubenswrapper[4809]: I1206 06:17:05.166464 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-58svx" event={"ID":"ecb6c22c-402b-4bdb-a791-29509985785a","Type":"ContainerDied","Data":"29529fe529b79a26d45b811f45d236fc54331abc835dbe69763e64c6f471e0e8"} Dec 06 06:17:05 crc kubenswrapper[4809]: I1206 06:17:05.172827 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-wxlvd" event={"ID":"b6e30c4b-8103-4399-b434-58bedae43fc5","Type":"ContainerStarted","Data":"dcaf40a21a826a0263dc333602e07e911b5ada26fa16b17644edb44c00f50d88"} Dec 06 06:17:05 crc kubenswrapper[4809]: I1206 06:17:05.173016 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-785d8bcb8c-wxlvd" podUID="b6e30c4b-8103-4399-b434-58bedae43fc5" containerName="dnsmasq-dns" containerID="cri-o://dcaf40a21a826a0263dc333602e07e911b5ada26fa16b17644edb44c00f50d88" gracePeriod=10 Dec 06 06:17:05 crc kubenswrapper[4809]: I1206 06:17:05.173175 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-785d8bcb8c-wxlvd" Dec 06 06:17:05 crc kubenswrapper[4809]: I1206 06:17:05.173737 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7899fdcd68-m4gx6" podStartSLOduration=7.1737274 podStartE2EDuration="7.1737274s" podCreationTimestamp="2025-12-06 06:16:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:17:05.151715036 +0000 UTC m=+1550.040698008" watchObservedRunningTime="2025-12-06 06:17:05.1737274 +0000 UTC m=+1550.062710342" Dec 06 06:17:05 crc kubenswrapper[4809]: I1206 06:17:05.191627 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-988458f5c-5zk66" podStartSLOduration=4.191276564 podStartE2EDuration="4.191276564s" podCreationTimestamp="2025-12-06 06:17:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:17:05.181227422 +0000 UTC m=+1550.070210374" watchObservedRunningTime="2025-12-06 06:17:05.191276564 +0000 UTC m=+1550.080259516" Dec 06 06:17:05 crc kubenswrapper[4809]: I1206 06:17:05.196028 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-m98hv" event={"ID":"8336af68-51f9-4f3a-8c92-61dd1aba9067","Type":"ContainerStarted","Data":"cbb746c60ce6a9134f7e45e0a79d24227920a2c5bece823b7c07cef89c91e92a"} Dec 06 06:17:05 crc kubenswrapper[4809]: I1206 06:17:05.202635 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-ckhg9" event={"ID":"edf1ab66-b483-4172-81fd-dd3f3b9b44aa","Type":"ContainerStarted","Data":"30b14dfc79cb4d967b8537d39fe468fc1409477b840153c0872891dfe7ab084b"} Dec 06 06:17:05 crc kubenswrapper[4809]: I1206 06:17:05.218622 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="df904bdc-a504-4deb-8dd2-d2acf7626200" containerName="glance-log" containerID="cri-o://58243ecc3fcf5a0158d7575c80f30f61d95ee72566dafb41c1a55d282ff97ec8" gracePeriod=30 Dec 06 06:17:05 crc kubenswrapper[4809]: I1206 06:17:05.218886 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"df904bdc-a504-4deb-8dd2-d2acf7626200","Type":"ContainerStarted","Data":"2300381bc4530f33d016676c2d9022f547c1d0cf3273d47b5bdc4ca71196cbef"} Dec 06 06:17:05 crc kubenswrapper[4809]: I1206 06:17:05.218948 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="df904bdc-a504-4deb-8dd2-d2acf7626200" containerName="glance-httpd" containerID="cri-o://2300381bc4530f33d016676c2d9022f547c1d0cf3273d47b5bdc4ca71196cbef" gracePeriod=30 Dec 06 06:17:05 crc kubenswrapper[4809]: I1206 06:17:05.223832 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=11.223808331 podStartE2EDuration="11.223808331s" podCreationTimestamp="2025-12-06 06:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:17:05.208540589 +0000 UTC m=+1550.097523531" watchObservedRunningTime="2025-12-06 06:17:05.223808331 +0000 UTC m=+1550.112791283" Dec 06 06:17:05 crc kubenswrapper[4809]: I1206 06:17:05.257749 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-ckhg9" podStartSLOduration=3.184515829 podStartE2EDuration="45.257728176s" podCreationTimestamp="2025-12-06 06:16:20 +0000 UTC" firstStartedPulling="2025-12-06 06:16:21.424783169 +0000 UTC m=+1506.313766111" lastFinishedPulling="2025-12-06 06:17:03.497995516 +0000 UTC m=+1548.386978458" observedRunningTime="2025-12-06 06:17:05.224267093 +0000 UTC m=+1550.113250055" watchObservedRunningTime="2025-12-06 06:17:05.257728176 +0000 UTC m=+1550.146711118" Dec 06 06:17:05 crc kubenswrapper[4809]: I1206 06:17:05.338501 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-m98hv" podStartSLOduration=4.565072059 podStartE2EDuration="45.338476435s" podCreationTimestamp="2025-12-06 06:16:20 +0000 UTC" firstStartedPulling="2025-12-06 06:16:22.731408884 +0000 UTC m=+1507.620391826" lastFinishedPulling="2025-12-06 06:17:03.50481326 +0000 UTC m=+1548.393796202" observedRunningTime="2025-12-06 06:17:05.283245524 +0000 UTC m=+1550.172228466" watchObservedRunningTime="2025-12-06 06:17:05.338476435 +0000 UTC m=+1550.227459377" Dec 06 06:17:05 crc kubenswrapper[4809]: I1206 06:17:05.350166 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-785d8bcb8c-wxlvd" podStartSLOduration=11.3501487 podStartE2EDuration="11.3501487s" podCreationTimestamp="2025-12-06 06:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:17:05.302490904 +0000 UTC m=+1550.191473846" watchObservedRunningTime="2025-12-06 06:17:05.3501487 +0000 UTC m=+1550.239131642" Dec 06 06:17:05 crc kubenswrapper[4809]: I1206 06:17:05.370112 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=11.370092998 podStartE2EDuration="11.370092998s" podCreationTimestamp="2025-12-06 06:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:17:05.330185261 +0000 UTC m=+1550.219168203" watchObservedRunningTime="2025-12-06 06:17:05.370092998 +0000 UTC m=+1550.259075940" Dec 06 06:17:06 crc kubenswrapper[4809]: I1206 06:17:06.218695 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-tc5gt" Dec 06 06:17:06 crc kubenswrapper[4809]: I1206 06:17:06.286605 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c805cd51-5aaa-4c78-b98e-29a5a6ae44a9-combined-ca-bundle\") pod \"c805cd51-5aaa-4c78-b98e-29a5a6ae44a9\" (UID: \"c805cd51-5aaa-4c78-b98e-29a5a6ae44a9\") " Dec 06 06:17:06 crc kubenswrapper[4809]: I1206 06:17:06.286679 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c805cd51-5aaa-4c78-b98e-29a5a6ae44a9-scripts\") pod \"c805cd51-5aaa-4c78-b98e-29a5a6ae44a9\" (UID: \"c805cd51-5aaa-4c78-b98e-29a5a6ae44a9\") " Dec 06 06:17:06 crc kubenswrapper[4809]: I1206 06:17:06.286787 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c805cd51-5aaa-4c78-b98e-29a5a6ae44a9-config-data\") pod \"c805cd51-5aaa-4c78-b98e-29a5a6ae44a9\" (UID: \"c805cd51-5aaa-4c78-b98e-29a5a6ae44a9\") " Dec 06 06:17:06 crc kubenswrapper[4809]: I1206 06:17:06.286825 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vpsfs\" (UniqueName: \"kubernetes.io/projected/c805cd51-5aaa-4c78-b98e-29a5a6ae44a9-kube-api-access-vpsfs\") pod \"c805cd51-5aaa-4c78-b98e-29a5a6ae44a9\" (UID: \"c805cd51-5aaa-4c78-b98e-29a5a6ae44a9\") " Dec 06 06:17:06 crc kubenswrapper[4809]: I1206 06:17:06.286868 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c805cd51-5aaa-4c78-b98e-29a5a6ae44a9-credential-keys\") pod \"c805cd51-5aaa-4c78-b98e-29a5a6ae44a9\" (UID: \"c805cd51-5aaa-4c78-b98e-29a5a6ae44a9\") " Dec 06 06:17:06 crc kubenswrapper[4809]: I1206 06:17:06.286904 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c805cd51-5aaa-4c78-b98e-29a5a6ae44a9-fernet-keys\") pod \"c805cd51-5aaa-4c78-b98e-29a5a6ae44a9\" (UID: \"c805cd51-5aaa-4c78-b98e-29a5a6ae44a9\") " Dec 06 06:17:06 crc kubenswrapper[4809]: I1206 06:17:06.293801 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c805cd51-5aaa-4c78-b98e-29a5a6ae44a9-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "c805cd51-5aaa-4c78-b98e-29a5a6ae44a9" (UID: "c805cd51-5aaa-4c78-b98e-29a5a6ae44a9"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:17:06 crc kubenswrapper[4809]: I1206 06:17:06.294335 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c805cd51-5aaa-4c78-b98e-29a5a6ae44a9-kube-api-access-vpsfs" (OuterVolumeSpecName: "kube-api-access-vpsfs") pod "c805cd51-5aaa-4c78-b98e-29a5a6ae44a9" (UID: "c805cd51-5aaa-4c78-b98e-29a5a6ae44a9"). InnerVolumeSpecName "kube-api-access-vpsfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:17:06 crc kubenswrapper[4809]: I1206 06:17:06.301420 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c805cd51-5aaa-4c78-b98e-29a5a6ae44a9-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "c805cd51-5aaa-4c78-b98e-29a5a6ae44a9" (UID: "c805cd51-5aaa-4c78-b98e-29a5a6ae44a9"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:17:06 crc kubenswrapper[4809]: I1206 06:17:06.302901 4809 generic.go:334] "Generic (PLEG): container finished" podID="b6e30c4b-8103-4399-b434-58bedae43fc5" containerID="dcaf40a21a826a0263dc333602e07e911b5ada26fa16b17644edb44c00f50d88" exitCode=0 Dec 06 06:17:06 crc kubenswrapper[4809]: I1206 06:17:06.303571 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-wxlvd" event={"ID":"b6e30c4b-8103-4399-b434-58bedae43fc5","Type":"ContainerDied","Data":"dcaf40a21a826a0263dc333602e07e911b5ada26fa16b17644edb44c00f50d88"} Dec 06 06:17:06 crc kubenswrapper[4809]: I1206 06:17:06.315440 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-tc5gt" event={"ID":"c805cd51-5aaa-4c78-b98e-29a5a6ae44a9","Type":"ContainerDied","Data":"a49b8a15dc5a86f3fddb084e2c1f54527c729dc39ff07863f28dbb445672ecec"} Dec 06 06:17:06 crc kubenswrapper[4809]: I1206 06:17:06.315476 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a49b8a15dc5a86f3fddb084e2c1f54527c729dc39ff07863f28dbb445672ecec" Dec 06 06:17:06 crc kubenswrapper[4809]: I1206 06:17:06.315529 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-tc5gt" Dec 06 06:17:06 crc kubenswrapper[4809]: I1206 06:17:06.318483 4809 generic.go:334] "Generic (PLEG): container finished" podID="df904bdc-a504-4deb-8dd2-d2acf7626200" containerID="2300381bc4530f33d016676c2d9022f547c1d0cf3273d47b5bdc4ca71196cbef" exitCode=0 Dec 06 06:17:06 crc kubenswrapper[4809]: I1206 06:17:06.318517 4809 generic.go:334] "Generic (PLEG): container finished" podID="df904bdc-a504-4deb-8dd2-d2acf7626200" containerID="58243ecc3fcf5a0158d7575c80f30f61d95ee72566dafb41c1a55d282ff97ec8" exitCode=143 Dec 06 06:17:06 crc kubenswrapper[4809]: I1206 06:17:06.318586 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"df904bdc-a504-4deb-8dd2-d2acf7626200","Type":"ContainerDied","Data":"2300381bc4530f33d016676c2d9022f547c1d0cf3273d47b5bdc4ca71196cbef"} Dec 06 06:17:06 crc kubenswrapper[4809]: I1206 06:17:06.318611 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"df904bdc-a504-4deb-8dd2-d2acf7626200","Type":"ContainerDied","Data":"58243ecc3fcf5a0158d7575c80f30f61d95ee72566dafb41c1a55d282ff97ec8"} Dec 06 06:17:06 crc kubenswrapper[4809]: I1206 06:17:06.321916 4809 generic.go:334] "Generic (PLEG): container finished" podID="5b6f4174-473c-4e12-bcbb-7dd6ab7633ee" containerID="5ab1c005b596383180356c364a3e4f5a45a30307e69f7e04a113bbc5382f1c40" exitCode=0 Dec 06 06:17:06 crc kubenswrapper[4809]: I1206 06:17:06.322027 4809 generic.go:334] "Generic (PLEG): container finished" podID="5b6f4174-473c-4e12-bcbb-7dd6ab7633ee" containerID="0e7b4374efdd9b613874cceea8597496abeca32c82b77e975e2722898bc35d40" exitCode=143 Dec 06 06:17:06 crc kubenswrapper[4809]: I1206 06:17:06.321971 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5b6f4174-473c-4e12-bcbb-7dd6ab7633ee","Type":"ContainerDied","Data":"5ab1c005b596383180356c364a3e4f5a45a30307e69f7e04a113bbc5382f1c40"} Dec 06 06:17:06 crc kubenswrapper[4809]: I1206 06:17:06.322390 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5b6f4174-473c-4e12-bcbb-7dd6ab7633ee","Type":"ContainerDied","Data":"0e7b4374efdd9b613874cceea8597496abeca32c82b77e975e2722898bc35d40"} Dec 06 06:17:06 crc kubenswrapper[4809]: I1206 06:17:06.322865 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c805cd51-5aaa-4c78-b98e-29a5a6ae44a9-scripts" (OuterVolumeSpecName: "scripts") pod "c805cd51-5aaa-4c78-b98e-29a5a6ae44a9" (UID: "c805cd51-5aaa-4c78-b98e-29a5a6ae44a9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:17:06 crc kubenswrapper[4809]: I1206 06:17:06.344753 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c805cd51-5aaa-4c78-b98e-29a5a6ae44a9-config-data" (OuterVolumeSpecName: "config-data") pod "c805cd51-5aaa-4c78-b98e-29a5a6ae44a9" (UID: "c805cd51-5aaa-4c78-b98e-29a5a6ae44a9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:17:06 crc kubenswrapper[4809]: I1206 06:17:06.349147 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c805cd51-5aaa-4c78-b98e-29a5a6ae44a9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c805cd51-5aaa-4c78-b98e-29a5a6ae44a9" (UID: "c805cd51-5aaa-4c78-b98e-29a5a6ae44a9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:17:06 crc kubenswrapper[4809]: I1206 06:17:06.389337 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c805cd51-5aaa-4c78-b98e-29a5a6ae44a9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:06 crc kubenswrapper[4809]: I1206 06:17:06.389381 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c805cd51-5aaa-4c78-b98e-29a5a6ae44a9-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:06 crc kubenswrapper[4809]: I1206 06:17:06.389392 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c805cd51-5aaa-4c78-b98e-29a5a6ae44a9-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:06 crc kubenswrapper[4809]: I1206 06:17:06.389400 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vpsfs\" (UniqueName: \"kubernetes.io/projected/c805cd51-5aaa-4c78-b98e-29a5a6ae44a9-kube-api-access-vpsfs\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:06 crc kubenswrapper[4809]: I1206 06:17:06.389410 4809 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c805cd51-5aaa-4c78-b98e-29a5a6ae44a9-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:06 crc kubenswrapper[4809]: I1206 06:17:06.389419 4809 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c805cd51-5aaa-4c78-b98e-29a5a6ae44a9-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.309534 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-wxlvd" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.310377 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.352565 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-6bb4ccfdc4-9jqw5"] Dec 06 06:17:07 crc kubenswrapper[4809]: E1206 06:17:07.353062 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6e30c4b-8103-4399-b434-58bedae43fc5" containerName="dnsmasq-dns" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.353076 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6e30c4b-8103-4399-b434-58bedae43fc5" containerName="dnsmasq-dns" Dec 06 06:17:07 crc kubenswrapper[4809]: E1206 06:17:07.353091 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df904bdc-a504-4deb-8dd2-d2acf7626200" containerName="glance-log" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.353099 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="df904bdc-a504-4deb-8dd2-d2acf7626200" containerName="glance-log" Dec 06 06:17:07 crc kubenswrapper[4809]: E1206 06:17:07.353121 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df904bdc-a504-4deb-8dd2-d2acf7626200" containerName="glance-httpd" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.353128 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="df904bdc-a504-4deb-8dd2-d2acf7626200" containerName="glance-httpd" Dec 06 06:17:07 crc kubenswrapper[4809]: E1206 06:17:07.353141 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40abc93a-e45f-44cb-8fb1-f14961fe3e74" containerName="init" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.353146 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="40abc93a-e45f-44cb-8fb1-f14961fe3e74" containerName="init" Dec 06 06:17:07 crc kubenswrapper[4809]: E1206 06:17:07.353172 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6e30c4b-8103-4399-b434-58bedae43fc5" containerName="init" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.353177 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6e30c4b-8103-4399-b434-58bedae43fc5" containerName="init" Dec 06 06:17:07 crc kubenswrapper[4809]: E1206 06:17:07.353192 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c805cd51-5aaa-4c78-b98e-29a5a6ae44a9" containerName="keystone-bootstrap" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.353198 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c805cd51-5aaa-4c78-b98e-29a5a6ae44a9" containerName="keystone-bootstrap" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.353554 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="df904bdc-a504-4deb-8dd2-d2acf7626200" containerName="glance-log" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.353568 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6e30c4b-8103-4399-b434-58bedae43fc5" containerName="dnsmasq-dns" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.353588 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="c805cd51-5aaa-4c78-b98e-29a5a6ae44a9" containerName="keystone-bootstrap" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.353625 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="df904bdc-a504-4deb-8dd2-d2acf7626200" containerName="glance-httpd" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.353648 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="40abc93a-e45f-44cb-8fb1-f14961fe3e74" containerName="init" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.354588 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6bb4ccfdc4-9jqw5" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.356091 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-58svx" event={"ID":"ecb6c22c-402b-4bdb-a791-29509985785a","Type":"ContainerStarted","Data":"bcf33d0fa0150896477a32304bd09c31c83fec771c3aec3d7578e7f98d8ed1e8"} Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.356422 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55f844cf75-58svx" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.357070 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.359830 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.360023 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.360152 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.360252 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.360349 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-2h2rg" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.405271 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-wxlvd" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.416706 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g42m7\" (UniqueName: \"kubernetes.io/projected/df904bdc-a504-4deb-8dd2-d2acf7626200-kube-api-access-g42m7\") pod \"df904bdc-a504-4deb-8dd2-d2acf7626200\" (UID: \"df904bdc-a504-4deb-8dd2-d2acf7626200\") " Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.421739 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df904bdc-a504-4deb-8dd2-d2acf7626200-combined-ca-bundle\") pod \"df904bdc-a504-4deb-8dd2-d2acf7626200\" (UID: \"df904bdc-a504-4deb-8dd2-d2acf7626200\") " Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.421857 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6e30c4b-8103-4399-b434-58bedae43fc5-config\") pod \"b6e30c4b-8103-4399-b434-58bedae43fc5\" (UID: \"b6e30c4b-8103-4399-b434-58bedae43fc5\") " Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.421889 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df904bdc-a504-4deb-8dd2-d2acf7626200-config-data\") pod \"df904bdc-a504-4deb-8dd2-d2acf7626200\" (UID: \"df904bdc-a504-4deb-8dd2-d2acf7626200\") " Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.421965 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-698pg\" (UniqueName: \"kubernetes.io/projected/b6e30c4b-8103-4399-b434-58bedae43fc5-kube-api-access-698pg\") pod \"b6e30c4b-8103-4399-b434-58bedae43fc5\" (UID: \"b6e30c4b-8103-4399-b434-58bedae43fc5\") " Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.422078 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/df904bdc-a504-4deb-8dd2-d2acf7626200-httpd-run\") pod \"df904bdc-a504-4deb-8dd2-d2acf7626200\" (UID: \"df904bdc-a504-4deb-8dd2-d2acf7626200\") " Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.422148 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df904bdc-a504-4deb-8dd2-d2acf7626200-logs\") pod \"df904bdc-a504-4deb-8dd2-d2acf7626200\" (UID: \"df904bdc-a504-4deb-8dd2-d2acf7626200\") " Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.422203 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"df904bdc-a504-4deb-8dd2-d2acf7626200\" (UID: \"df904bdc-a504-4deb-8dd2-d2acf7626200\") " Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.422226 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b6e30c4b-8103-4399-b434-58bedae43fc5-dns-swift-storage-0\") pod \"b6e30c4b-8103-4399-b434-58bedae43fc5\" (UID: \"b6e30c4b-8103-4399-b434-58bedae43fc5\") " Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.422242 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b6e30c4b-8103-4399-b434-58bedae43fc5-ovsdbserver-nb\") pod \"b6e30c4b-8103-4399-b434-58bedae43fc5\" (UID: \"b6e30c4b-8103-4399-b434-58bedae43fc5\") " Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.422257 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b6e30c4b-8103-4399-b434-58bedae43fc5-ovsdbserver-sb\") pod \"b6e30c4b-8103-4399-b434-58bedae43fc5\" (UID: \"b6e30c4b-8103-4399-b434-58bedae43fc5\") " Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.422282 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df904bdc-a504-4deb-8dd2-d2acf7626200-scripts\") pod \"df904bdc-a504-4deb-8dd2-d2acf7626200\" (UID: \"df904bdc-a504-4deb-8dd2-d2acf7626200\") " Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.422460 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b6e30c4b-8103-4399-b434-58bedae43fc5-dns-svc\") pod \"b6e30c4b-8103-4399-b434-58bedae43fc5\" (UID: \"b6e30c4b-8103-4399-b434-58bedae43fc5\") " Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.422823 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df904bdc-a504-4deb-8dd2-d2acf7626200-logs" (OuterVolumeSpecName: "logs") pod "df904bdc-a504-4deb-8dd2-d2acf7626200" (UID: "df904bdc-a504-4deb-8dd2-d2acf7626200"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.435696 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df904bdc-a504-4deb-8dd2-d2acf7626200-kube-api-access-g42m7" (OuterVolumeSpecName: "kube-api-access-g42m7") pod "df904bdc-a504-4deb-8dd2-d2acf7626200" (UID: "df904bdc-a504-4deb-8dd2-d2acf7626200"). InnerVolumeSpecName "kube-api-access-g42m7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.436288 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df904bdc-a504-4deb-8dd2-d2acf7626200-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "df904bdc-a504-4deb-8dd2-d2acf7626200" (UID: "df904bdc-a504-4deb-8dd2-d2acf7626200"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.436396 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.440486 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df904bdc-a504-4deb-8dd2-d2acf7626200-logs\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.440541 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g42m7\" (UniqueName: \"kubernetes.io/projected/df904bdc-a504-4deb-8dd2-d2acf7626200-kube-api-access-g42m7\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.440553 4809 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/df904bdc-a504-4deb-8dd2-d2acf7626200-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.455079 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55f844cf75-58svx" podStartSLOduration=10.455054123 podStartE2EDuration="10.455054123s" podCreationTimestamp="2025-12-06 06:16:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:17:07.441530289 +0000 UTC m=+1552.330513231" watchObservedRunningTime="2025-12-06 06:17:07.455054123 +0000 UTC m=+1552.344037155" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.456290 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df904bdc-a504-4deb-8dd2-d2acf7626200-scripts" (OuterVolumeSpecName: "scripts") pod "df904bdc-a504-4deb-8dd2-d2acf7626200" (UID: "df904bdc-a504-4deb-8dd2-d2acf7626200"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.460352 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6e30c4b-8103-4399-b434-58bedae43fc5-kube-api-access-698pg" (OuterVolumeSpecName: "kube-api-access-698pg") pod "b6e30c4b-8103-4399-b434-58bedae43fc5" (UID: "b6e30c4b-8103-4399-b434-58bedae43fc5"). InnerVolumeSpecName "kube-api-access-698pg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.476627 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "df904bdc-a504-4deb-8dd2-d2acf7626200" (UID: "df904bdc-a504-4deb-8dd2-d2acf7626200"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.496160 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.509261 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-wxlvd" event={"ID":"b6e30c4b-8103-4399-b434-58bedae43fc5","Type":"ContainerDied","Data":"a3d0de81ce3713c8508abf317d9721e0b720ace598ca136ca7e0f7781bcc148a"} Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.509316 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6bb4ccfdc4-9jqw5"] Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.509338 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"df904bdc-a504-4deb-8dd2-d2acf7626200","Type":"ContainerDied","Data":"b26c1ce80fa736673bf48ee790d716702359b3240f0ea5cef6a82299ac840a37"} Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.509371 4809 scope.go:117] "RemoveContainer" containerID="dcaf40a21a826a0263dc333602e07e911b5ada26fa16b17644edb44c00f50d88" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.542756 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a-combined-ca-bundle\") pod \"keystone-6bb4ccfdc4-9jqw5\" (UID: \"b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a\") " pod="openstack/keystone-6bb4ccfdc4-9jqw5" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.542859 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a-fernet-keys\") pod \"keystone-6bb4ccfdc4-9jqw5\" (UID: \"b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a\") " pod="openstack/keystone-6bb4ccfdc4-9jqw5" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.542962 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a-internal-tls-certs\") pod \"keystone-6bb4ccfdc4-9jqw5\" (UID: \"b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a\") " pod="openstack/keystone-6bb4ccfdc4-9jqw5" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.543000 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a-credential-keys\") pod \"keystone-6bb4ccfdc4-9jqw5\" (UID: \"b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a\") " pod="openstack/keystone-6bb4ccfdc4-9jqw5" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.543112 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a-scripts\") pod \"keystone-6bb4ccfdc4-9jqw5\" (UID: \"b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a\") " pod="openstack/keystone-6bb4ccfdc4-9jqw5" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.543154 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ksjz\" (UniqueName: \"kubernetes.io/projected/b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a-kube-api-access-6ksjz\") pod \"keystone-6bb4ccfdc4-9jqw5\" (UID: \"b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a\") " pod="openstack/keystone-6bb4ccfdc4-9jqw5" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.543179 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a-public-tls-certs\") pod \"keystone-6bb4ccfdc4-9jqw5\" (UID: \"b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a\") " pod="openstack/keystone-6bb4ccfdc4-9jqw5" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.543267 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a-config-data\") pod \"keystone-6bb4ccfdc4-9jqw5\" (UID: \"b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a\") " pod="openstack/keystone-6bb4ccfdc4-9jqw5" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.543449 4809 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.543470 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df904bdc-a504-4deb-8dd2-d2acf7626200-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.543480 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-698pg\" (UniqueName: \"kubernetes.io/projected/b6e30c4b-8103-4399-b434-58bedae43fc5-kube-api-access-698pg\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.543769 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6e30c4b-8103-4399-b434-58bedae43fc5-config" (OuterVolumeSpecName: "config") pod "b6e30c4b-8103-4399-b434-58bedae43fc5" (UID: "b6e30c4b-8103-4399-b434-58bedae43fc5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.577671 4809 scope.go:117] "RemoveContainer" containerID="87f24a96e341bc224827b16c9293236499ae758fa60cd7ad890fa2e69a045e17" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.591509 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6e30c4b-8103-4399-b434-58bedae43fc5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b6e30c4b-8103-4399-b434-58bedae43fc5" (UID: "b6e30c4b-8103-4399-b434-58bedae43fc5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.611063 4809 scope.go:117] "RemoveContainer" containerID="2300381bc4530f33d016676c2d9022f547c1d0cf3273d47b5bdc4ca71196cbef" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.615285 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df904bdc-a504-4deb-8dd2-d2acf7626200-config-data" (OuterVolumeSpecName: "config-data") pod "df904bdc-a504-4deb-8dd2-d2acf7626200" (UID: "df904bdc-a504-4deb-8dd2-d2acf7626200"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.629205 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6e30c4b-8103-4399-b434-58bedae43fc5-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b6e30c4b-8103-4399-b434-58bedae43fc5" (UID: "b6e30c4b-8103-4399-b434-58bedae43fc5"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.642598 4809 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.644795 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6gh5\" (UniqueName: \"kubernetes.io/projected/5b6f4174-473c-4e12-bcbb-7dd6ab7633ee-kube-api-access-v6gh5\") pod \"5b6f4174-473c-4e12-bcbb-7dd6ab7633ee\" (UID: \"5b6f4174-473c-4e12-bcbb-7dd6ab7633ee\") " Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.644876 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b6f4174-473c-4e12-bcbb-7dd6ab7633ee-config-data\") pod \"5b6f4174-473c-4e12-bcbb-7dd6ab7633ee\" (UID: \"5b6f4174-473c-4e12-bcbb-7dd6ab7633ee\") " Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.644950 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5b6f4174-473c-4e12-bcbb-7dd6ab7633ee-httpd-run\") pod \"5b6f4174-473c-4e12-bcbb-7dd6ab7633ee\" (UID: \"5b6f4174-473c-4e12-bcbb-7dd6ab7633ee\") " Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.644978 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b6f4174-473c-4e12-bcbb-7dd6ab7633ee-combined-ca-bundle\") pod \"5b6f4174-473c-4e12-bcbb-7dd6ab7633ee\" (UID: \"5b6f4174-473c-4e12-bcbb-7dd6ab7633ee\") " Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.644997 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b6f4174-473c-4e12-bcbb-7dd6ab7633ee-logs\") pod \"5b6f4174-473c-4e12-bcbb-7dd6ab7633ee\" (UID: \"5b6f4174-473c-4e12-bcbb-7dd6ab7633ee\") " Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.645201 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"5b6f4174-473c-4e12-bcbb-7dd6ab7633ee\" (UID: \"5b6f4174-473c-4e12-bcbb-7dd6ab7633ee\") " Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.645255 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b6f4174-473c-4e12-bcbb-7dd6ab7633ee-scripts\") pod \"5b6f4174-473c-4e12-bcbb-7dd6ab7633ee\" (UID: \"5b6f4174-473c-4e12-bcbb-7dd6ab7633ee\") " Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.645550 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a-fernet-keys\") pod \"keystone-6bb4ccfdc4-9jqw5\" (UID: \"b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a\") " pod="openstack/keystone-6bb4ccfdc4-9jqw5" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.645603 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a-internal-tls-certs\") pod \"keystone-6bb4ccfdc4-9jqw5\" (UID: \"b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a\") " pod="openstack/keystone-6bb4ccfdc4-9jqw5" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.645626 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a-credential-keys\") pod \"keystone-6bb4ccfdc4-9jqw5\" (UID: \"b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a\") " pod="openstack/keystone-6bb4ccfdc4-9jqw5" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.645684 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a-scripts\") pod \"keystone-6bb4ccfdc4-9jqw5\" (UID: \"b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a\") " pod="openstack/keystone-6bb4ccfdc4-9jqw5" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.645718 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ksjz\" (UniqueName: \"kubernetes.io/projected/b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a-kube-api-access-6ksjz\") pod \"keystone-6bb4ccfdc4-9jqw5\" (UID: \"b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a\") " pod="openstack/keystone-6bb4ccfdc4-9jqw5" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.645746 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a-public-tls-certs\") pod \"keystone-6bb4ccfdc4-9jqw5\" (UID: \"b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a\") " pod="openstack/keystone-6bb4ccfdc4-9jqw5" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.645813 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a-config-data\") pod \"keystone-6bb4ccfdc4-9jqw5\" (UID: \"b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a\") " pod="openstack/keystone-6bb4ccfdc4-9jqw5" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.645868 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a-combined-ca-bundle\") pod \"keystone-6bb4ccfdc4-9jqw5\" (UID: \"b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a\") " pod="openstack/keystone-6bb4ccfdc4-9jqw5" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.645921 4809 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.645946 4809 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b6e30c4b-8103-4399-b434-58bedae43fc5-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.645957 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b6e30c4b-8103-4399-b434-58bedae43fc5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.645967 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6e30c4b-8103-4399-b434-58bedae43fc5-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.645976 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df904bdc-a504-4deb-8dd2-d2acf7626200-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.645548 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b6f4174-473c-4e12-bcbb-7dd6ab7633ee-logs" (OuterVolumeSpecName: "logs") pod "5b6f4174-473c-4e12-bcbb-7dd6ab7633ee" (UID: "5b6f4174-473c-4e12-bcbb-7dd6ab7633ee"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.648241 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b6f4174-473c-4e12-bcbb-7dd6ab7633ee-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "5b6f4174-473c-4e12-bcbb-7dd6ab7633ee" (UID: "5b6f4174-473c-4e12-bcbb-7dd6ab7633ee"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.648606 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b6f4174-473c-4e12-bcbb-7dd6ab7633ee-kube-api-access-v6gh5" (OuterVolumeSpecName: "kube-api-access-v6gh5") pod "5b6f4174-473c-4e12-bcbb-7dd6ab7633ee" (UID: "5b6f4174-473c-4e12-bcbb-7dd6ab7633ee"). InnerVolumeSpecName "kube-api-access-v6gh5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.653550 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a-public-tls-certs\") pod \"keystone-6bb4ccfdc4-9jqw5\" (UID: \"b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a\") " pod="openstack/keystone-6bb4ccfdc4-9jqw5" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.653743 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a-credential-keys\") pod \"keystone-6bb4ccfdc4-9jqw5\" (UID: \"b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a\") " pod="openstack/keystone-6bb4ccfdc4-9jqw5" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.654101 4809 scope.go:117] "RemoveContainer" containerID="58243ecc3fcf5a0158d7575c80f30f61d95ee72566dafb41c1a55d282ff97ec8" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.654153 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a-internal-tls-certs\") pod \"keystone-6bb4ccfdc4-9jqw5\" (UID: \"b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a\") " pod="openstack/keystone-6bb4ccfdc4-9jqw5" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.654254 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a-combined-ca-bundle\") pod \"keystone-6bb4ccfdc4-9jqw5\" (UID: \"b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a\") " pod="openstack/keystone-6bb4ccfdc4-9jqw5" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.656315 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a-fernet-keys\") pod \"keystone-6bb4ccfdc4-9jqw5\" (UID: \"b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a\") " pod="openstack/keystone-6bb4ccfdc4-9jqw5" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.656486 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "5b6f4174-473c-4e12-bcbb-7dd6ab7633ee" (UID: "5b6f4174-473c-4e12-bcbb-7dd6ab7633ee"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.657787 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a-config-data\") pod \"keystone-6bb4ccfdc4-9jqw5\" (UID: \"b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a\") " pod="openstack/keystone-6bb4ccfdc4-9jqw5" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.658227 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a-scripts\") pod \"keystone-6bb4ccfdc4-9jqw5\" (UID: \"b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a\") " pod="openstack/keystone-6bb4ccfdc4-9jqw5" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.658294 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df904bdc-a504-4deb-8dd2-d2acf7626200-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "df904bdc-a504-4deb-8dd2-d2acf7626200" (UID: "df904bdc-a504-4deb-8dd2-d2acf7626200"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.665866 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ksjz\" (UniqueName: \"kubernetes.io/projected/b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a-kube-api-access-6ksjz\") pod \"keystone-6bb4ccfdc4-9jqw5\" (UID: \"b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a\") " pod="openstack/keystone-6bb4ccfdc4-9jqw5" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.682669 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b6f4174-473c-4e12-bcbb-7dd6ab7633ee-scripts" (OuterVolumeSpecName: "scripts") pod "5b6f4174-473c-4e12-bcbb-7dd6ab7633ee" (UID: "5b6f4174-473c-4e12-bcbb-7dd6ab7633ee"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.690616 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6e30c4b-8103-4399-b434-58bedae43fc5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b6e30c4b-8103-4399-b434-58bedae43fc5" (UID: "b6e30c4b-8103-4399-b434-58bedae43fc5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.692280 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6e30c4b-8103-4399-b434-58bedae43fc5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b6e30c4b-8103-4399-b434-58bedae43fc5" (UID: "b6e30c4b-8103-4399-b434-58bedae43fc5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.724335 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b6f4174-473c-4e12-bcbb-7dd6ab7633ee-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5b6f4174-473c-4e12-bcbb-7dd6ab7633ee" (UID: "5b6f4174-473c-4e12-bcbb-7dd6ab7633ee"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.749489 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6bb4ccfdc4-9jqw5" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.750150 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b6e30c4b-8103-4399-b434-58bedae43fc5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.750193 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b6e30c4b-8103-4399-b434-58bedae43fc5-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.750202 4809 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5b6f4174-473c-4e12-bcbb-7dd6ab7633ee-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.750211 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b6f4174-473c-4e12-bcbb-7dd6ab7633ee-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.750221 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b6f4174-473c-4e12-bcbb-7dd6ab7633ee-logs\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.750230 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df904bdc-a504-4deb-8dd2-d2acf7626200-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.750250 4809 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.750262 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b6f4174-473c-4e12-bcbb-7dd6ab7633ee-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.750272 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6gh5\" (UniqueName: \"kubernetes.io/projected/5b6f4174-473c-4e12-bcbb-7dd6ab7633ee-kube-api-access-v6gh5\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.763519 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b6f4174-473c-4e12-bcbb-7dd6ab7633ee-config-data" (OuterVolumeSpecName: "config-data") pod "5b6f4174-473c-4e12-bcbb-7dd6ab7633ee" (UID: "5b6f4174-473c-4e12-bcbb-7dd6ab7633ee"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.791545 4809 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.852679 4809 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.852712 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b6f4174-473c-4e12-bcbb-7dd6ab7633ee-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.959100 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-wxlvd"] Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.975670 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-wxlvd"] Dec 06 06:17:07 crc kubenswrapper[4809]: I1206 06:17:07.993163 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.025962 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.038894 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 06:17:08 crc kubenswrapper[4809]: E1206 06:17:08.039472 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b6f4174-473c-4e12-bcbb-7dd6ab7633ee" containerName="glance-log" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.039499 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b6f4174-473c-4e12-bcbb-7dd6ab7633ee" containerName="glance-log" Dec 06 06:17:08 crc kubenswrapper[4809]: E1206 06:17:08.039547 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b6f4174-473c-4e12-bcbb-7dd6ab7633ee" containerName="glance-httpd" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.039557 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b6f4174-473c-4e12-bcbb-7dd6ab7633ee" containerName="glance-httpd" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.039818 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b6f4174-473c-4e12-bcbb-7dd6ab7633ee" containerName="glance-log" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.039851 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b6f4174-473c-4e12-bcbb-7dd6ab7633ee" containerName="glance-httpd" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.041217 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.043401 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.044134 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.058300 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.167392 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc087d56-1746-4ae2-a3d1-7a44f3084390-scripts\") pod \"glance-default-external-api-0\" (UID: \"dc087d56-1746-4ae2-a3d1-7a44f3084390\") " pod="openstack/glance-default-external-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.167462 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc087d56-1746-4ae2-a3d1-7a44f3084390-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"dc087d56-1746-4ae2-a3d1-7a44f3084390\") " pod="openstack/glance-default-external-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.167652 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dc087d56-1746-4ae2-a3d1-7a44f3084390-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"dc087d56-1746-4ae2-a3d1-7a44f3084390\") " pod="openstack/glance-default-external-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.167685 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc087d56-1746-4ae2-a3d1-7a44f3084390-logs\") pod \"glance-default-external-api-0\" (UID: \"dc087d56-1746-4ae2-a3d1-7a44f3084390\") " pod="openstack/glance-default-external-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.167738 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc087d56-1746-4ae2-a3d1-7a44f3084390-config-data\") pod \"glance-default-external-api-0\" (UID: \"dc087d56-1746-4ae2-a3d1-7a44f3084390\") " pod="openstack/glance-default-external-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.167763 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"dc087d56-1746-4ae2-a3d1-7a44f3084390\") " pod="openstack/glance-default-external-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.167779 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc087d56-1746-4ae2-a3d1-7a44f3084390-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"dc087d56-1746-4ae2-a3d1-7a44f3084390\") " pod="openstack/glance-default-external-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.167794 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gthwl\" (UniqueName: \"kubernetes.io/projected/dc087d56-1746-4ae2-a3d1-7a44f3084390-kube-api-access-gthwl\") pod \"glance-default-external-api-0\" (UID: \"dc087d56-1746-4ae2-a3d1-7a44f3084390\") " pod="openstack/glance-default-external-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.269424 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dc087d56-1746-4ae2-a3d1-7a44f3084390-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"dc087d56-1746-4ae2-a3d1-7a44f3084390\") " pod="openstack/glance-default-external-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.269479 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc087d56-1746-4ae2-a3d1-7a44f3084390-logs\") pod \"glance-default-external-api-0\" (UID: \"dc087d56-1746-4ae2-a3d1-7a44f3084390\") " pod="openstack/glance-default-external-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.269532 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc087d56-1746-4ae2-a3d1-7a44f3084390-config-data\") pod \"glance-default-external-api-0\" (UID: \"dc087d56-1746-4ae2-a3d1-7a44f3084390\") " pod="openstack/glance-default-external-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.269564 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"dc087d56-1746-4ae2-a3d1-7a44f3084390\") " pod="openstack/glance-default-external-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.269592 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc087d56-1746-4ae2-a3d1-7a44f3084390-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"dc087d56-1746-4ae2-a3d1-7a44f3084390\") " pod="openstack/glance-default-external-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.269614 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gthwl\" (UniqueName: \"kubernetes.io/projected/dc087d56-1746-4ae2-a3d1-7a44f3084390-kube-api-access-gthwl\") pod \"glance-default-external-api-0\" (UID: \"dc087d56-1746-4ae2-a3d1-7a44f3084390\") " pod="openstack/glance-default-external-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.269653 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc087d56-1746-4ae2-a3d1-7a44f3084390-scripts\") pod \"glance-default-external-api-0\" (UID: \"dc087d56-1746-4ae2-a3d1-7a44f3084390\") " pod="openstack/glance-default-external-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.269694 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc087d56-1746-4ae2-a3d1-7a44f3084390-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"dc087d56-1746-4ae2-a3d1-7a44f3084390\") " pod="openstack/glance-default-external-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.270456 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"dc087d56-1746-4ae2-a3d1-7a44f3084390\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-external-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.270503 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc087d56-1746-4ae2-a3d1-7a44f3084390-logs\") pod \"glance-default-external-api-0\" (UID: \"dc087d56-1746-4ae2-a3d1-7a44f3084390\") " pod="openstack/glance-default-external-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.270790 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dc087d56-1746-4ae2-a3d1-7a44f3084390-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"dc087d56-1746-4ae2-a3d1-7a44f3084390\") " pod="openstack/glance-default-external-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.276831 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc087d56-1746-4ae2-a3d1-7a44f3084390-scripts\") pod \"glance-default-external-api-0\" (UID: \"dc087d56-1746-4ae2-a3d1-7a44f3084390\") " pod="openstack/glance-default-external-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.277330 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc087d56-1746-4ae2-a3d1-7a44f3084390-config-data\") pod \"glance-default-external-api-0\" (UID: \"dc087d56-1746-4ae2-a3d1-7a44f3084390\") " pod="openstack/glance-default-external-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.277548 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6bb4ccfdc4-9jqw5"] Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.280064 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc087d56-1746-4ae2-a3d1-7a44f3084390-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"dc087d56-1746-4ae2-a3d1-7a44f3084390\") " pod="openstack/glance-default-external-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.281120 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc087d56-1746-4ae2-a3d1-7a44f3084390-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"dc087d56-1746-4ae2-a3d1-7a44f3084390\") " pod="openstack/glance-default-external-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.298576 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gthwl\" (UniqueName: \"kubernetes.io/projected/dc087d56-1746-4ae2-a3d1-7a44f3084390-kube-api-access-gthwl\") pod \"glance-default-external-api-0\" (UID: \"dc087d56-1746-4ae2-a3d1-7a44f3084390\") " pod="openstack/glance-default-external-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.360122 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"dc087d56-1746-4ae2-a3d1-7a44f3084390\") " pod="openstack/glance-default-external-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.449115 4809 generic.go:334] "Generic (PLEG): container finished" podID="662e24e4-5dd6-4d18-822c-2863017f15af" containerID="12a159c9d2e59d6ab6c567a3ac994fcd15176cf11ff241ba517b9ef132bec4bc" exitCode=0 Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.449237 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-4gxwk" event={"ID":"662e24e4-5dd6-4d18-822c-2863017f15af","Type":"ContainerDied","Data":"12a159c9d2e59d6ab6c567a3ac994fcd15176cf11ff241ba517b9ef132bec4bc"} Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.460239 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5b6f4174-473c-4e12-bcbb-7dd6ab7633ee","Type":"ContainerDied","Data":"5dda841c122fa4d761f1a9b61cd62e6392c4aa39121faab6151c5d97cfdd2eb2"} Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.460288 4809 scope.go:117] "RemoveContainer" containerID="5ab1c005b596383180356c364a3e4f5a45a30307e69f7e04a113bbc5382f1c40" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.460400 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.473659 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6bb4ccfdc4-9jqw5" event={"ID":"b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a","Type":"ContainerStarted","Data":"d6fbf318ad8159b614a3c61af198de299efeaaac8364f4cb524fe4bc4432ef3c"} Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.526976 4809 scope.go:117] "RemoveContainer" containerID="0e7b4374efdd9b613874cceea8597496abeca32c82b77e975e2722898bc35d40" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.560557 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.575086 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.596246 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.598813 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.601232 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.602522 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.603834 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.664820 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.684173 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jg5f7\" (UniqueName: \"kubernetes.io/projected/c40b4907-61d2-4ec7-b611-cb01d399c8d0-kube-api-access-jg5f7\") pod \"glance-default-internal-api-0\" (UID: \"c40b4907-61d2-4ec7-b611-cb01d399c8d0\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.684671 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c40b4907-61d2-4ec7-b611-cb01d399c8d0-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"c40b4907-61d2-4ec7-b611-cb01d399c8d0\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.684708 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"c40b4907-61d2-4ec7-b611-cb01d399c8d0\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.684775 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c40b4907-61d2-4ec7-b611-cb01d399c8d0-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c40b4907-61d2-4ec7-b611-cb01d399c8d0\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.684836 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c40b4907-61d2-4ec7-b611-cb01d399c8d0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c40b4907-61d2-4ec7-b611-cb01d399c8d0\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.684866 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c40b4907-61d2-4ec7-b611-cb01d399c8d0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c40b4907-61d2-4ec7-b611-cb01d399c8d0\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.685010 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c40b4907-61d2-4ec7-b611-cb01d399c8d0-logs\") pod \"glance-default-internal-api-0\" (UID: \"c40b4907-61d2-4ec7-b611-cb01d399c8d0\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.685078 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c40b4907-61d2-4ec7-b611-cb01d399c8d0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c40b4907-61d2-4ec7-b611-cb01d399c8d0\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.789702 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c40b4907-61d2-4ec7-b611-cb01d399c8d0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c40b4907-61d2-4ec7-b611-cb01d399c8d0\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.789997 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c40b4907-61d2-4ec7-b611-cb01d399c8d0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c40b4907-61d2-4ec7-b611-cb01d399c8d0\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.790102 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c40b4907-61d2-4ec7-b611-cb01d399c8d0-logs\") pod \"glance-default-internal-api-0\" (UID: \"c40b4907-61d2-4ec7-b611-cb01d399c8d0\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.790172 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c40b4907-61d2-4ec7-b611-cb01d399c8d0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c40b4907-61d2-4ec7-b611-cb01d399c8d0\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.790221 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jg5f7\" (UniqueName: \"kubernetes.io/projected/c40b4907-61d2-4ec7-b611-cb01d399c8d0-kube-api-access-jg5f7\") pod \"glance-default-internal-api-0\" (UID: \"c40b4907-61d2-4ec7-b611-cb01d399c8d0\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.790305 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c40b4907-61d2-4ec7-b611-cb01d399c8d0-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"c40b4907-61d2-4ec7-b611-cb01d399c8d0\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.790339 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"c40b4907-61d2-4ec7-b611-cb01d399c8d0\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.790404 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c40b4907-61d2-4ec7-b611-cb01d399c8d0-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c40b4907-61d2-4ec7-b611-cb01d399c8d0\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.793431 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"c40b4907-61d2-4ec7-b611-cb01d399c8d0\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-internal-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.793886 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c40b4907-61d2-4ec7-b611-cb01d399c8d0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c40b4907-61d2-4ec7-b611-cb01d399c8d0\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.793908 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c40b4907-61d2-4ec7-b611-cb01d399c8d0-logs\") pod \"glance-default-internal-api-0\" (UID: \"c40b4907-61d2-4ec7-b611-cb01d399c8d0\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.803907 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c40b4907-61d2-4ec7-b611-cb01d399c8d0-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c40b4907-61d2-4ec7-b611-cb01d399c8d0\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.804664 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c40b4907-61d2-4ec7-b611-cb01d399c8d0-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"c40b4907-61d2-4ec7-b611-cb01d399c8d0\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.804811 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c40b4907-61d2-4ec7-b611-cb01d399c8d0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c40b4907-61d2-4ec7-b611-cb01d399c8d0\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.819197 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jg5f7\" (UniqueName: \"kubernetes.io/projected/c40b4907-61d2-4ec7-b611-cb01d399c8d0-kube-api-access-jg5f7\") pod \"glance-default-internal-api-0\" (UID: \"c40b4907-61d2-4ec7-b611-cb01d399c8d0\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.820981 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c40b4907-61d2-4ec7-b611-cb01d399c8d0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c40b4907-61d2-4ec7-b611-cb01d399c8d0\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.871524 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"c40b4907-61d2-4ec7-b611-cb01d399c8d0\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:17:08 crc kubenswrapper[4809]: I1206 06:17:08.948492 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 06:17:09 crc kubenswrapper[4809]: I1206 06:17:09.350176 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 06:17:09 crc kubenswrapper[4809]: I1206 06:17:09.422723 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b6f4174-473c-4e12-bcbb-7dd6ab7633ee" path="/var/lib/kubelet/pods/5b6f4174-473c-4e12-bcbb-7dd6ab7633ee/volumes" Dec 06 06:17:09 crc kubenswrapper[4809]: I1206 06:17:09.426094 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6e30c4b-8103-4399-b434-58bedae43fc5" path="/var/lib/kubelet/pods/b6e30c4b-8103-4399-b434-58bedae43fc5/volumes" Dec 06 06:17:09 crc kubenswrapper[4809]: I1206 06:17:09.434999 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df904bdc-a504-4deb-8dd2-d2acf7626200" path="/var/lib/kubelet/pods/df904bdc-a504-4deb-8dd2-d2acf7626200/volumes" Dec 06 06:17:09 crc kubenswrapper[4809]: I1206 06:17:09.543278 4809 generic.go:334] "Generic (PLEG): container finished" podID="8336af68-51f9-4f3a-8c92-61dd1aba9067" containerID="cbb746c60ce6a9134f7e45e0a79d24227920a2c5bece823b7c07cef89c91e92a" exitCode=0 Dec 06 06:17:09 crc kubenswrapper[4809]: I1206 06:17:09.543382 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-m98hv" event={"ID":"8336af68-51f9-4f3a-8c92-61dd1aba9067","Type":"ContainerDied","Data":"cbb746c60ce6a9134f7e45e0a79d24227920a2c5bece823b7c07cef89c91e92a"} Dec 06 06:17:09 crc kubenswrapper[4809]: I1206 06:17:09.574221 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dc087d56-1746-4ae2-a3d1-7a44f3084390","Type":"ContainerStarted","Data":"aa0efffb726b2380651dccfca520dab8405c3b056961bbd1a8f7ffa1b06ef98f"} Dec 06 06:17:09 crc kubenswrapper[4809]: I1206 06:17:09.664759 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6bb4ccfdc4-9jqw5" event={"ID":"b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a","Type":"ContainerStarted","Data":"5348d33f163ae4f491818f7eba4f9e5372ef10dbbb4b4549b4b2942c6bd2c419"} Dec 06 06:17:09 crc kubenswrapper[4809]: I1206 06:17:09.664828 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-6bb4ccfdc4-9jqw5" Dec 06 06:17:09 crc kubenswrapper[4809]: I1206 06:17:09.728211 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-6bb4ccfdc4-9jqw5" podStartSLOduration=2.728191047 podStartE2EDuration="2.728191047s" podCreationTimestamp="2025-12-06 06:17:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:17:09.714635851 +0000 UTC m=+1554.603618793" watchObservedRunningTime="2025-12-06 06:17:09.728191047 +0000 UTC m=+1554.617173989" Dec 06 06:17:09 crc kubenswrapper[4809]: I1206 06:17:09.775196 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 06:17:10 crc kubenswrapper[4809]: I1206 06:17:10.677533 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dc087d56-1746-4ae2-a3d1-7a44f3084390","Type":"ContainerStarted","Data":"eceb1dd6a956119561296e96f6dd57c1a36fc552142dab032ead78002d98f409"} Dec 06 06:17:13 crc kubenswrapper[4809]: I1206 06:17:13.238171 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55f844cf75-58svx" Dec 06 06:17:13 crc kubenswrapper[4809]: I1206 06:17:13.313970 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-r89dg"] Dec 06 06:17:13 crc kubenswrapper[4809]: I1206 06:17:13.314203 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-r89dg" podUID="0c24a4f1-4cfb-4aae-ac2e-fa7680751c26" containerName="dnsmasq-dns" containerID="cri-o://43136fcc5a12ae304cd1d4d1dc504b5c5485dce32274385a16b0fa0c26d6de99" gracePeriod=10 Dec 06 06:17:13 crc kubenswrapper[4809]: W1206 06:17:13.777132 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc40b4907_61d2_4ec7_b611_cb01d399c8d0.slice/crio-7487ebf3d482625a35a71ae604e34b91a246b0fef4380fb68114da4b9bf4c673 WatchSource:0}: Error finding container 7487ebf3d482625a35a71ae604e34b91a246b0fef4380fb68114da4b9bf4c673: Status 404 returned error can't find the container with id 7487ebf3d482625a35a71ae604e34b91a246b0fef4380fb68114da4b9bf4c673 Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.218891 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-4gxwk" Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.228763 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-m98hv" Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.238566 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-r89dg" Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.367277 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c24a4f1-4cfb-4aae-ac2e-fa7680751c26-ovsdbserver-sb\") pod \"0c24a4f1-4cfb-4aae-ac2e-fa7680751c26\" (UID: \"0c24a4f1-4cfb-4aae-ac2e-fa7680751c26\") " Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.367362 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r96sr\" (UniqueName: \"kubernetes.io/projected/8336af68-51f9-4f3a-8c92-61dd1aba9067-kube-api-access-r96sr\") pod \"8336af68-51f9-4f3a-8c92-61dd1aba9067\" (UID: \"8336af68-51f9-4f3a-8c92-61dd1aba9067\") " Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.367422 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c24a4f1-4cfb-4aae-ac2e-fa7680751c26-config\") pod \"0c24a4f1-4cfb-4aae-ac2e-fa7680751c26\" (UID: \"0c24a4f1-4cfb-4aae-ac2e-fa7680751c26\") " Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.367518 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8336af68-51f9-4f3a-8c92-61dd1aba9067-logs\") pod \"8336af68-51f9-4f3a-8c92-61dd1aba9067\" (UID: \"8336af68-51f9-4f3a-8c92-61dd1aba9067\") " Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.367560 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qpp7j\" (UniqueName: \"kubernetes.io/projected/662e24e4-5dd6-4d18-822c-2863017f15af-kube-api-access-qpp7j\") pod \"662e24e4-5dd6-4d18-822c-2863017f15af\" (UID: \"662e24e4-5dd6-4d18-822c-2863017f15af\") " Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.367611 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8336af68-51f9-4f3a-8c92-61dd1aba9067-scripts\") pod \"8336af68-51f9-4f3a-8c92-61dd1aba9067\" (UID: \"8336af68-51f9-4f3a-8c92-61dd1aba9067\") " Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.367640 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5c2k\" (UniqueName: \"kubernetes.io/projected/0c24a4f1-4cfb-4aae-ac2e-fa7680751c26-kube-api-access-v5c2k\") pod \"0c24a4f1-4cfb-4aae-ac2e-fa7680751c26\" (UID: \"0c24a4f1-4cfb-4aae-ac2e-fa7680751c26\") " Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.367740 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/662e24e4-5dd6-4d18-822c-2863017f15af-combined-ca-bundle\") pod \"662e24e4-5dd6-4d18-822c-2863017f15af\" (UID: \"662e24e4-5dd6-4d18-822c-2863017f15af\") " Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.367775 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c24a4f1-4cfb-4aae-ac2e-fa7680751c26-ovsdbserver-nb\") pod \"0c24a4f1-4cfb-4aae-ac2e-fa7680751c26\" (UID: \"0c24a4f1-4cfb-4aae-ac2e-fa7680751c26\") " Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.367814 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/662e24e4-5dd6-4d18-822c-2863017f15af-db-sync-config-data\") pod \"662e24e4-5dd6-4d18-822c-2863017f15af\" (UID: \"662e24e4-5dd6-4d18-822c-2863017f15af\") " Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.367860 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8336af68-51f9-4f3a-8c92-61dd1aba9067-config-data\") pod \"8336af68-51f9-4f3a-8c92-61dd1aba9067\" (UID: \"8336af68-51f9-4f3a-8c92-61dd1aba9067\") " Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.367913 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c24a4f1-4cfb-4aae-ac2e-fa7680751c26-dns-svc\") pod \"0c24a4f1-4cfb-4aae-ac2e-fa7680751c26\" (UID: \"0c24a4f1-4cfb-4aae-ac2e-fa7680751c26\") " Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.367968 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8336af68-51f9-4f3a-8c92-61dd1aba9067-combined-ca-bundle\") pod \"8336af68-51f9-4f3a-8c92-61dd1aba9067\" (UID: \"8336af68-51f9-4f3a-8c92-61dd1aba9067\") " Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.370163 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8336af68-51f9-4f3a-8c92-61dd1aba9067-logs" (OuterVolumeSpecName: "logs") pod "8336af68-51f9-4f3a-8c92-61dd1aba9067" (UID: "8336af68-51f9-4f3a-8c92-61dd1aba9067"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.372681 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8336af68-51f9-4f3a-8c92-61dd1aba9067-kube-api-access-r96sr" (OuterVolumeSpecName: "kube-api-access-r96sr") pod "8336af68-51f9-4f3a-8c92-61dd1aba9067" (UID: "8336af68-51f9-4f3a-8c92-61dd1aba9067"). InnerVolumeSpecName "kube-api-access-r96sr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.379837 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c24a4f1-4cfb-4aae-ac2e-fa7680751c26-kube-api-access-v5c2k" (OuterVolumeSpecName: "kube-api-access-v5c2k") pod "0c24a4f1-4cfb-4aae-ac2e-fa7680751c26" (UID: "0c24a4f1-4cfb-4aae-ac2e-fa7680751c26"). InnerVolumeSpecName "kube-api-access-v5c2k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.381501 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/662e24e4-5dd6-4d18-822c-2863017f15af-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "662e24e4-5dd6-4d18-822c-2863017f15af" (UID: "662e24e4-5dd6-4d18-822c-2863017f15af"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.382966 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8336af68-51f9-4f3a-8c92-61dd1aba9067-scripts" (OuterVolumeSpecName: "scripts") pod "8336af68-51f9-4f3a-8c92-61dd1aba9067" (UID: "8336af68-51f9-4f3a-8c92-61dd1aba9067"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.392948 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/662e24e4-5dd6-4d18-822c-2863017f15af-kube-api-access-qpp7j" (OuterVolumeSpecName: "kube-api-access-qpp7j") pod "662e24e4-5dd6-4d18-822c-2863017f15af" (UID: "662e24e4-5dd6-4d18-822c-2863017f15af"). InnerVolumeSpecName "kube-api-access-qpp7j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.433629 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8336af68-51f9-4f3a-8c92-61dd1aba9067-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8336af68-51f9-4f3a-8c92-61dd1aba9067" (UID: "8336af68-51f9-4f3a-8c92-61dd1aba9067"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.471643 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r96sr\" (UniqueName: \"kubernetes.io/projected/8336af68-51f9-4f3a-8c92-61dd1aba9067-kube-api-access-r96sr\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.471676 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8336af68-51f9-4f3a-8c92-61dd1aba9067-logs\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.471689 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qpp7j\" (UniqueName: \"kubernetes.io/projected/662e24e4-5dd6-4d18-822c-2863017f15af-kube-api-access-qpp7j\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.471779 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8336af68-51f9-4f3a-8c92-61dd1aba9067-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.471789 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5c2k\" (UniqueName: \"kubernetes.io/projected/0c24a4f1-4cfb-4aae-ac2e-fa7680751c26-kube-api-access-v5c2k\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.471798 4809 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/662e24e4-5dd6-4d18-822c-2863017f15af-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.471808 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8336af68-51f9-4f3a-8c92-61dd1aba9067-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.533882 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/662e24e4-5dd6-4d18-822c-2863017f15af-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "662e24e4-5dd6-4d18-822c-2863017f15af" (UID: "662e24e4-5dd6-4d18-822c-2863017f15af"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.574508 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/662e24e4-5dd6-4d18-822c-2863017f15af-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.630377 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8336af68-51f9-4f3a-8c92-61dd1aba9067-config-data" (OuterVolumeSpecName: "config-data") pod "8336af68-51f9-4f3a-8c92-61dd1aba9067" (UID: "8336af68-51f9-4f3a-8c92-61dd1aba9067"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.676125 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8336af68-51f9-4f3a-8c92-61dd1aba9067-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.684373 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c24a4f1-4cfb-4aae-ac2e-fa7680751c26-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0c24a4f1-4cfb-4aae-ac2e-fa7680751c26" (UID: "0c24a4f1-4cfb-4aae-ac2e-fa7680751c26"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.696975 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c24a4f1-4cfb-4aae-ac2e-fa7680751c26-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0c24a4f1-4cfb-4aae-ac2e-fa7680751c26" (UID: "0c24a4f1-4cfb-4aae-ac2e-fa7680751c26"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.701106 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c24a4f1-4cfb-4aae-ac2e-fa7680751c26-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0c24a4f1-4cfb-4aae-ac2e-fa7680751c26" (UID: "0c24a4f1-4cfb-4aae-ac2e-fa7680751c26"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.702194 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c24a4f1-4cfb-4aae-ac2e-fa7680751c26-config" (OuterVolumeSpecName: "config") pod "0c24a4f1-4cfb-4aae-ac2e-fa7680751c26" (UID: "0c24a4f1-4cfb-4aae-ac2e-fa7680751c26"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.751845 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-4gxwk" event={"ID":"662e24e4-5dd6-4d18-822c-2863017f15af","Type":"ContainerDied","Data":"a5bb6b2166abf5b42fe39fdb1928631a4d7d9d596113bc95699f0ce321b401c9"} Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.751890 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a5bb6b2166abf5b42fe39fdb1928631a4d7d9d596113bc95699f0ce321b401c9" Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.751887 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-4gxwk" Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.755165 4809 generic.go:334] "Generic (PLEG): container finished" podID="edf1ab66-b483-4172-81fd-dd3f3b9b44aa" containerID="30b14dfc79cb4d967b8537d39fe468fc1409477b840153c0872891dfe7ab084b" exitCode=0 Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.755235 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-ckhg9" event={"ID":"edf1ab66-b483-4172-81fd-dd3f3b9b44aa","Type":"ContainerDied","Data":"30b14dfc79cb4d967b8537d39fe468fc1409477b840153c0872891dfe7ab084b"} Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.757639 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c40b4907-61d2-4ec7-b611-cb01d399c8d0","Type":"ContainerStarted","Data":"7487ebf3d482625a35a71ae604e34b91a246b0fef4380fb68114da4b9bf4c673"} Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.763865 4809 generic.go:334] "Generic (PLEG): container finished" podID="0c24a4f1-4cfb-4aae-ac2e-fa7680751c26" containerID="43136fcc5a12ae304cd1d4d1dc504b5c5485dce32274385a16b0fa0c26d6de99" exitCode=0 Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.763959 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-r89dg" event={"ID":"0c24a4f1-4cfb-4aae-ac2e-fa7680751c26","Type":"ContainerDied","Data":"43136fcc5a12ae304cd1d4d1dc504b5c5485dce32274385a16b0fa0c26d6de99"} Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.763994 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-r89dg" event={"ID":"0c24a4f1-4cfb-4aae-ac2e-fa7680751c26","Type":"ContainerDied","Data":"c6d176c838ee7982f8dba1d9a555740fc1ea025201e0fd30b602de6d5b97fcb9"} Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.764016 4809 scope.go:117] "RemoveContainer" containerID="43136fcc5a12ae304cd1d4d1dc504b5c5485dce32274385a16b0fa0c26d6de99" Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.764042 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-r89dg" Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.772901 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"93e2d67f-9481-4ca3-83e7-4f7c0673b9d4","Type":"ContainerStarted","Data":"04ae50b6ffa0e470d79b81e343165917548537b8df78e6568beeaa62cb16d06e"} Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.782009 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c24a4f1-4cfb-4aae-ac2e-fa7680751c26-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.782333 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c24a4f1-4cfb-4aae-ac2e-fa7680751c26-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.782347 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c24a4f1-4cfb-4aae-ac2e-fa7680751c26-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.782359 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c24a4f1-4cfb-4aae-ac2e-fa7680751c26-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.796259 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-m98hv" event={"ID":"8336af68-51f9-4f3a-8c92-61dd1aba9067","Type":"ContainerDied","Data":"e670b82e7f278746dda04ed70357e80a34f69e6381934a87d255e3f054730829"} Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.796307 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e670b82e7f278746dda04ed70357e80a34f69e6381934a87d255e3f054730829" Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.796386 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-m98hv" Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.808041 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dc087d56-1746-4ae2-a3d1-7a44f3084390","Type":"ContainerStarted","Data":"76b82b56f3e1173ce4aabac3674d4645c6199e126e35b484ef68d60197933ced"} Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.838342 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=7.838321776 podStartE2EDuration="7.838321776s" podCreationTimestamp="2025-12-06 06:17:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:17:14.830581747 +0000 UTC m=+1559.719564709" watchObservedRunningTime="2025-12-06 06:17:14.838321776 +0000 UTC m=+1559.727304718" Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.926231 4809 scope.go:117] "RemoveContainer" containerID="66bc7af1b8f7b0ea275598143ea77ba9570c3706b7d8f18d1b05c2d5b7f4b7a5" Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.949311 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-r89dg"] Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.965091 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-r89dg"] Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.977872 4809 scope.go:117] "RemoveContainer" containerID="43136fcc5a12ae304cd1d4d1dc504b5c5485dce32274385a16b0fa0c26d6de99" Dec 06 06:17:14 crc kubenswrapper[4809]: E1206 06:17:14.978592 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43136fcc5a12ae304cd1d4d1dc504b5c5485dce32274385a16b0fa0c26d6de99\": container with ID starting with 43136fcc5a12ae304cd1d4d1dc504b5c5485dce32274385a16b0fa0c26d6de99 not found: ID does not exist" containerID="43136fcc5a12ae304cd1d4d1dc504b5c5485dce32274385a16b0fa0c26d6de99" Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.978636 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43136fcc5a12ae304cd1d4d1dc504b5c5485dce32274385a16b0fa0c26d6de99"} err="failed to get container status \"43136fcc5a12ae304cd1d4d1dc504b5c5485dce32274385a16b0fa0c26d6de99\": rpc error: code = NotFound desc = could not find container \"43136fcc5a12ae304cd1d4d1dc504b5c5485dce32274385a16b0fa0c26d6de99\": container with ID starting with 43136fcc5a12ae304cd1d4d1dc504b5c5485dce32274385a16b0fa0c26d6de99 not found: ID does not exist" Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.978658 4809 scope.go:117] "RemoveContainer" containerID="66bc7af1b8f7b0ea275598143ea77ba9570c3706b7d8f18d1b05c2d5b7f4b7a5" Dec 06 06:17:14 crc kubenswrapper[4809]: E1206 06:17:14.978976 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66bc7af1b8f7b0ea275598143ea77ba9570c3706b7d8f18d1b05c2d5b7f4b7a5\": container with ID starting with 66bc7af1b8f7b0ea275598143ea77ba9570c3706b7d8f18d1b05c2d5b7f4b7a5 not found: ID does not exist" containerID="66bc7af1b8f7b0ea275598143ea77ba9570c3706b7d8f18d1b05c2d5b7f4b7a5" Dec 06 06:17:14 crc kubenswrapper[4809]: I1206 06:17:14.979040 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66bc7af1b8f7b0ea275598143ea77ba9570c3706b7d8f18d1b05c2d5b7f4b7a5"} err="failed to get container status \"66bc7af1b8f7b0ea275598143ea77ba9570c3706b7d8f18d1b05c2d5b7f4b7a5\": rpc error: code = NotFound desc = could not find container \"66bc7af1b8f7b0ea275598143ea77ba9570c3706b7d8f18d1b05c2d5b7f4b7a5\": container with ID starting with 66bc7af1b8f7b0ea275598143ea77ba9570c3706b7d8f18d1b05c2d5b7f4b7a5 not found: ID does not exist" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.423853 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c24a4f1-4cfb-4aae-ac2e-fa7680751c26" path="/var/lib/kubelet/pods/0c24a4f1-4cfb-4aae-ac2e-fa7680751c26/volumes" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.474805 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-5fc7d5fdc6-llxdt"] Dec 06 06:17:15 crc kubenswrapper[4809]: E1206 06:17:15.475210 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c24a4f1-4cfb-4aae-ac2e-fa7680751c26" containerName="dnsmasq-dns" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.475238 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c24a4f1-4cfb-4aae-ac2e-fa7680751c26" containerName="dnsmasq-dns" Dec 06 06:17:15 crc kubenswrapper[4809]: E1206 06:17:15.475259 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8336af68-51f9-4f3a-8c92-61dd1aba9067" containerName="placement-db-sync" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.475268 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8336af68-51f9-4f3a-8c92-61dd1aba9067" containerName="placement-db-sync" Dec 06 06:17:15 crc kubenswrapper[4809]: E1206 06:17:15.475313 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c24a4f1-4cfb-4aae-ac2e-fa7680751c26" containerName="init" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.475330 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c24a4f1-4cfb-4aae-ac2e-fa7680751c26" containerName="init" Dec 06 06:17:15 crc kubenswrapper[4809]: E1206 06:17:15.475342 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="662e24e4-5dd6-4d18-822c-2863017f15af" containerName="barbican-db-sync" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.475350 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="662e24e4-5dd6-4d18-822c-2863017f15af" containerName="barbican-db-sync" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.475603 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="8336af68-51f9-4f3a-8c92-61dd1aba9067" containerName="placement-db-sync" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.475648 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="662e24e4-5dd6-4d18-822c-2863017f15af" containerName="barbican-db-sync" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.475666 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c24a4f1-4cfb-4aae-ac2e-fa7680751c26" containerName="dnsmasq-dns" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.477897 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5fc7d5fdc6-llxdt"] Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.478045 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5fc7d5fdc6-llxdt" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.485645 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.485888 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.486459 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.486487 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.491217 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-vjw2g" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.503036 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-646ffc98d7-5l7d6"] Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.510409 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-646ffc98d7-5l7d6" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.515465 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-s2gvb" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.515659 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.515808 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.523584 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-646ffc98d7-5l7d6"] Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.631247 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/79161677-35e4-41b3-8fb4-f2bd4beebe60-config-data-custom\") pod \"barbican-worker-646ffc98d7-5l7d6\" (UID: \"79161677-35e4-41b3-8fb4-f2bd4beebe60\") " pod="openstack/barbican-worker-646ffc98d7-5l7d6" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.631327 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nds75\" (UniqueName: \"kubernetes.io/projected/b8ba7416-80be-4115-b960-58d147b56e9d-kube-api-access-nds75\") pod \"placement-5fc7d5fdc6-llxdt\" (UID: \"b8ba7416-80be-4115-b960-58d147b56e9d\") " pod="openstack/placement-5fc7d5fdc6-llxdt" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.631521 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79161677-35e4-41b3-8fb4-f2bd4beebe60-combined-ca-bundle\") pod \"barbican-worker-646ffc98d7-5l7d6\" (UID: \"79161677-35e4-41b3-8fb4-f2bd4beebe60\") " pod="openstack/barbican-worker-646ffc98d7-5l7d6" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.631728 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b8ba7416-80be-4115-b960-58d147b56e9d-internal-tls-certs\") pod \"placement-5fc7d5fdc6-llxdt\" (UID: \"b8ba7416-80be-4115-b960-58d147b56e9d\") " pod="openstack/placement-5fc7d5fdc6-llxdt" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.631794 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8ba7416-80be-4115-b960-58d147b56e9d-scripts\") pod \"placement-5fc7d5fdc6-llxdt\" (UID: \"b8ba7416-80be-4115-b960-58d147b56e9d\") " pod="openstack/placement-5fc7d5fdc6-llxdt" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.631889 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b8ba7416-80be-4115-b960-58d147b56e9d-logs\") pod \"placement-5fc7d5fdc6-llxdt\" (UID: \"b8ba7416-80be-4115-b960-58d147b56e9d\") " pod="openstack/placement-5fc7d5fdc6-llxdt" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.632113 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8ba7416-80be-4115-b960-58d147b56e9d-combined-ca-bundle\") pod \"placement-5fc7d5fdc6-llxdt\" (UID: \"b8ba7416-80be-4115-b960-58d147b56e9d\") " pod="openstack/placement-5fc7d5fdc6-llxdt" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.632237 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79161677-35e4-41b3-8fb4-f2bd4beebe60-logs\") pod \"barbican-worker-646ffc98d7-5l7d6\" (UID: \"79161677-35e4-41b3-8fb4-f2bd4beebe60\") " pod="openstack/barbican-worker-646ffc98d7-5l7d6" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.632366 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8ba7416-80be-4115-b960-58d147b56e9d-config-data\") pod \"placement-5fc7d5fdc6-llxdt\" (UID: \"b8ba7416-80be-4115-b960-58d147b56e9d\") " pod="openstack/placement-5fc7d5fdc6-llxdt" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.639544 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-5cdb89b766-6mc5c"] Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.649999 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5cdb89b766-6mc5c" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.652826 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79161677-35e4-41b3-8fb4-f2bd4beebe60-config-data\") pod \"barbican-worker-646ffc98d7-5l7d6\" (UID: \"79161677-35e4-41b3-8fb4-f2bd4beebe60\") " pod="openstack/barbican-worker-646ffc98d7-5l7d6" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.652953 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b8ba7416-80be-4115-b960-58d147b56e9d-public-tls-certs\") pod \"placement-5fc7d5fdc6-llxdt\" (UID: \"b8ba7416-80be-4115-b960-58d147b56e9d\") " pod="openstack/placement-5fc7d5fdc6-llxdt" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.653094 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58z6l\" (UniqueName: \"kubernetes.io/projected/79161677-35e4-41b3-8fb4-f2bd4beebe60-kube-api-access-58z6l\") pod \"barbican-worker-646ffc98d7-5l7d6\" (UID: \"79161677-35e4-41b3-8fb4-f2bd4beebe60\") " pod="openstack/barbican-worker-646ffc98d7-5l7d6" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.657012 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.682087 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5cdb89b766-6mc5c"] Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.706440 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-k4rfq"] Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.709043 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-k4rfq" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.755568 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnqcz\" (UniqueName: \"kubernetes.io/projected/1d925ed8-47dc-4bf2-a476-2526d257d5d0-kube-api-access-rnqcz\") pod \"barbican-keystone-listener-5cdb89b766-6mc5c\" (UID: \"1d925ed8-47dc-4bf2-a476-2526d257d5d0\") " pod="openstack/barbican-keystone-listener-5cdb89b766-6mc5c" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.755708 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79161677-35e4-41b3-8fb4-f2bd4beebe60-combined-ca-bundle\") pod \"barbican-worker-646ffc98d7-5l7d6\" (UID: \"79161677-35e4-41b3-8fb4-f2bd4beebe60\") " pod="openstack/barbican-worker-646ffc98d7-5l7d6" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.755772 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d925ed8-47dc-4bf2-a476-2526d257d5d0-combined-ca-bundle\") pod \"barbican-keystone-listener-5cdb89b766-6mc5c\" (UID: \"1d925ed8-47dc-4bf2-a476-2526d257d5d0\") " pod="openstack/barbican-keystone-listener-5cdb89b766-6mc5c" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.755908 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b8ba7416-80be-4115-b960-58d147b56e9d-internal-tls-certs\") pod \"placement-5fc7d5fdc6-llxdt\" (UID: \"b8ba7416-80be-4115-b960-58d147b56e9d\") " pod="openstack/placement-5fc7d5fdc6-llxdt" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.755949 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8ba7416-80be-4115-b960-58d147b56e9d-scripts\") pod \"placement-5fc7d5fdc6-llxdt\" (UID: \"b8ba7416-80be-4115-b960-58d147b56e9d\") " pod="openstack/placement-5fc7d5fdc6-llxdt" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.756001 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b8ba7416-80be-4115-b960-58d147b56e9d-logs\") pod \"placement-5fc7d5fdc6-llxdt\" (UID: \"b8ba7416-80be-4115-b960-58d147b56e9d\") " pod="openstack/placement-5fc7d5fdc6-llxdt" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.756095 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8ba7416-80be-4115-b960-58d147b56e9d-combined-ca-bundle\") pod \"placement-5fc7d5fdc6-llxdt\" (UID: \"b8ba7416-80be-4115-b960-58d147b56e9d\") " pod="openstack/placement-5fc7d5fdc6-llxdt" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.756125 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79161677-35e4-41b3-8fb4-f2bd4beebe60-logs\") pod \"barbican-worker-646ffc98d7-5l7d6\" (UID: \"79161677-35e4-41b3-8fb4-f2bd4beebe60\") " pod="openstack/barbican-worker-646ffc98d7-5l7d6" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.756167 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8ba7416-80be-4115-b960-58d147b56e9d-config-data\") pod \"placement-5fc7d5fdc6-llxdt\" (UID: \"b8ba7416-80be-4115-b960-58d147b56e9d\") " pod="openstack/placement-5fc7d5fdc6-llxdt" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.756198 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d925ed8-47dc-4bf2-a476-2526d257d5d0-logs\") pod \"barbican-keystone-listener-5cdb89b766-6mc5c\" (UID: \"1d925ed8-47dc-4bf2-a476-2526d257d5d0\") " pod="openstack/barbican-keystone-listener-5cdb89b766-6mc5c" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.756249 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79161677-35e4-41b3-8fb4-f2bd4beebe60-config-data\") pod \"barbican-worker-646ffc98d7-5l7d6\" (UID: \"79161677-35e4-41b3-8fb4-f2bd4beebe60\") " pod="openstack/barbican-worker-646ffc98d7-5l7d6" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.756306 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b8ba7416-80be-4115-b960-58d147b56e9d-public-tls-certs\") pod \"placement-5fc7d5fdc6-llxdt\" (UID: \"b8ba7416-80be-4115-b960-58d147b56e9d\") " pod="openstack/placement-5fc7d5fdc6-llxdt" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.756418 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d925ed8-47dc-4bf2-a476-2526d257d5d0-config-data\") pod \"barbican-keystone-listener-5cdb89b766-6mc5c\" (UID: \"1d925ed8-47dc-4bf2-a476-2526d257d5d0\") " pod="openstack/barbican-keystone-listener-5cdb89b766-6mc5c" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.756493 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58z6l\" (UniqueName: \"kubernetes.io/projected/79161677-35e4-41b3-8fb4-f2bd4beebe60-kube-api-access-58z6l\") pod \"barbican-worker-646ffc98d7-5l7d6\" (UID: \"79161677-35e4-41b3-8fb4-f2bd4beebe60\") " pod="openstack/barbican-worker-646ffc98d7-5l7d6" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.756577 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1d925ed8-47dc-4bf2-a476-2526d257d5d0-config-data-custom\") pod \"barbican-keystone-listener-5cdb89b766-6mc5c\" (UID: \"1d925ed8-47dc-4bf2-a476-2526d257d5d0\") " pod="openstack/barbican-keystone-listener-5cdb89b766-6mc5c" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.756627 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/79161677-35e4-41b3-8fb4-f2bd4beebe60-config-data-custom\") pod \"barbican-worker-646ffc98d7-5l7d6\" (UID: \"79161677-35e4-41b3-8fb4-f2bd4beebe60\") " pod="openstack/barbican-worker-646ffc98d7-5l7d6" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.756656 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nds75\" (UniqueName: \"kubernetes.io/projected/b8ba7416-80be-4115-b960-58d147b56e9d-kube-api-access-nds75\") pod \"placement-5fc7d5fdc6-llxdt\" (UID: \"b8ba7416-80be-4115-b960-58d147b56e9d\") " pod="openstack/placement-5fc7d5fdc6-llxdt" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.758456 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b8ba7416-80be-4115-b960-58d147b56e9d-logs\") pod \"placement-5fc7d5fdc6-llxdt\" (UID: \"b8ba7416-80be-4115-b960-58d147b56e9d\") " pod="openstack/placement-5fc7d5fdc6-llxdt" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.761767 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79161677-35e4-41b3-8fb4-f2bd4beebe60-logs\") pod \"barbican-worker-646ffc98d7-5l7d6\" (UID: \"79161677-35e4-41b3-8fb4-f2bd4beebe60\") " pod="openstack/barbican-worker-646ffc98d7-5l7d6" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.767399 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8ba7416-80be-4115-b960-58d147b56e9d-config-data\") pod \"placement-5fc7d5fdc6-llxdt\" (UID: \"b8ba7416-80be-4115-b960-58d147b56e9d\") " pod="openstack/placement-5fc7d5fdc6-llxdt" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.770552 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8ba7416-80be-4115-b960-58d147b56e9d-scripts\") pod \"placement-5fc7d5fdc6-llxdt\" (UID: \"b8ba7416-80be-4115-b960-58d147b56e9d\") " pod="openstack/placement-5fc7d5fdc6-llxdt" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.770601 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8ba7416-80be-4115-b960-58d147b56e9d-combined-ca-bundle\") pod \"placement-5fc7d5fdc6-llxdt\" (UID: \"b8ba7416-80be-4115-b960-58d147b56e9d\") " pod="openstack/placement-5fc7d5fdc6-llxdt" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.771184 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b8ba7416-80be-4115-b960-58d147b56e9d-internal-tls-certs\") pod \"placement-5fc7d5fdc6-llxdt\" (UID: \"b8ba7416-80be-4115-b960-58d147b56e9d\") " pod="openstack/placement-5fc7d5fdc6-llxdt" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.771269 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79161677-35e4-41b3-8fb4-f2bd4beebe60-combined-ca-bundle\") pod \"barbican-worker-646ffc98d7-5l7d6\" (UID: \"79161677-35e4-41b3-8fb4-f2bd4beebe60\") " pod="openstack/barbican-worker-646ffc98d7-5l7d6" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.773021 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nds75\" (UniqueName: \"kubernetes.io/projected/b8ba7416-80be-4115-b960-58d147b56e9d-kube-api-access-nds75\") pod \"placement-5fc7d5fdc6-llxdt\" (UID: \"b8ba7416-80be-4115-b960-58d147b56e9d\") " pod="openstack/placement-5fc7d5fdc6-llxdt" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.773504 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b8ba7416-80be-4115-b960-58d147b56e9d-public-tls-certs\") pod \"placement-5fc7d5fdc6-llxdt\" (UID: \"b8ba7416-80be-4115-b960-58d147b56e9d\") " pod="openstack/placement-5fc7d5fdc6-llxdt" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.774232 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79161677-35e4-41b3-8fb4-f2bd4beebe60-config-data\") pod \"barbican-worker-646ffc98d7-5l7d6\" (UID: \"79161677-35e4-41b3-8fb4-f2bd4beebe60\") " pod="openstack/barbican-worker-646ffc98d7-5l7d6" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.778864 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58z6l\" (UniqueName: \"kubernetes.io/projected/79161677-35e4-41b3-8fb4-f2bd4beebe60-kube-api-access-58z6l\") pod \"barbican-worker-646ffc98d7-5l7d6\" (UID: \"79161677-35e4-41b3-8fb4-f2bd4beebe60\") " pod="openstack/barbican-worker-646ffc98d7-5l7d6" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.788557 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-k4rfq"] Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.789612 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/79161677-35e4-41b3-8fb4-f2bd4beebe60-config-data-custom\") pod \"barbican-worker-646ffc98d7-5l7d6\" (UID: \"79161677-35e4-41b3-8fb4-f2bd4beebe60\") " pod="openstack/barbican-worker-646ffc98d7-5l7d6" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.823734 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5675ff94b4-kfxhm"] Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.827760 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-xs2c8" event={"ID":"c75cf833-5179-4e18-8f74-3faed72757ed","Type":"ContainerStarted","Data":"13167b3c813f9d5bc3d34d0d09928c7bb068ae3337cb629047c05d9accdf14ab"} Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.827988 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5675ff94b4-kfxhm" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.832819 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.834305 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c40b4907-61d2-4ec7-b611-cb01d399c8d0","Type":"ContainerStarted","Data":"e66f47a441764738daccb70723b5c733763a3cad98cdd4e042a369cc6b7bce32"} Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.836290 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5675ff94b4-kfxhm"] Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.858552 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/90f11553-7c39-4768-8669-85112803eb21-dns-svc\") pod \"dnsmasq-dns-85ff748b95-k4rfq\" (UID: \"90f11553-7c39-4768-8669-85112803eb21\") " pod="openstack/dnsmasq-dns-85ff748b95-k4rfq" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.858631 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d925ed8-47dc-4bf2-a476-2526d257d5d0-logs\") pod \"barbican-keystone-listener-5cdb89b766-6mc5c\" (UID: \"1d925ed8-47dc-4bf2-a476-2526d257d5d0\") " pod="openstack/barbican-keystone-listener-5cdb89b766-6mc5c" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.858706 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90f11553-7c39-4768-8669-85112803eb21-config\") pod \"dnsmasq-dns-85ff748b95-k4rfq\" (UID: \"90f11553-7c39-4768-8669-85112803eb21\") " pod="openstack/dnsmasq-dns-85ff748b95-k4rfq" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.858734 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgdx2\" (UniqueName: \"kubernetes.io/projected/90f11553-7c39-4768-8669-85112803eb21-kube-api-access-tgdx2\") pod \"dnsmasq-dns-85ff748b95-k4rfq\" (UID: \"90f11553-7c39-4768-8669-85112803eb21\") " pod="openstack/dnsmasq-dns-85ff748b95-k4rfq" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.858796 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d925ed8-47dc-4bf2-a476-2526d257d5d0-config-data\") pod \"barbican-keystone-listener-5cdb89b766-6mc5c\" (UID: \"1d925ed8-47dc-4bf2-a476-2526d257d5d0\") " pod="openstack/barbican-keystone-listener-5cdb89b766-6mc5c" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.858889 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/90f11553-7c39-4768-8669-85112803eb21-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-k4rfq\" (UID: \"90f11553-7c39-4768-8669-85112803eb21\") " pod="openstack/dnsmasq-dns-85ff748b95-k4rfq" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.858918 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1d925ed8-47dc-4bf2-a476-2526d257d5d0-config-data-custom\") pod \"barbican-keystone-listener-5cdb89b766-6mc5c\" (UID: \"1d925ed8-47dc-4bf2-a476-2526d257d5d0\") " pod="openstack/barbican-keystone-listener-5cdb89b766-6mc5c" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.859200 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d925ed8-47dc-4bf2-a476-2526d257d5d0-logs\") pod \"barbican-keystone-listener-5cdb89b766-6mc5c\" (UID: \"1d925ed8-47dc-4bf2-a476-2526d257d5d0\") " pod="openstack/barbican-keystone-listener-5cdb89b766-6mc5c" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.859687 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnqcz\" (UniqueName: \"kubernetes.io/projected/1d925ed8-47dc-4bf2-a476-2526d257d5d0-kube-api-access-rnqcz\") pod \"barbican-keystone-listener-5cdb89b766-6mc5c\" (UID: \"1d925ed8-47dc-4bf2-a476-2526d257d5d0\") " pod="openstack/barbican-keystone-listener-5cdb89b766-6mc5c" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.859751 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d925ed8-47dc-4bf2-a476-2526d257d5d0-combined-ca-bundle\") pod \"barbican-keystone-listener-5cdb89b766-6mc5c\" (UID: \"1d925ed8-47dc-4bf2-a476-2526d257d5d0\") " pod="openstack/barbican-keystone-listener-5cdb89b766-6mc5c" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.859779 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/90f11553-7c39-4768-8669-85112803eb21-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-k4rfq\" (UID: \"90f11553-7c39-4768-8669-85112803eb21\") " pod="openstack/dnsmasq-dns-85ff748b95-k4rfq" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.859981 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/90f11553-7c39-4768-8669-85112803eb21-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-k4rfq\" (UID: \"90f11553-7c39-4768-8669-85112803eb21\") " pod="openstack/dnsmasq-dns-85ff748b95-k4rfq" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.862434 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5fc7d5fdc6-llxdt" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.864971 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d925ed8-47dc-4bf2-a476-2526d257d5d0-config-data\") pod \"barbican-keystone-listener-5cdb89b766-6mc5c\" (UID: \"1d925ed8-47dc-4bf2-a476-2526d257d5d0\") " pod="openstack/barbican-keystone-listener-5cdb89b766-6mc5c" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.868526 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1d925ed8-47dc-4bf2-a476-2526d257d5d0-config-data-custom\") pod \"barbican-keystone-listener-5cdb89b766-6mc5c\" (UID: \"1d925ed8-47dc-4bf2-a476-2526d257d5d0\") " pod="openstack/barbican-keystone-listener-5cdb89b766-6mc5c" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.868878 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d925ed8-47dc-4bf2-a476-2526d257d5d0-combined-ca-bundle\") pod \"barbican-keystone-listener-5cdb89b766-6mc5c\" (UID: \"1d925ed8-47dc-4bf2-a476-2526d257d5d0\") " pod="openstack/barbican-keystone-listener-5cdb89b766-6mc5c" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.879259 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnqcz\" (UniqueName: \"kubernetes.io/projected/1d925ed8-47dc-4bf2-a476-2526d257d5d0-kube-api-access-rnqcz\") pod \"barbican-keystone-listener-5cdb89b766-6mc5c\" (UID: \"1d925ed8-47dc-4bf2-a476-2526d257d5d0\") " pod="openstack/barbican-keystone-listener-5cdb89b766-6mc5c" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.895227 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-646ffc98d7-5l7d6" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.962348 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/90f11553-7c39-4768-8669-85112803eb21-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-k4rfq\" (UID: \"90f11553-7c39-4768-8669-85112803eb21\") " pod="openstack/dnsmasq-dns-85ff748b95-k4rfq" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.962446 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvms5\" (UniqueName: \"kubernetes.io/projected/db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9-kube-api-access-pvms5\") pod \"barbican-api-5675ff94b4-kfxhm\" (UID: \"db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9\") " pod="openstack/barbican-api-5675ff94b4-kfxhm" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.962481 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9-config-data\") pod \"barbican-api-5675ff94b4-kfxhm\" (UID: \"db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9\") " pod="openstack/barbican-api-5675ff94b4-kfxhm" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.962509 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/90f11553-7c39-4768-8669-85112803eb21-dns-svc\") pod \"dnsmasq-dns-85ff748b95-k4rfq\" (UID: \"90f11553-7c39-4768-8669-85112803eb21\") " pod="openstack/dnsmasq-dns-85ff748b95-k4rfq" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.962548 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9-config-data-custom\") pod \"barbican-api-5675ff94b4-kfxhm\" (UID: \"db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9\") " pod="openstack/barbican-api-5675ff94b4-kfxhm" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.962580 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9-combined-ca-bundle\") pod \"barbican-api-5675ff94b4-kfxhm\" (UID: \"db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9\") " pod="openstack/barbican-api-5675ff94b4-kfxhm" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.962627 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90f11553-7c39-4768-8669-85112803eb21-config\") pod \"dnsmasq-dns-85ff748b95-k4rfq\" (UID: \"90f11553-7c39-4768-8669-85112803eb21\") " pod="openstack/dnsmasq-dns-85ff748b95-k4rfq" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.962652 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgdx2\" (UniqueName: \"kubernetes.io/projected/90f11553-7c39-4768-8669-85112803eb21-kube-api-access-tgdx2\") pod \"dnsmasq-dns-85ff748b95-k4rfq\" (UID: \"90f11553-7c39-4768-8669-85112803eb21\") " pod="openstack/dnsmasq-dns-85ff748b95-k4rfq" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.962907 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/90f11553-7c39-4768-8669-85112803eb21-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-k4rfq\" (UID: \"90f11553-7c39-4768-8669-85112803eb21\") " pod="openstack/dnsmasq-dns-85ff748b95-k4rfq" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.963046 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/90f11553-7c39-4768-8669-85112803eb21-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-k4rfq\" (UID: \"90f11553-7c39-4768-8669-85112803eb21\") " pod="openstack/dnsmasq-dns-85ff748b95-k4rfq" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.963158 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9-logs\") pod \"barbican-api-5675ff94b4-kfxhm\" (UID: \"db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9\") " pod="openstack/barbican-api-5675ff94b4-kfxhm" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.966680 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/90f11553-7c39-4768-8669-85112803eb21-dns-svc\") pod \"dnsmasq-dns-85ff748b95-k4rfq\" (UID: \"90f11553-7c39-4768-8669-85112803eb21\") " pod="openstack/dnsmasq-dns-85ff748b95-k4rfq" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.966752 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/90f11553-7c39-4768-8669-85112803eb21-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-k4rfq\" (UID: \"90f11553-7c39-4768-8669-85112803eb21\") " pod="openstack/dnsmasq-dns-85ff748b95-k4rfq" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.967705 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90f11553-7c39-4768-8669-85112803eb21-config\") pod \"dnsmasq-dns-85ff748b95-k4rfq\" (UID: \"90f11553-7c39-4768-8669-85112803eb21\") " pod="openstack/dnsmasq-dns-85ff748b95-k4rfq" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.967826 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/90f11553-7c39-4768-8669-85112803eb21-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-k4rfq\" (UID: \"90f11553-7c39-4768-8669-85112803eb21\") " pod="openstack/dnsmasq-dns-85ff748b95-k4rfq" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.977214 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/90f11553-7c39-4768-8669-85112803eb21-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-k4rfq\" (UID: \"90f11553-7c39-4768-8669-85112803eb21\") " pod="openstack/dnsmasq-dns-85ff748b95-k4rfq" Dec 06 06:17:15 crc kubenswrapper[4809]: I1206 06:17:15.990644 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgdx2\" (UniqueName: \"kubernetes.io/projected/90f11553-7c39-4768-8669-85112803eb21-kube-api-access-tgdx2\") pod \"dnsmasq-dns-85ff748b95-k4rfq\" (UID: \"90f11553-7c39-4768-8669-85112803eb21\") " pod="openstack/dnsmasq-dns-85ff748b95-k4rfq" Dec 06 06:17:16 crc kubenswrapper[4809]: I1206 06:17:16.004415 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5cdb89b766-6mc5c" Dec 06 06:17:16 crc kubenswrapper[4809]: I1206 06:17:16.057086 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-k4rfq" Dec 06 06:17:16 crc kubenswrapper[4809]: I1206 06:17:16.064693 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9-logs\") pod \"barbican-api-5675ff94b4-kfxhm\" (UID: \"db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9\") " pod="openstack/barbican-api-5675ff94b4-kfxhm" Dec 06 06:17:16 crc kubenswrapper[4809]: I1206 06:17:16.064774 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvms5\" (UniqueName: \"kubernetes.io/projected/db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9-kube-api-access-pvms5\") pod \"barbican-api-5675ff94b4-kfxhm\" (UID: \"db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9\") " pod="openstack/barbican-api-5675ff94b4-kfxhm" Dec 06 06:17:16 crc kubenswrapper[4809]: I1206 06:17:16.064801 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9-config-data\") pod \"barbican-api-5675ff94b4-kfxhm\" (UID: \"db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9\") " pod="openstack/barbican-api-5675ff94b4-kfxhm" Dec 06 06:17:16 crc kubenswrapper[4809]: I1206 06:17:16.064828 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9-config-data-custom\") pod \"barbican-api-5675ff94b4-kfxhm\" (UID: \"db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9\") " pod="openstack/barbican-api-5675ff94b4-kfxhm" Dec 06 06:17:16 crc kubenswrapper[4809]: I1206 06:17:16.064851 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9-combined-ca-bundle\") pod \"barbican-api-5675ff94b4-kfxhm\" (UID: \"db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9\") " pod="openstack/barbican-api-5675ff94b4-kfxhm" Dec 06 06:17:16 crc kubenswrapper[4809]: I1206 06:17:16.071082 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9-logs\") pod \"barbican-api-5675ff94b4-kfxhm\" (UID: \"db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9\") " pod="openstack/barbican-api-5675ff94b4-kfxhm" Dec 06 06:17:16 crc kubenswrapper[4809]: I1206 06:17:16.071923 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9-combined-ca-bundle\") pod \"barbican-api-5675ff94b4-kfxhm\" (UID: \"db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9\") " pod="openstack/barbican-api-5675ff94b4-kfxhm" Dec 06 06:17:16 crc kubenswrapper[4809]: I1206 06:17:16.079067 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9-config-data\") pod \"barbican-api-5675ff94b4-kfxhm\" (UID: \"db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9\") " pod="openstack/barbican-api-5675ff94b4-kfxhm" Dec 06 06:17:16 crc kubenswrapper[4809]: I1206 06:17:16.080336 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9-config-data-custom\") pod \"barbican-api-5675ff94b4-kfxhm\" (UID: \"db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9\") " pod="openstack/barbican-api-5675ff94b4-kfxhm" Dec 06 06:17:16 crc kubenswrapper[4809]: I1206 06:17:16.094801 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvms5\" (UniqueName: \"kubernetes.io/projected/db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9-kube-api-access-pvms5\") pod \"barbican-api-5675ff94b4-kfxhm\" (UID: \"db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9\") " pod="openstack/barbican-api-5675ff94b4-kfxhm" Dec 06 06:17:16 crc kubenswrapper[4809]: I1206 06:17:16.260059 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5675ff94b4-kfxhm" Dec 06 06:17:16 crc kubenswrapper[4809]: I1206 06:17:16.441466 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-ckhg9" Dec 06 06:17:16 crc kubenswrapper[4809]: I1206 06:17:16.586690 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edf1ab66-b483-4172-81fd-dd3f3b9b44aa-config-data\") pod \"edf1ab66-b483-4172-81fd-dd3f3b9b44aa\" (UID: \"edf1ab66-b483-4172-81fd-dd3f3b9b44aa\") " Dec 06 06:17:16 crc kubenswrapper[4809]: I1206 06:17:16.587237 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkz67\" (UniqueName: \"kubernetes.io/projected/edf1ab66-b483-4172-81fd-dd3f3b9b44aa-kube-api-access-zkz67\") pod \"edf1ab66-b483-4172-81fd-dd3f3b9b44aa\" (UID: \"edf1ab66-b483-4172-81fd-dd3f3b9b44aa\") " Dec 06 06:17:16 crc kubenswrapper[4809]: I1206 06:17:16.587450 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edf1ab66-b483-4172-81fd-dd3f3b9b44aa-combined-ca-bundle\") pod \"edf1ab66-b483-4172-81fd-dd3f3b9b44aa\" (UID: \"edf1ab66-b483-4172-81fd-dd3f3b9b44aa\") " Dec 06 06:17:16 crc kubenswrapper[4809]: I1206 06:17:16.598224 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edf1ab66-b483-4172-81fd-dd3f3b9b44aa-kube-api-access-zkz67" (OuterVolumeSpecName: "kube-api-access-zkz67") pod "edf1ab66-b483-4172-81fd-dd3f3b9b44aa" (UID: "edf1ab66-b483-4172-81fd-dd3f3b9b44aa"). InnerVolumeSpecName "kube-api-access-zkz67". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:17:16 crc kubenswrapper[4809]: I1206 06:17:16.647416 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edf1ab66-b483-4172-81fd-dd3f3b9b44aa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "edf1ab66-b483-4172-81fd-dd3f3b9b44aa" (UID: "edf1ab66-b483-4172-81fd-dd3f3b9b44aa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:17:16 crc kubenswrapper[4809]: I1206 06:17:16.689735 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edf1ab66-b483-4172-81fd-dd3f3b9b44aa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:16 crc kubenswrapper[4809]: I1206 06:17:16.689765 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkz67\" (UniqueName: \"kubernetes.io/projected/edf1ab66-b483-4172-81fd-dd3f3b9b44aa-kube-api-access-zkz67\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:16 crc kubenswrapper[4809]: I1206 06:17:16.733067 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edf1ab66-b483-4172-81fd-dd3f3b9b44aa-config-data" (OuterVolumeSpecName: "config-data") pod "edf1ab66-b483-4172-81fd-dd3f3b9b44aa" (UID: "edf1ab66-b483-4172-81fd-dd3f3b9b44aa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:17:16 crc kubenswrapper[4809]: I1206 06:17:16.780305 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-646ffc98d7-5l7d6"] Dec 06 06:17:16 crc kubenswrapper[4809]: I1206 06:17:16.792004 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edf1ab66-b483-4172-81fd-dd3f3b9b44aa-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:16 crc kubenswrapper[4809]: I1206 06:17:16.804869 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5fc7d5fdc6-llxdt"] Dec 06 06:17:16 crc kubenswrapper[4809]: W1206 06:17:16.868100 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod79161677_35e4_41b3_8fb4_f2bd4beebe60.slice/crio-7ead9bb28ca24da5a8ad27a4673205934e63c087feda6069b06e97490d6a7aa9 WatchSource:0}: Error finding container 7ead9bb28ca24da5a8ad27a4673205934e63c087feda6069b06e97490d6a7aa9: Status 404 returned error can't find the container with id 7ead9bb28ca24da5a8ad27a4673205934e63c087feda6069b06e97490d6a7aa9 Dec 06 06:17:16 crc kubenswrapper[4809]: I1206 06:17:16.921388 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-ckhg9" event={"ID":"edf1ab66-b483-4172-81fd-dd3f3b9b44aa","Type":"ContainerDied","Data":"a9627522d80b87d58fb1947fb4b4e91e8a19b1bd560edd020ccf2a4d8fd042e4"} Dec 06 06:17:16 crc kubenswrapper[4809]: I1206 06:17:16.921820 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a9627522d80b87d58fb1947fb4b4e91e8a19b1bd560edd020ccf2a4d8fd042e4" Dec 06 06:17:16 crc kubenswrapper[4809]: I1206 06:17:16.921693 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-ckhg9" Dec 06 06:17:16 crc kubenswrapper[4809]: I1206 06:17:16.926899 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c40b4907-61d2-4ec7-b611-cb01d399c8d0","Type":"ContainerStarted","Data":"328becf5ebf16384ad2244583014f70780407728394585b140c94f7a4e16e55b"} Dec 06 06:17:16 crc kubenswrapper[4809]: I1206 06:17:16.969009 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-xs2c8" podStartSLOduration=5.332016462 podStartE2EDuration="56.968976344s" podCreationTimestamp="2025-12-06 06:16:20 +0000 UTC" firstStartedPulling="2025-12-06 06:16:22.4318052 +0000 UTC m=+1507.320788142" lastFinishedPulling="2025-12-06 06:17:14.068765082 +0000 UTC m=+1558.957748024" observedRunningTime="2025-12-06 06:17:16.95027976 +0000 UTC m=+1561.839262702" watchObservedRunningTime="2025-12-06 06:17:16.968976344 +0000 UTC m=+1561.857959286" Dec 06 06:17:17 crc kubenswrapper[4809]: I1206 06:17:17.019213 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=9.019185739 podStartE2EDuration="9.019185739s" podCreationTimestamp="2025-12-06 06:17:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:17:16.978677846 +0000 UTC m=+1561.867660798" watchObservedRunningTime="2025-12-06 06:17:17.019185739 +0000 UTC m=+1561.908168681" Dec 06 06:17:17 crc kubenswrapper[4809]: I1206 06:17:17.101911 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-k4rfq"] Dec 06 06:17:17 crc kubenswrapper[4809]: I1206 06:17:17.182085 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5675ff94b4-kfxhm"] Dec 06 06:17:17 crc kubenswrapper[4809]: W1206 06:17:17.267954 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddb2e7d63_181d_43f7_8cdd_7bfb8f95b6d9.slice/crio-531badae15a0972281ef0ca64fd43b28aab8c0c00afac77c3ff2814d1ec069e6 WatchSource:0}: Error finding container 531badae15a0972281ef0ca64fd43b28aab8c0c00afac77c3ff2814d1ec069e6: Status 404 returned error can't find the container with id 531badae15a0972281ef0ca64fd43b28aab8c0c00afac77c3ff2814d1ec069e6 Dec 06 06:17:17 crc kubenswrapper[4809]: I1206 06:17:17.373385 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5cdb89b766-6mc5c"] Dec 06 06:17:17 crc kubenswrapper[4809]: W1206 06:17:17.385827 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1d925ed8_47dc_4bf2_a476_2526d257d5d0.slice/crio-d672bf418b1c8da932d567f69812825cf4e27972bf083b7c9a5cda0d43a74038 WatchSource:0}: Error finding container d672bf418b1c8da932d567f69812825cf4e27972bf083b7c9a5cda0d43a74038: Status 404 returned error can't find the container with id d672bf418b1c8da932d567f69812825cf4e27972bf083b7c9a5cda0d43a74038 Dec 06 06:17:18 crc kubenswrapper[4809]: I1206 06:17:18.033728 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5fc7d5fdc6-llxdt" event={"ID":"b8ba7416-80be-4115-b960-58d147b56e9d","Type":"ContainerStarted","Data":"881d3d01d473789c30d2b9f57416b93184a5a520dff5e61095a090c7aa6b1587"} Dec 06 06:17:18 crc kubenswrapper[4809]: I1206 06:17:18.048508 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5fc7d5fdc6-llxdt" event={"ID":"b8ba7416-80be-4115-b960-58d147b56e9d","Type":"ContainerStarted","Data":"32bc078f281c459a5cb0b75be3e579462a69a1e599b87c5f6d62ccfd179db94f"} Dec 06 06:17:18 crc kubenswrapper[4809]: I1206 06:17:18.052166 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-646ffc98d7-5l7d6" event={"ID":"79161677-35e4-41b3-8fb4-f2bd4beebe60","Type":"ContainerStarted","Data":"7ead9bb28ca24da5a8ad27a4673205934e63c087feda6069b06e97490d6a7aa9"} Dec 06 06:17:18 crc kubenswrapper[4809]: I1206 06:17:18.089163 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-k4rfq" event={"ID":"90f11553-7c39-4768-8669-85112803eb21","Type":"ContainerStarted","Data":"7f39a7c141b936ffb87141024c5ebc8375a90519fdb2000f95a93b6f885d25a0"} Dec 06 06:17:18 crc kubenswrapper[4809]: I1206 06:17:18.089202 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-k4rfq" event={"ID":"90f11553-7c39-4768-8669-85112803eb21","Type":"ContainerStarted","Data":"3cc3ba849aa2b6e3870642cabdb432ad61b3f823251d281d314e51d0e445ad86"} Dec 06 06:17:18 crc kubenswrapper[4809]: I1206 06:17:18.119565 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5675ff94b4-kfxhm" event={"ID":"db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9","Type":"ContainerStarted","Data":"594bd3483b5ff6789ef10412230d25201c6a9a66c05826c4f8915a1a7976f0d3"} Dec 06 06:17:18 crc kubenswrapper[4809]: I1206 06:17:18.119607 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5675ff94b4-kfxhm" event={"ID":"db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9","Type":"ContainerStarted","Data":"531badae15a0972281ef0ca64fd43b28aab8c0c00afac77c3ff2814d1ec069e6"} Dec 06 06:17:18 crc kubenswrapper[4809]: I1206 06:17:18.122575 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5cdb89b766-6mc5c" event={"ID":"1d925ed8-47dc-4bf2-a476-2526d257d5d0","Type":"ContainerStarted","Data":"d672bf418b1c8da932d567f69812825cf4e27972bf083b7c9a5cda0d43a74038"} Dec 06 06:17:18 crc kubenswrapper[4809]: I1206 06:17:18.666257 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 06 06:17:18 crc kubenswrapper[4809]: I1206 06:17:18.666600 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 06 06:17:18 crc kubenswrapper[4809]: I1206 06:17:18.726700 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 06 06:17:18 crc kubenswrapper[4809]: I1206 06:17:18.749250 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 06 06:17:18 crc kubenswrapper[4809]: I1206 06:17:18.950300 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 06 06:17:18 crc kubenswrapper[4809]: I1206 06:17:18.950358 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 06 06:17:19 crc kubenswrapper[4809]: I1206 06:17:19.003661 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5bb7f9f986-k7477"] Dec 06 06:17:19 crc kubenswrapper[4809]: E1206 06:17:19.004159 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edf1ab66-b483-4172-81fd-dd3f3b9b44aa" containerName="heat-db-sync" Dec 06 06:17:19 crc kubenswrapper[4809]: I1206 06:17:19.004181 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="edf1ab66-b483-4172-81fd-dd3f3b9b44aa" containerName="heat-db-sync" Dec 06 06:17:19 crc kubenswrapper[4809]: I1206 06:17:19.004399 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="edf1ab66-b483-4172-81fd-dd3f3b9b44aa" containerName="heat-db-sync" Dec 06 06:17:19 crc kubenswrapper[4809]: I1206 06:17:19.006171 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5bb7f9f986-k7477" Dec 06 06:17:19 crc kubenswrapper[4809]: I1206 06:17:19.007368 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 06 06:17:19 crc kubenswrapper[4809]: I1206 06:17:19.009390 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 06 06:17:19 crc kubenswrapper[4809]: I1206 06:17:19.009480 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 06 06:17:19 crc kubenswrapper[4809]: I1206 06:17:19.033706 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 06 06:17:19 crc kubenswrapper[4809]: I1206 06:17:19.052994 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5bb7f9f986-k7477"] Dec 06 06:17:19 crc kubenswrapper[4809]: I1206 06:17:19.143847 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5fc7d5fdc6-llxdt" event={"ID":"b8ba7416-80be-4115-b960-58d147b56e9d","Type":"ContainerStarted","Data":"2558b1a8f84340c6f265952698a2256e3db8169b1bfa5ae93266b5af5b7770e8"} Dec 06 06:17:19 crc kubenswrapper[4809]: I1206 06:17:19.145074 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5fc7d5fdc6-llxdt" Dec 06 06:17:19 crc kubenswrapper[4809]: I1206 06:17:19.145150 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5fc7d5fdc6-llxdt" Dec 06 06:17:19 crc kubenswrapper[4809]: I1206 06:17:19.150887 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5675ff94b4-kfxhm" event={"ID":"db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9","Type":"ContainerStarted","Data":"b9295161ebe317bbd2d4cba00e6910036c3f9d287808b173f9dbffad7a300f73"} Dec 06 06:17:19 crc kubenswrapper[4809]: I1206 06:17:19.153516 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5675ff94b4-kfxhm" Dec 06 06:17:19 crc kubenswrapper[4809]: I1206 06:17:19.153568 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5675ff94b4-kfxhm" Dec 06 06:17:19 crc kubenswrapper[4809]: I1206 06:17:19.165527 4809 generic.go:334] "Generic (PLEG): container finished" podID="90f11553-7c39-4768-8669-85112803eb21" containerID="7f39a7c141b936ffb87141024c5ebc8375a90519fdb2000f95a93b6f885d25a0" exitCode=0 Dec 06 06:17:19 crc kubenswrapper[4809]: I1206 06:17:19.166965 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-k4rfq" event={"ID":"90f11553-7c39-4768-8669-85112803eb21","Type":"ContainerDied","Data":"7f39a7c141b936ffb87141024c5ebc8375a90519fdb2000f95a93b6f885d25a0"} Dec 06 06:17:19 crc kubenswrapper[4809]: I1206 06:17:19.166999 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-k4rfq" event={"ID":"90f11553-7c39-4768-8669-85112803eb21","Type":"ContainerStarted","Data":"6d46c7bf2cc29fb06ab2974a2f57728ed4bc42abd66266594d85cf362760e738"} Dec 06 06:17:19 crc kubenswrapper[4809]: I1206 06:17:19.167581 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-85ff748b95-k4rfq" Dec 06 06:17:19 crc kubenswrapper[4809]: I1206 06:17:19.168556 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 06 06:17:19 crc kubenswrapper[4809]: I1206 06:17:19.168670 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 06 06:17:19 crc kubenswrapper[4809]: I1206 06:17:19.168698 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 06 06:17:19 crc kubenswrapper[4809]: I1206 06:17:19.168840 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 06 06:17:19 crc kubenswrapper[4809]: I1206 06:17:19.185457 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-5fc7d5fdc6-llxdt" podStartSLOduration=4.185435348 podStartE2EDuration="4.185435348s" podCreationTimestamp="2025-12-06 06:17:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:17:19.17991644 +0000 UTC m=+1564.068899402" watchObservedRunningTime="2025-12-06 06:17:19.185435348 +0000 UTC m=+1564.074418290" Dec 06 06:17:19 crc kubenswrapper[4809]: I1206 06:17:19.202506 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7df2cd07-a4c8-4934-8755-7fa7a4c4b3fa-logs\") pod \"barbican-api-5bb7f9f986-k7477\" (UID: \"7df2cd07-a4c8-4934-8755-7fa7a4c4b3fa\") " pod="openstack/barbican-api-5bb7f9f986-k7477" Dec 06 06:17:19 crc kubenswrapper[4809]: I1206 06:17:19.202580 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7df2cd07-a4c8-4934-8755-7fa7a4c4b3fa-config-data-custom\") pod \"barbican-api-5bb7f9f986-k7477\" (UID: \"7df2cd07-a4c8-4934-8755-7fa7a4c4b3fa\") " pod="openstack/barbican-api-5bb7f9f986-k7477" Dec 06 06:17:19 crc kubenswrapper[4809]: I1206 06:17:19.202616 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7df2cd07-a4c8-4934-8755-7fa7a4c4b3fa-config-data\") pod \"barbican-api-5bb7f9f986-k7477\" (UID: \"7df2cd07-a4c8-4934-8755-7fa7a4c4b3fa\") " pod="openstack/barbican-api-5bb7f9f986-k7477" Dec 06 06:17:19 crc kubenswrapper[4809]: I1206 06:17:19.202634 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7df2cd07-a4c8-4934-8755-7fa7a4c4b3fa-internal-tls-certs\") pod \"barbican-api-5bb7f9f986-k7477\" (UID: \"7df2cd07-a4c8-4934-8755-7fa7a4c4b3fa\") " pod="openstack/barbican-api-5bb7f9f986-k7477" Dec 06 06:17:19 crc kubenswrapper[4809]: I1206 06:17:19.202723 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7df2cd07-a4c8-4934-8755-7fa7a4c4b3fa-public-tls-certs\") pod \"barbican-api-5bb7f9f986-k7477\" (UID: \"7df2cd07-a4c8-4934-8755-7fa7a4c4b3fa\") " pod="openstack/barbican-api-5bb7f9f986-k7477" Dec 06 06:17:19 crc kubenswrapper[4809]: I1206 06:17:19.203039 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7df2cd07-a4c8-4934-8755-7fa7a4c4b3fa-combined-ca-bundle\") pod \"barbican-api-5bb7f9f986-k7477\" (UID: \"7df2cd07-a4c8-4934-8755-7fa7a4c4b3fa\") " pod="openstack/barbican-api-5bb7f9f986-k7477" Dec 06 06:17:19 crc kubenswrapper[4809]: I1206 06:17:19.203178 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwtn2\" (UniqueName: \"kubernetes.io/projected/7df2cd07-a4c8-4934-8755-7fa7a4c4b3fa-kube-api-access-jwtn2\") pod \"barbican-api-5bb7f9f986-k7477\" (UID: \"7df2cd07-a4c8-4934-8755-7fa7a4c4b3fa\") " pod="openstack/barbican-api-5bb7f9f986-k7477" Dec 06 06:17:19 crc kubenswrapper[4809]: I1206 06:17:19.206428 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5675ff94b4-kfxhm" podStartSLOduration=4.206407814 podStartE2EDuration="4.206407814s" podCreationTimestamp="2025-12-06 06:17:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:17:19.201128592 +0000 UTC m=+1564.090111534" watchObservedRunningTime="2025-12-06 06:17:19.206407814 +0000 UTC m=+1564.095390756" Dec 06 06:17:19 crc kubenswrapper[4809]: I1206 06:17:19.305700 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7df2cd07-a4c8-4934-8755-7fa7a4c4b3fa-config-data\") pod \"barbican-api-5bb7f9f986-k7477\" (UID: \"7df2cd07-a4c8-4934-8755-7fa7a4c4b3fa\") " pod="openstack/barbican-api-5bb7f9f986-k7477" Dec 06 06:17:19 crc kubenswrapper[4809]: I1206 06:17:19.305764 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7df2cd07-a4c8-4934-8755-7fa7a4c4b3fa-internal-tls-certs\") pod \"barbican-api-5bb7f9f986-k7477\" (UID: \"7df2cd07-a4c8-4934-8755-7fa7a4c4b3fa\") " pod="openstack/barbican-api-5bb7f9f986-k7477" Dec 06 06:17:19 crc kubenswrapper[4809]: I1206 06:17:19.305802 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7df2cd07-a4c8-4934-8755-7fa7a4c4b3fa-public-tls-certs\") pod \"barbican-api-5bb7f9f986-k7477\" (UID: \"7df2cd07-a4c8-4934-8755-7fa7a4c4b3fa\") " pod="openstack/barbican-api-5bb7f9f986-k7477" Dec 06 06:17:19 crc kubenswrapper[4809]: I1206 06:17:19.306019 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7df2cd07-a4c8-4934-8755-7fa7a4c4b3fa-combined-ca-bundle\") pod \"barbican-api-5bb7f9f986-k7477\" (UID: \"7df2cd07-a4c8-4934-8755-7fa7a4c4b3fa\") " pod="openstack/barbican-api-5bb7f9f986-k7477" Dec 06 06:17:19 crc kubenswrapper[4809]: I1206 06:17:19.306101 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwtn2\" (UniqueName: \"kubernetes.io/projected/7df2cd07-a4c8-4934-8755-7fa7a4c4b3fa-kube-api-access-jwtn2\") pod \"barbican-api-5bb7f9f986-k7477\" (UID: \"7df2cd07-a4c8-4934-8755-7fa7a4c4b3fa\") " pod="openstack/barbican-api-5bb7f9f986-k7477" Dec 06 06:17:19 crc kubenswrapper[4809]: I1206 06:17:19.306367 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7df2cd07-a4c8-4934-8755-7fa7a4c4b3fa-logs\") pod \"barbican-api-5bb7f9f986-k7477\" (UID: \"7df2cd07-a4c8-4934-8755-7fa7a4c4b3fa\") " pod="openstack/barbican-api-5bb7f9f986-k7477" Dec 06 06:17:19 crc kubenswrapper[4809]: I1206 06:17:19.306500 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7df2cd07-a4c8-4934-8755-7fa7a4c4b3fa-config-data-custom\") pod \"barbican-api-5bb7f9f986-k7477\" (UID: \"7df2cd07-a4c8-4934-8755-7fa7a4c4b3fa\") " pod="openstack/barbican-api-5bb7f9f986-k7477" Dec 06 06:17:19 crc kubenswrapper[4809]: I1206 06:17:19.307891 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7df2cd07-a4c8-4934-8755-7fa7a4c4b3fa-logs\") pod \"barbican-api-5bb7f9f986-k7477\" (UID: \"7df2cd07-a4c8-4934-8755-7fa7a4c4b3fa\") " pod="openstack/barbican-api-5bb7f9f986-k7477" Dec 06 06:17:19 crc kubenswrapper[4809]: I1206 06:17:19.317469 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7df2cd07-a4c8-4934-8755-7fa7a4c4b3fa-internal-tls-certs\") pod \"barbican-api-5bb7f9f986-k7477\" (UID: \"7df2cd07-a4c8-4934-8755-7fa7a4c4b3fa\") " pod="openstack/barbican-api-5bb7f9f986-k7477" Dec 06 06:17:19 crc kubenswrapper[4809]: I1206 06:17:19.317982 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7df2cd07-a4c8-4934-8755-7fa7a4c4b3fa-public-tls-certs\") pod \"barbican-api-5bb7f9f986-k7477\" (UID: \"7df2cd07-a4c8-4934-8755-7fa7a4c4b3fa\") " pod="openstack/barbican-api-5bb7f9f986-k7477" Dec 06 06:17:19 crc kubenswrapper[4809]: I1206 06:17:19.318090 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7df2cd07-a4c8-4934-8755-7fa7a4c4b3fa-config-data-custom\") pod \"barbican-api-5bb7f9f986-k7477\" (UID: \"7df2cd07-a4c8-4934-8755-7fa7a4c4b3fa\") " pod="openstack/barbican-api-5bb7f9f986-k7477" Dec 06 06:17:19 crc kubenswrapper[4809]: I1206 06:17:19.318507 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7df2cd07-a4c8-4934-8755-7fa7a4c4b3fa-combined-ca-bundle\") pod \"barbican-api-5bb7f9f986-k7477\" (UID: \"7df2cd07-a4c8-4934-8755-7fa7a4c4b3fa\") " pod="openstack/barbican-api-5bb7f9f986-k7477" Dec 06 06:17:19 crc kubenswrapper[4809]: I1206 06:17:19.329823 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7df2cd07-a4c8-4934-8755-7fa7a4c4b3fa-config-data\") pod \"barbican-api-5bb7f9f986-k7477\" (UID: \"7df2cd07-a4c8-4934-8755-7fa7a4c4b3fa\") " pod="openstack/barbican-api-5bb7f9f986-k7477" Dec 06 06:17:19 crc kubenswrapper[4809]: I1206 06:17:19.334907 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwtn2\" (UniqueName: \"kubernetes.io/projected/7df2cd07-a4c8-4934-8755-7fa7a4c4b3fa-kube-api-access-jwtn2\") pod \"barbican-api-5bb7f9f986-k7477\" (UID: \"7df2cd07-a4c8-4934-8755-7fa7a4c4b3fa\") " pod="openstack/barbican-api-5bb7f9f986-k7477" Dec 06 06:17:19 crc kubenswrapper[4809]: I1206 06:17:19.344656 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5bb7f9f986-k7477" Dec 06 06:17:21 crc kubenswrapper[4809]: I1206 06:17:21.083792 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-85ff748b95-k4rfq" podStartSLOduration=6.083765948 podStartE2EDuration="6.083765948s" podCreationTimestamp="2025-12-06 06:17:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:17:19.231297356 +0000 UTC m=+1564.120280298" watchObservedRunningTime="2025-12-06 06:17:21.083765948 +0000 UTC m=+1565.972748890" Dec 06 06:17:21 crc kubenswrapper[4809]: I1206 06:17:21.086383 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5bb7f9f986-k7477"] Dec 06 06:17:21 crc kubenswrapper[4809]: W1206 06:17:21.113221 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7df2cd07_a4c8_4934_8755_7fa7a4c4b3fa.slice/crio-5a27e76825dcce4d0d2ad5ac4a7eceb617c3477aa3b5bde6266fe50e4a182346 WatchSource:0}: Error finding container 5a27e76825dcce4d0d2ad5ac4a7eceb617c3477aa3b5bde6266fe50e4a182346: Status 404 returned error can't find the container with id 5a27e76825dcce4d0d2ad5ac4a7eceb617c3477aa3b5bde6266fe50e4a182346 Dec 06 06:17:21 crc kubenswrapper[4809]: I1206 06:17:21.202416 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-646ffc98d7-5l7d6" event={"ID":"79161677-35e4-41b3-8fb4-f2bd4beebe60","Type":"ContainerStarted","Data":"cb8ec6fc9a20ac3aec0fdf28f97f7d19828f402d1a9ff6af3e784798d73c7575"} Dec 06 06:17:21 crc kubenswrapper[4809]: I1206 06:17:21.204280 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5bb7f9f986-k7477" event={"ID":"7df2cd07-a4c8-4934-8755-7fa7a4c4b3fa","Type":"ContainerStarted","Data":"5a27e76825dcce4d0d2ad5ac4a7eceb617c3477aa3b5bde6266fe50e4a182346"} Dec 06 06:17:21 crc kubenswrapper[4809]: I1206 06:17:21.209295 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5cdb89b766-6mc5c" event={"ID":"1d925ed8-47dc-4bf2-a476-2526d257d5d0","Type":"ContainerStarted","Data":"d52170e174cc05613bd6b5f99a76d140715d6d16ede8bdf03304c94432aadaec"} Dec 06 06:17:21 crc kubenswrapper[4809]: I1206 06:17:21.209637 4809 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 06:17:21 crc kubenswrapper[4809]: I1206 06:17:21.209800 4809 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 06:17:21 crc kubenswrapper[4809]: I1206 06:17:21.209899 4809 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 06:17:22 crc kubenswrapper[4809]: I1206 06:17:22.224622 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-646ffc98d7-5l7d6" event={"ID":"79161677-35e4-41b3-8fb4-f2bd4beebe60","Type":"ContainerStarted","Data":"ca194000a404f1e1e32d102c286e28cd3d46011fef4111f2e4f6a227cb5506ec"} Dec 06 06:17:22 crc kubenswrapper[4809]: I1206 06:17:22.235447 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5bb7f9f986-k7477" event={"ID":"7df2cd07-a4c8-4934-8755-7fa7a4c4b3fa","Type":"ContainerStarted","Data":"232f90d78ec78ae69fd2aad438e379ba4c5a880eebf67e71353b2c22381dd18d"} Dec 06 06:17:22 crc kubenswrapper[4809]: I1206 06:17:22.235509 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5bb7f9f986-k7477" event={"ID":"7df2cd07-a4c8-4934-8755-7fa7a4c4b3fa","Type":"ContainerStarted","Data":"f4f3d5f829d7d72a261152e8d5b9355209abab7f39c6d1ce98a125fb5a77338b"} Dec 06 06:17:22 crc kubenswrapper[4809]: I1206 06:17:22.236575 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5bb7f9f986-k7477" Dec 06 06:17:22 crc kubenswrapper[4809]: I1206 06:17:22.236612 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5bb7f9f986-k7477" Dec 06 06:17:22 crc kubenswrapper[4809]: I1206 06:17:22.245447 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-646ffc98d7-5l7d6" podStartSLOduration=3.724522362 podStartE2EDuration="7.245424602s" podCreationTimestamp="2025-12-06 06:17:15 +0000 UTC" firstStartedPulling="2025-12-06 06:17:16.89135417 +0000 UTC m=+1561.780337112" lastFinishedPulling="2025-12-06 06:17:20.41225641 +0000 UTC m=+1565.301239352" observedRunningTime="2025-12-06 06:17:22.24465103 +0000 UTC m=+1567.133633982" watchObservedRunningTime="2025-12-06 06:17:22.245424602 +0000 UTC m=+1567.134407544" Dec 06 06:17:22 crc kubenswrapper[4809]: I1206 06:17:22.260550 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5cdb89b766-6mc5c" event={"ID":"1d925ed8-47dc-4bf2-a476-2526d257d5d0","Type":"ContainerStarted","Data":"12f1c8b881c6544a508d9d94243eb0657c1be5202729cf057c5c61914b1adcf2"} Dec 06 06:17:22 crc kubenswrapper[4809]: I1206 06:17:22.280297 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5bb7f9f986-k7477" podStartSLOduration=4.2802772319999995 podStartE2EDuration="4.280277232s" podCreationTimestamp="2025-12-06 06:17:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:17:22.267705773 +0000 UTC m=+1567.156688715" watchObservedRunningTime="2025-12-06 06:17:22.280277232 +0000 UTC m=+1567.169260174" Dec 06 06:17:22 crc kubenswrapper[4809]: I1206 06:17:22.297322 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-5cdb89b766-6mc5c" podStartSLOduration=4.264721847 podStartE2EDuration="7.297302631s" podCreationTimestamp="2025-12-06 06:17:15 +0000 UTC" firstStartedPulling="2025-12-06 06:17:17.391611918 +0000 UTC m=+1562.280594870" lastFinishedPulling="2025-12-06 06:17:20.424192712 +0000 UTC m=+1565.313175654" observedRunningTime="2025-12-06 06:17:22.290325993 +0000 UTC m=+1567.179308935" watchObservedRunningTime="2025-12-06 06:17:22.297302631 +0000 UTC m=+1567.186285573" Dec 06 06:17:22 crc kubenswrapper[4809]: I1206 06:17:22.795481 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 06 06:17:22 crc kubenswrapper[4809]: I1206 06:17:22.795583 4809 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 06:17:22 crc kubenswrapper[4809]: I1206 06:17:22.839653 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 06 06:17:22 crc kubenswrapper[4809]: I1206 06:17:22.839844 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 06 06:17:23 crc kubenswrapper[4809]: I1206 06:17:23.354772 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 06 06:17:24 crc kubenswrapper[4809]: I1206 06:17:24.282230 4809 generic.go:334] "Generic (PLEG): container finished" podID="c75cf833-5179-4e18-8f74-3faed72757ed" containerID="13167b3c813f9d5bc3d34d0d09928c7bb068ae3337cb629047c05d9accdf14ab" exitCode=0 Dec 06 06:17:24 crc kubenswrapper[4809]: I1206 06:17:24.282441 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-xs2c8" event={"ID":"c75cf833-5179-4e18-8f74-3faed72757ed","Type":"ContainerDied","Data":"13167b3c813f9d5bc3d34d0d09928c7bb068ae3337cb629047c05d9accdf14ab"} Dec 06 06:17:26 crc kubenswrapper[4809]: I1206 06:17:26.059217 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-85ff748b95-k4rfq" Dec 06 06:17:26 crc kubenswrapper[4809]: I1206 06:17:26.147789 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-58svx"] Dec 06 06:17:26 crc kubenswrapper[4809]: I1206 06:17:26.148413 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55f844cf75-58svx" podUID="ecb6c22c-402b-4bdb-a791-29509985785a" containerName="dnsmasq-dns" containerID="cri-o://bcf33d0fa0150896477a32304bd09c31c83fec771c3aec3d7578e7f98d8ed1e8" gracePeriod=10 Dec 06 06:17:28 crc kubenswrapper[4809]: I1206 06:17:28.236618 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-55f844cf75-58svx" podUID="ecb6c22c-402b-4bdb-a791-29509985785a" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.186:5353: connect: connection refused" Dec 06 06:17:28 crc kubenswrapper[4809]: I1206 06:17:28.344105 4809 generic.go:334] "Generic (PLEG): container finished" podID="ecb6c22c-402b-4bdb-a791-29509985785a" containerID="bcf33d0fa0150896477a32304bd09c31c83fec771c3aec3d7578e7f98d8ed1e8" exitCode=0 Dec 06 06:17:28 crc kubenswrapper[4809]: I1206 06:17:28.344152 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-58svx" event={"ID":"ecb6c22c-402b-4bdb-a791-29509985785a","Type":"ContainerDied","Data":"bcf33d0fa0150896477a32304bd09c31c83fec771c3aec3d7578e7f98d8ed1e8"} Dec 06 06:17:28 crc kubenswrapper[4809]: I1206 06:17:28.402504 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7899fdcd68-m4gx6" Dec 06 06:17:28 crc kubenswrapper[4809]: I1206 06:17:28.695257 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5675ff94b4-kfxhm" Dec 06 06:17:28 crc kubenswrapper[4809]: I1206 06:17:28.700777 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5675ff94b4-kfxhm" Dec 06 06:17:29 crc kubenswrapper[4809]: I1206 06:17:29.334581 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-xs2c8" Dec 06 06:17:29 crc kubenswrapper[4809]: I1206 06:17:29.363758 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-xs2c8" Dec 06 06:17:29 crc kubenswrapper[4809]: I1206 06:17:29.364200 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-xs2c8" event={"ID":"c75cf833-5179-4e18-8f74-3faed72757ed","Type":"ContainerDied","Data":"cd2c57afa753ad81e6088dc8cff8286bd25fc0a3b10e34b89956257ae4809bc3"} Dec 06 06:17:29 crc kubenswrapper[4809]: I1206 06:17:29.364245 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd2c57afa753ad81e6088dc8cff8286bd25fc0a3b10e34b89956257ae4809bc3" Dec 06 06:17:29 crc kubenswrapper[4809]: I1206 06:17:29.538245 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c75cf833-5179-4e18-8f74-3faed72757ed-combined-ca-bundle\") pod \"c75cf833-5179-4e18-8f74-3faed72757ed\" (UID: \"c75cf833-5179-4e18-8f74-3faed72757ed\") " Dec 06 06:17:29 crc kubenswrapper[4809]: I1206 06:17:29.538670 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c75cf833-5179-4e18-8f74-3faed72757ed-etc-machine-id\") pod \"c75cf833-5179-4e18-8f74-3faed72757ed\" (UID: \"c75cf833-5179-4e18-8f74-3faed72757ed\") " Dec 06 06:17:29 crc kubenswrapper[4809]: I1206 06:17:29.538735 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c75cf833-5179-4e18-8f74-3faed72757ed-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "c75cf833-5179-4e18-8f74-3faed72757ed" (UID: "c75cf833-5179-4e18-8f74-3faed72757ed"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:17:29 crc kubenswrapper[4809]: I1206 06:17:29.538819 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxm74\" (UniqueName: \"kubernetes.io/projected/c75cf833-5179-4e18-8f74-3faed72757ed-kube-api-access-wxm74\") pod \"c75cf833-5179-4e18-8f74-3faed72757ed\" (UID: \"c75cf833-5179-4e18-8f74-3faed72757ed\") " Dec 06 06:17:29 crc kubenswrapper[4809]: I1206 06:17:29.538949 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c75cf833-5179-4e18-8f74-3faed72757ed-db-sync-config-data\") pod \"c75cf833-5179-4e18-8f74-3faed72757ed\" (UID: \"c75cf833-5179-4e18-8f74-3faed72757ed\") " Dec 06 06:17:29 crc kubenswrapper[4809]: I1206 06:17:29.539008 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c75cf833-5179-4e18-8f74-3faed72757ed-scripts\") pod \"c75cf833-5179-4e18-8f74-3faed72757ed\" (UID: \"c75cf833-5179-4e18-8f74-3faed72757ed\") " Dec 06 06:17:29 crc kubenswrapper[4809]: I1206 06:17:29.539040 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c75cf833-5179-4e18-8f74-3faed72757ed-config-data\") pod \"c75cf833-5179-4e18-8f74-3faed72757ed\" (UID: \"c75cf833-5179-4e18-8f74-3faed72757ed\") " Dec 06 06:17:29 crc kubenswrapper[4809]: I1206 06:17:29.539629 4809 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c75cf833-5179-4e18-8f74-3faed72757ed-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:29 crc kubenswrapper[4809]: I1206 06:17:29.555249 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c75cf833-5179-4e18-8f74-3faed72757ed-kube-api-access-wxm74" (OuterVolumeSpecName: "kube-api-access-wxm74") pod "c75cf833-5179-4e18-8f74-3faed72757ed" (UID: "c75cf833-5179-4e18-8f74-3faed72757ed"). InnerVolumeSpecName "kube-api-access-wxm74". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:17:29 crc kubenswrapper[4809]: I1206 06:17:29.555510 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c75cf833-5179-4e18-8f74-3faed72757ed-scripts" (OuterVolumeSpecName: "scripts") pod "c75cf833-5179-4e18-8f74-3faed72757ed" (UID: "c75cf833-5179-4e18-8f74-3faed72757ed"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:17:29 crc kubenswrapper[4809]: I1206 06:17:29.568117 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c75cf833-5179-4e18-8f74-3faed72757ed-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "c75cf833-5179-4e18-8f74-3faed72757ed" (UID: "c75cf833-5179-4e18-8f74-3faed72757ed"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:17:29 crc kubenswrapper[4809]: I1206 06:17:29.587169 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c75cf833-5179-4e18-8f74-3faed72757ed-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c75cf833-5179-4e18-8f74-3faed72757ed" (UID: "c75cf833-5179-4e18-8f74-3faed72757ed"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:17:29 crc kubenswrapper[4809]: I1206 06:17:29.641757 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c75cf833-5179-4e18-8f74-3faed72757ed-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:29 crc kubenswrapper[4809]: I1206 06:17:29.641800 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxm74\" (UniqueName: \"kubernetes.io/projected/c75cf833-5179-4e18-8f74-3faed72757ed-kube-api-access-wxm74\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:29 crc kubenswrapper[4809]: I1206 06:17:29.641812 4809 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c75cf833-5179-4e18-8f74-3faed72757ed-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:29 crc kubenswrapper[4809]: I1206 06:17:29.641821 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c75cf833-5179-4e18-8f74-3faed72757ed-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:29 crc kubenswrapper[4809]: I1206 06:17:29.687604 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c75cf833-5179-4e18-8f74-3faed72757ed-config-data" (OuterVolumeSpecName: "config-data") pod "c75cf833-5179-4e18-8f74-3faed72757ed" (UID: "c75cf833-5179-4e18-8f74-3faed72757ed"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:17:29 crc kubenswrapper[4809]: I1206 06:17:29.744036 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c75cf833-5179-4e18-8f74-3faed72757ed-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:30 crc kubenswrapper[4809]: I1206 06:17:30.776721 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 06:17:30 crc kubenswrapper[4809]: E1206 06:17:30.777653 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c75cf833-5179-4e18-8f74-3faed72757ed" containerName="cinder-db-sync" Dec 06 06:17:30 crc kubenswrapper[4809]: I1206 06:17:30.777672 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c75cf833-5179-4e18-8f74-3faed72757ed" containerName="cinder-db-sync" Dec 06 06:17:30 crc kubenswrapper[4809]: I1206 06:17:30.777997 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="c75cf833-5179-4e18-8f74-3faed72757ed" containerName="cinder-db-sync" Dec 06 06:17:30 crc kubenswrapper[4809]: I1206 06:17:30.779687 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 06 06:17:30 crc kubenswrapper[4809]: I1206 06:17:30.783696 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 06 06:17:30 crc kubenswrapper[4809]: I1206 06:17:30.783968 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-hwb92" Dec 06 06:17:30 crc kubenswrapper[4809]: I1206 06:17:30.788311 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 06 06:17:30 crc kubenswrapper[4809]: I1206 06:17:30.788547 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 06 06:17:30 crc kubenswrapper[4809]: I1206 06:17:30.806993 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 06:17:30 crc kubenswrapper[4809]: I1206 06:17:30.873175 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-6dldb"] Dec 06 06:17:30 crc kubenswrapper[4809]: I1206 06:17:30.875067 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-6dldb" Dec 06 06:17:30 crc kubenswrapper[4809]: I1206 06:17:30.915686 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-6dldb"] Dec 06 06:17:30 crc kubenswrapper[4809]: I1206 06:17:30.971468 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8757fd43-0a69-40c8-8832-53a4acd0a912-config\") pod \"dnsmasq-dns-5c9776ccc5-6dldb\" (UID: \"8757fd43-0a69-40c8-8832-53a4acd0a912\") " pod="openstack/dnsmasq-dns-5c9776ccc5-6dldb" Dec 06 06:17:30 crc kubenswrapper[4809]: I1206 06:17:30.971530 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61b3cb73-1301-4073-9492-ef9f6810be3f-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"61b3cb73-1301-4073-9492-ef9f6810be3f\") " pod="openstack/cinder-scheduler-0" Dec 06 06:17:30 crc kubenswrapper[4809]: I1206 06:17:30.971571 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6jm8\" (UniqueName: \"kubernetes.io/projected/61b3cb73-1301-4073-9492-ef9f6810be3f-kube-api-access-c6jm8\") pod \"cinder-scheduler-0\" (UID: \"61b3cb73-1301-4073-9492-ef9f6810be3f\") " pod="openstack/cinder-scheduler-0" Dec 06 06:17:30 crc kubenswrapper[4809]: I1206 06:17:30.971610 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8757fd43-0a69-40c8-8832-53a4acd0a912-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-6dldb\" (UID: \"8757fd43-0a69-40c8-8832-53a4acd0a912\") " pod="openstack/dnsmasq-dns-5c9776ccc5-6dldb" Dec 06 06:17:30 crc kubenswrapper[4809]: I1206 06:17:30.971627 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61b3cb73-1301-4073-9492-ef9f6810be3f-config-data\") pod \"cinder-scheduler-0\" (UID: \"61b3cb73-1301-4073-9492-ef9f6810be3f\") " pod="openstack/cinder-scheduler-0" Dec 06 06:17:30 crc kubenswrapper[4809]: I1206 06:17:30.971643 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/61b3cb73-1301-4073-9492-ef9f6810be3f-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"61b3cb73-1301-4073-9492-ef9f6810be3f\") " pod="openstack/cinder-scheduler-0" Dec 06 06:17:30 crc kubenswrapper[4809]: I1206 06:17:30.971667 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/61b3cb73-1301-4073-9492-ef9f6810be3f-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"61b3cb73-1301-4073-9492-ef9f6810be3f\") " pod="openstack/cinder-scheduler-0" Dec 06 06:17:30 crc kubenswrapper[4809]: I1206 06:17:30.971715 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbv45\" (UniqueName: \"kubernetes.io/projected/8757fd43-0a69-40c8-8832-53a4acd0a912-kube-api-access-bbv45\") pod \"dnsmasq-dns-5c9776ccc5-6dldb\" (UID: \"8757fd43-0a69-40c8-8832-53a4acd0a912\") " pod="openstack/dnsmasq-dns-5c9776ccc5-6dldb" Dec 06 06:17:30 crc kubenswrapper[4809]: I1206 06:17:30.971738 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8757fd43-0a69-40c8-8832-53a4acd0a912-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-6dldb\" (UID: \"8757fd43-0a69-40c8-8832-53a4acd0a912\") " pod="openstack/dnsmasq-dns-5c9776ccc5-6dldb" Dec 06 06:17:30 crc kubenswrapper[4809]: I1206 06:17:30.971756 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/61b3cb73-1301-4073-9492-ef9f6810be3f-scripts\") pod \"cinder-scheduler-0\" (UID: \"61b3cb73-1301-4073-9492-ef9f6810be3f\") " pod="openstack/cinder-scheduler-0" Dec 06 06:17:30 crc kubenswrapper[4809]: I1206 06:17:30.971785 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8757fd43-0a69-40c8-8832-53a4acd0a912-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-6dldb\" (UID: \"8757fd43-0a69-40c8-8832-53a4acd0a912\") " pod="openstack/dnsmasq-dns-5c9776ccc5-6dldb" Dec 06 06:17:30 crc kubenswrapper[4809]: I1206 06:17:30.971817 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8757fd43-0a69-40c8-8832-53a4acd0a912-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-6dldb\" (UID: \"8757fd43-0a69-40c8-8832-53a4acd0a912\") " pod="openstack/dnsmasq-dns-5c9776ccc5-6dldb" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.035802 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.037654 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.046598 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.063647 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.074829 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/61b3cb73-1301-4073-9492-ef9f6810be3f-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"61b3cb73-1301-4073-9492-ef9f6810be3f\") " pod="openstack/cinder-scheduler-0" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.074978 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbv45\" (UniqueName: \"kubernetes.io/projected/8757fd43-0a69-40c8-8832-53a4acd0a912-kube-api-access-bbv45\") pod \"dnsmasq-dns-5c9776ccc5-6dldb\" (UID: \"8757fd43-0a69-40c8-8832-53a4acd0a912\") " pod="openstack/dnsmasq-dns-5c9776ccc5-6dldb" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.075018 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8757fd43-0a69-40c8-8832-53a4acd0a912-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-6dldb\" (UID: \"8757fd43-0a69-40c8-8832-53a4acd0a912\") " pod="openstack/dnsmasq-dns-5c9776ccc5-6dldb" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.075052 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/61b3cb73-1301-4073-9492-ef9f6810be3f-scripts\") pod \"cinder-scheduler-0\" (UID: \"61b3cb73-1301-4073-9492-ef9f6810be3f\") " pod="openstack/cinder-scheduler-0" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.075096 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8757fd43-0a69-40c8-8832-53a4acd0a912-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-6dldb\" (UID: \"8757fd43-0a69-40c8-8832-53a4acd0a912\") " pod="openstack/dnsmasq-dns-5c9776ccc5-6dldb" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.075158 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8757fd43-0a69-40c8-8832-53a4acd0a912-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-6dldb\" (UID: \"8757fd43-0a69-40c8-8832-53a4acd0a912\") " pod="openstack/dnsmasq-dns-5c9776ccc5-6dldb" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.075302 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8757fd43-0a69-40c8-8832-53a4acd0a912-config\") pod \"dnsmasq-dns-5c9776ccc5-6dldb\" (UID: \"8757fd43-0a69-40c8-8832-53a4acd0a912\") " pod="openstack/dnsmasq-dns-5c9776ccc5-6dldb" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.075339 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61b3cb73-1301-4073-9492-ef9f6810be3f-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"61b3cb73-1301-4073-9492-ef9f6810be3f\") " pod="openstack/cinder-scheduler-0" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.075397 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6jm8\" (UniqueName: \"kubernetes.io/projected/61b3cb73-1301-4073-9492-ef9f6810be3f-kube-api-access-c6jm8\") pod \"cinder-scheduler-0\" (UID: \"61b3cb73-1301-4073-9492-ef9f6810be3f\") " pod="openstack/cinder-scheduler-0" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.075451 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8757fd43-0a69-40c8-8832-53a4acd0a912-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-6dldb\" (UID: \"8757fd43-0a69-40c8-8832-53a4acd0a912\") " pod="openstack/dnsmasq-dns-5c9776ccc5-6dldb" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.075485 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61b3cb73-1301-4073-9492-ef9f6810be3f-config-data\") pod \"cinder-scheduler-0\" (UID: \"61b3cb73-1301-4073-9492-ef9f6810be3f\") " pod="openstack/cinder-scheduler-0" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.075504 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/61b3cb73-1301-4073-9492-ef9f6810be3f-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"61b3cb73-1301-4073-9492-ef9f6810be3f\") " pod="openstack/cinder-scheduler-0" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.078467 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8757fd43-0a69-40c8-8832-53a4acd0a912-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-6dldb\" (UID: \"8757fd43-0a69-40c8-8832-53a4acd0a912\") " pod="openstack/dnsmasq-dns-5c9776ccc5-6dldb" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.078504 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8757fd43-0a69-40c8-8832-53a4acd0a912-config\") pod \"dnsmasq-dns-5c9776ccc5-6dldb\" (UID: \"8757fd43-0a69-40c8-8832-53a4acd0a912\") " pod="openstack/dnsmasq-dns-5c9776ccc5-6dldb" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.078528 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/61b3cb73-1301-4073-9492-ef9f6810be3f-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"61b3cb73-1301-4073-9492-ef9f6810be3f\") " pod="openstack/cinder-scheduler-0" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.079350 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8757fd43-0a69-40c8-8832-53a4acd0a912-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-6dldb\" (UID: \"8757fd43-0a69-40c8-8832-53a4acd0a912\") " pod="openstack/dnsmasq-dns-5c9776ccc5-6dldb" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.079843 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8757fd43-0a69-40c8-8832-53a4acd0a912-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-6dldb\" (UID: \"8757fd43-0a69-40c8-8832-53a4acd0a912\") " pod="openstack/dnsmasq-dns-5c9776ccc5-6dldb" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.086792 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8757fd43-0a69-40c8-8832-53a4acd0a912-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-6dldb\" (UID: \"8757fd43-0a69-40c8-8832-53a4acd0a912\") " pod="openstack/dnsmasq-dns-5c9776ccc5-6dldb" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.088052 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/61b3cb73-1301-4073-9492-ef9f6810be3f-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"61b3cb73-1301-4073-9492-ef9f6810be3f\") " pod="openstack/cinder-scheduler-0" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.103882 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61b3cb73-1301-4073-9492-ef9f6810be3f-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"61b3cb73-1301-4073-9492-ef9f6810be3f\") " pod="openstack/cinder-scheduler-0" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.104428 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/61b3cb73-1301-4073-9492-ef9f6810be3f-scripts\") pod \"cinder-scheduler-0\" (UID: \"61b3cb73-1301-4073-9492-ef9f6810be3f\") " pod="openstack/cinder-scheduler-0" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.104543 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61b3cb73-1301-4073-9492-ef9f6810be3f-config-data\") pod \"cinder-scheduler-0\" (UID: \"61b3cb73-1301-4073-9492-ef9f6810be3f\") " pod="openstack/cinder-scheduler-0" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.108970 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbv45\" (UniqueName: \"kubernetes.io/projected/8757fd43-0a69-40c8-8832-53a4acd0a912-kube-api-access-bbv45\") pod \"dnsmasq-dns-5c9776ccc5-6dldb\" (UID: \"8757fd43-0a69-40c8-8832-53a4acd0a912\") " pod="openstack/dnsmasq-dns-5c9776ccc5-6dldb" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.113539 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6jm8\" (UniqueName: \"kubernetes.io/projected/61b3cb73-1301-4073-9492-ef9f6810be3f-kube-api-access-c6jm8\") pod \"cinder-scheduler-0\" (UID: \"61b3cb73-1301-4073-9492-ef9f6810be3f\") " pod="openstack/cinder-scheduler-0" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.155138 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.176896 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cf04523-ccb3-40bd-93d1-6dd59edf7530-config-data\") pod \"cinder-api-0\" (UID: \"7cf04523-ccb3-40bd-93d1-6dd59edf7530\") " pod="openstack/cinder-api-0" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.177112 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7cf04523-ccb3-40bd-93d1-6dd59edf7530-etc-machine-id\") pod \"cinder-api-0\" (UID: \"7cf04523-ccb3-40bd-93d1-6dd59edf7530\") " pod="openstack/cinder-api-0" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.177134 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7cf04523-ccb3-40bd-93d1-6dd59edf7530-scripts\") pod \"cinder-api-0\" (UID: \"7cf04523-ccb3-40bd-93d1-6dd59edf7530\") " pod="openstack/cinder-api-0" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.177158 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9zlt\" (UniqueName: \"kubernetes.io/projected/7cf04523-ccb3-40bd-93d1-6dd59edf7530-kube-api-access-h9zlt\") pod \"cinder-api-0\" (UID: \"7cf04523-ccb3-40bd-93d1-6dd59edf7530\") " pod="openstack/cinder-api-0" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.177207 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7cf04523-ccb3-40bd-93d1-6dd59edf7530-config-data-custom\") pod \"cinder-api-0\" (UID: \"7cf04523-ccb3-40bd-93d1-6dd59edf7530\") " pod="openstack/cinder-api-0" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.177229 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7cf04523-ccb3-40bd-93d1-6dd59edf7530-logs\") pod \"cinder-api-0\" (UID: \"7cf04523-ccb3-40bd-93d1-6dd59edf7530\") " pod="openstack/cinder-api-0" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.177245 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cf04523-ccb3-40bd-93d1-6dd59edf7530-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"7cf04523-ccb3-40bd-93d1-6dd59edf7530\") " pod="openstack/cinder-api-0" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.281190 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cf04523-ccb3-40bd-93d1-6dd59edf7530-config-data\") pod \"cinder-api-0\" (UID: \"7cf04523-ccb3-40bd-93d1-6dd59edf7530\") " pod="openstack/cinder-api-0" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.281787 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7cf04523-ccb3-40bd-93d1-6dd59edf7530-etc-machine-id\") pod \"cinder-api-0\" (UID: \"7cf04523-ccb3-40bd-93d1-6dd59edf7530\") " pod="openstack/cinder-api-0" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.281808 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7cf04523-ccb3-40bd-93d1-6dd59edf7530-scripts\") pod \"cinder-api-0\" (UID: \"7cf04523-ccb3-40bd-93d1-6dd59edf7530\") " pod="openstack/cinder-api-0" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.281824 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9zlt\" (UniqueName: \"kubernetes.io/projected/7cf04523-ccb3-40bd-93d1-6dd59edf7530-kube-api-access-h9zlt\") pod \"cinder-api-0\" (UID: \"7cf04523-ccb3-40bd-93d1-6dd59edf7530\") " pod="openstack/cinder-api-0" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.281870 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7cf04523-ccb3-40bd-93d1-6dd59edf7530-config-data-custom\") pod \"cinder-api-0\" (UID: \"7cf04523-ccb3-40bd-93d1-6dd59edf7530\") " pod="openstack/cinder-api-0" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.281894 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7cf04523-ccb3-40bd-93d1-6dd59edf7530-logs\") pod \"cinder-api-0\" (UID: \"7cf04523-ccb3-40bd-93d1-6dd59edf7530\") " pod="openstack/cinder-api-0" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.281910 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cf04523-ccb3-40bd-93d1-6dd59edf7530-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"7cf04523-ccb3-40bd-93d1-6dd59edf7530\") " pod="openstack/cinder-api-0" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.282283 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7cf04523-ccb3-40bd-93d1-6dd59edf7530-etc-machine-id\") pod \"cinder-api-0\" (UID: \"7cf04523-ccb3-40bd-93d1-6dd59edf7530\") " pod="openstack/cinder-api-0" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.282567 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7cf04523-ccb3-40bd-93d1-6dd59edf7530-logs\") pod \"cinder-api-0\" (UID: \"7cf04523-ccb3-40bd-93d1-6dd59edf7530\") " pod="openstack/cinder-api-0" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.287391 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7cf04523-ccb3-40bd-93d1-6dd59edf7530-scripts\") pod \"cinder-api-0\" (UID: \"7cf04523-ccb3-40bd-93d1-6dd59edf7530\") " pod="openstack/cinder-api-0" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.292106 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cf04523-ccb3-40bd-93d1-6dd59edf7530-config-data\") pod \"cinder-api-0\" (UID: \"7cf04523-ccb3-40bd-93d1-6dd59edf7530\") " pod="openstack/cinder-api-0" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.293283 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cf04523-ccb3-40bd-93d1-6dd59edf7530-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"7cf04523-ccb3-40bd-93d1-6dd59edf7530\") " pod="openstack/cinder-api-0" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.302775 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7cf04523-ccb3-40bd-93d1-6dd59edf7530-config-data-custom\") pod \"cinder-api-0\" (UID: \"7cf04523-ccb3-40bd-93d1-6dd59edf7530\") " pod="openstack/cinder-api-0" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.306459 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9zlt\" (UniqueName: \"kubernetes.io/projected/7cf04523-ccb3-40bd-93d1-6dd59edf7530-kube-api-access-h9zlt\") pod \"cinder-api-0\" (UID: \"7cf04523-ccb3-40bd-93d1-6dd59edf7530\") " pod="openstack/cinder-api-0" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.330159 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-6dldb" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.363540 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.844329 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-988458f5c-5zk66" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.929318 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5bb7f9f986-k7477" Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.936315 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7899fdcd68-m4gx6"] Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.936590 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7899fdcd68-m4gx6" podUID="4bba6eef-372a-4b9a-a27c-76f834e3519b" containerName="neutron-api" containerID="cri-o://79601c67c345fc5970db8138981011bc46985c80eaa23e7012e78c7dc020e526" gracePeriod=30 Dec 06 06:17:31 crc kubenswrapper[4809]: I1206 06:17:31.936759 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7899fdcd68-m4gx6" podUID="4bba6eef-372a-4b9a-a27c-76f834e3519b" containerName="neutron-httpd" containerID="cri-o://ec39c808be73c216788b8c2a0005c75163a22a80031d12660535a4f3142004ed" gracePeriod=30 Dec 06 06:17:32 crc kubenswrapper[4809]: I1206 06:17:32.058924 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5bb7f9f986-k7477" Dec 06 06:17:32 crc kubenswrapper[4809]: I1206 06:17:32.149499 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5675ff94b4-kfxhm"] Dec 06 06:17:32 crc kubenswrapper[4809]: I1206 06:17:32.149760 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5675ff94b4-kfxhm" podUID="db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9" containerName="barbican-api-log" containerID="cri-o://594bd3483b5ff6789ef10412230d25201c6a9a66c05826c4f8915a1a7976f0d3" gracePeriod=30 Dec 06 06:17:32 crc kubenswrapper[4809]: I1206 06:17:32.151030 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5675ff94b4-kfxhm" podUID="db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9" containerName="barbican-api" containerID="cri-o://b9295161ebe317bbd2d4cba00e6910036c3f9d287808b173f9dbffad7a300f73" gracePeriod=30 Dec 06 06:17:32 crc kubenswrapper[4809]: I1206 06:17:32.427176 4809 generic.go:334] "Generic (PLEG): container finished" podID="db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9" containerID="594bd3483b5ff6789ef10412230d25201c6a9a66c05826c4f8915a1a7976f0d3" exitCode=143 Dec 06 06:17:32 crc kubenswrapper[4809]: I1206 06:17:32.427271 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5675ff94b4-kfxhm" event={"ID":"db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9","Type":"ContainerDied","Data":"594bd3483b5ff6789ef10412230d25201c6a9a66c05826c4f8915a1a7976f0d3"} Dec 06 06:17:32 crc kubenswrapper[4809]: I1206 06:17:32.437113 4809 generic.go:334] "Generic (PLEG): container finished" podID="4bba6eef-372a-4b9a-a27c-76f834e3519b" containerID="ec39c808be73c216788b8c2a0005c75163a22a80031d12660535a4f3142004ed" exitCode=0 Dec 06 06:17:32 crc kubenswrapper[4809]: I1206 06:17:32.437303 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7899fdcd68-m4gx6" event={"ID":"4bba6eef-372a-4b9a-a27c-76f834e3519b","Type":"ContainerDied","Data":"ec39c808be73c216788b8c2a0005c75163a22a80031d12660535a4f3142004ed"} Dec 06 06:17:33 crc kubenswrapper[4809]: I1206 06:17:33.129798 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 06 06:17:34 crc kubenswrapper[4809]: I1206 06:17:34.497623 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:17:34 crc kubenswrapper[4809]: I1206 06:17:34.498089 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.261754 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5675ff94b4-kfxhm" podUID="db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.196:9311/healthcheck\": dial tcp 10.217.0.196:9311: connect: connection refused" Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.262033 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5675ff94b4-kfxhm" podUID="db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.196:9311/healthcheck\": dial tcp 10.217.0.196:9311: connect: connection refused" Dec 06 06:17:36 crc kubenswrapper[4809]: E1206 06:17:36.295848 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/ubi9/httpd-24:latest" Dec 06 06:17:36 crc kubenswrapper[4809]: E1206 06:17:36.296265 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:proxy-httpd,Image:registry.redhat.io/ubi9/httpd-24:latest,Command:[/usr/sbin/httpd],Args:[-DFOREGROUND],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:proxy-httpd,HostPort:0,ContainerPort:3000,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf/httpd.conf,SubPath:httpd.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf.d/ssl.conf,SubPath:ssl.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:run-httpd,ReadOnly:false,MountPath:/run/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:log-httpd,ReadOnly:false,MountPath:/var/log/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6zc4q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(93e2d67f-9481-4ca3-83e7-4f7c0673b9d4): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 06:17:36 crc kubenswrapper[4809]: E1206 06:17:36.297828 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"proxy-httpd\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="93e2d67f-9481-4ca3-83e7-4f7c0673b9d4" Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.319478 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-58svx" Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.410404 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ecb6c22c-402b-4bdb-a791-29509985785a-dns-swift-storage-0\") pod \"ecb6c22c-402b-4bdb-a791-29509985785a\" (UID: \"ecb6c22c-402b-4bdb-a791-29509985785a\") " Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.410645 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecb6c22c-402b-4bdb-a791-29509985785a-config\") pod \"ecb6c22c-402b-4bdb-a791-29509985785a\" (UID: \"ecb6c22c-402b-4bdb-a791-29509985785a\") " Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.410706 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ecb6c22c-402b-4bdb-a791-29509985785a-ovsdbserver-sb\") pod \"ecb6c22c-402b-4bdb-a791-29509985785a\" (UID: \"ecb6c22c-402b-4bdb-a791-29509985785a\") " Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.411153 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ecb6c22c-402b-4bdb-a791-29509985785a-ovsdbserver-nb\") pod \"ecb6c22c-402b-4bdb-a791-29509985785a\" (UID: \"ecb6c22c-402b-4bdb-a791-29509985785a\") " Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.411215 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ecb6c22c-402b-4bdb-a791-29509985785a-dns-svc\") pod \"ecb6c22c-402b-4bdb-a791-29509985785a\" (UID: \"ecb6c22c-402b-4bdb-a791-29509985785a\") " Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.411334 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8qj7\" (UniqueName: \"kubernetes.io/projected/ecb6c22c-402b-4bdb-a791-29509985785a-kube-api-access-t8qj7\") pod \"ecb6c22c-402b-4bdb-a791-29509985785a\" (UID: \"ecb6c22c-402b-4bdb-a791-29509985785a\") " Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.422968 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecb6c22c-402b-4bdb-a791-29509985785a-kube-api-access-t8qj7" (OuterVolumeSpecName: "kube-api-access-t8qj7") pod "ecb6c22c-402b-4bdb-a791-29509985785a" (UID: "ecb6c22c-402b-4bdb-a791-29509985785a"). InnerVolumeSpecName "kube-api-access-t8qj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.477280 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ecb6c22c-402b-4bdb-a791-29509985785a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ecb6c22c-402b-4bdb-a791-29509985785a" (UID: "ecb6c22c-402b-4bdb-a791-29509985785a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.488217 4809 generic.go:334] "Generic (PLEG): container finished" podID="db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9" containerID="b9295161ebe317bbd2d4cba00e6910036c3f9d287808b173f9dbffad7a300f73" exitCode=0 Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.488269 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5675ff94b4-kfxhm" event={"ID":"db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9","Type":"ContainerDied","Data":"b9295161ebe317bbd2d4cba00e6910036c3f9d287808b173f9dbffad7a300f73"} Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.491027 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="93e2d67f-9481-4ca3-83e7-4f7c0673b9d4" containerName="ceilometer-central-agent" containerID="cri-o://081257dd2f4b335328e16317962b5d8e7e6fec0da5505d2793bea03badf12dd1" gracePeriod=30 Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.491124 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-58svx" Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.491677 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-58svx" event={"ID":"ecb6c22c-402b-4bdb-a791-29509985785a","Type":"ContainerDied","Data":"3ef6aa1add3472dbb20bb435c15329561d4fb74293136cdaae01957e26374cec"} Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.491712 4809 scope.go:117] "RemoveContainer" containerID="bcf33d0fa0150896477a32304bd09c31c83fec771c3aec3d7578e7f98d8ed1e8" Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.492131 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="93e2d67f-9481-4ca3-83e7-4f7c0673b9d4" containerName="sg-core" containerID="cri-o://04ae50b6ffa0e470d79b81e343165917548537b8df78e6568beeaa62cb16d06e" gracePeriod=30 Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.492163 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ecb6c22c-402b-4bdb-a791-29509985785a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ecb6c22c-402b-4bdb-a791-29509985785a" (UID: "ecb6c22c-402b-4bdb-a791-29509985785a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.492179 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="93e2d67f-9481-4ca3-83e7-4f7c0673b9d4" containerName="ceilometer-notification-agent" containerID="cri-o://ce4c3b3f255d533aa317c246f5f07360cdecb16cf7ab4271977b7f91345a258f" gracePeriod=30 Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.495311 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ecb6c22c-402b-4bdb-a791-29509985785a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ecb6c22c-402b-4bdb-a791-29509985785a" (UID: "ecb6c22c-402b-4bdb-a791-29509985785a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.515877 4809 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ecb6c22c-402b-4bdb-a791-29509985785a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.515913 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ecb6c22c-402b-4bdb-a791-29509985785a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.515922 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ecb6c22c-402b-4bdb-a791-29509985785a-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.515944 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8qj7\" (UniqueName: \"kubernetes.io/projected/ecb6c22c-402b-4bdb-a791-29509985785a-kube-api-access-t8qj7\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.563769 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ecb6c22c-402b-4bdb-a791-29509985785a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ecb6c22c-402b-4bdb-a791-29509985785a" (UID: "ecb6c22c-402b-4bdb-a791-29509985785a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.578945 4809 scope.go:117] "RemoveContainer" containerID="29529fe529b79a26d45b811f45d236fc54331abc835dbe69763e64c6f471e0e8" Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.587219 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ecb6c22c-402b-4bdb-a791-29509985785a-config" (OuterVolumeSpecName: "config") pod "ecb6c22c-402b-4bdb-a791-29509985785a" (UID: "ecb6c22c-402b-4bdb-a791-29509985785a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.619190 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecb6c22c-402b-4bdb-a791-29509985785a-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.619231 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ecb6c22c-402b-4bdb-a791-29509985785a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.723888 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5675ff94b4-kfxhm" Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.823109 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9-config-data\") pod \"db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9\" (UID: \"db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9\") " Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.823443 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9-logs\") pod \"db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9\" (UID: \"db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9\") " Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.823483 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9-config-data-custom\") pod \"db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9\" (UID: \"db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9\") " Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.823634 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvms5\" (UniqueName: \"kubernetes.io/projected/db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9-kube-api-access-pvms5\") pod \"db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9\" (UID: \"db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9\") " Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.823841 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9-combined-ca-bundle\") pod \"db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9\" (UID: \"db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9\") " Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.825553 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9-logs" (OuterVolumeSpecName: "logs") pod "db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9" (UID: "db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.831465 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9" (UID: "db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.838687 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-58svx"] Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.862857 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9-kube-api-access-pvms5" (OuterVolumeSpecName: "kube-api-access-pvms5") pod "db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9" (UID: "db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9"). InnerVolumeSpecName "kube-api-access-pvms5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.872998 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-58svx"] Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.883212 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9" (UID: "db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.897818 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9-config-data" (OuterVolumeSpecName: "config-data") pod "db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9" (UID: "db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.927213 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.927520 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9-logs\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.927601 4809 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.927679 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvms5\" (UniqueName: \"kubernetes.io/projected/db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9-kube-api-access-pvms5\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:36 crc kubenswrapper[4809]: I1206 06:17:36.927781 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:37 crc kubenswrapper[4809]: E1206 06:17:37.038884 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podecb6c22c_402b_4bdb_a791_29509985785a.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93e2d67f_9481_4ca3_83e7_4f7c0673b9d4.slice/crio-conmon-081257dd2f4b335328e16317962b5d8e7e6fec0da5505d2793bea03badf12dd1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93e2d67f_9481_4ca3_83e7_4f7c0673b9d4.slice/crio-081257dd2f4b335328e16317962b5d8e7e6fec0da5505d2793bea03badf12dd1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podecb6c22c_402b_4bdb_a791_29509985785a.slice/crio-3ef6aa1add3472dbb20bb435c15329561d4fb74293136cdaae01957e26374cec\": RecentStats: unable to find data in memory cache]" Dec 06 06:17:37 crc kubenswrapper[4809]: I1206 06:17:37.109014 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 06 06:17:37 crc kubenswrapper[4809]: I1206 06:17:37.122119 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 06:17:37 crc kubenswrapper[4809]: I1206 06:17:37.150418 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-6dldb"] Dec 06 06:17:37 crc kubenswrapper[4809]: W1206 06:17:37.165079 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod61b3cb73_1301_4073_9492_ef9f6810be3f.slice/crio-bab6899d52381c591757d6a96c7302eba4559ecefbf8783d11cb13c116c291f7 WatchSource:0}: Error finding container bab6899d52381c591757d6a96c7302eba4559ecefbf8783d11cb13c116c291f7: Status 404 returned error can't find the container with id bab6899d52381c591757d6a96c7302eba4559ecefbf8783d11cb13c116c291f7 Dec 06 06:17:37 crc kubenswrapper[4809]: I1206 06:17:37.414879 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecb6c22c-402b-4bdb-a791-29509985785a" path="/var/lib/kubelet/pods/ecb6c22c-402b-4bdb-a791-29509985785a/volumes" Dec 06 06:17:37 crc kubenswrapper[4809]: I1206 06:17:37.511860 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5675ff94b4-kfxhm" event={"ID":"db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9","Type":"ContainerDied","Data":"531badae15a0972281ef0ca64fd43b28aab8c0c00afac77c3ff2814d1ec069e6"} Dec 06 06:17:37 crc kubenswrapper[4809]: I1206 06:17:37.511925 4809 scope.go:117] "RemoveContainer" containerID="b9295161ebe317bbd2d4cba00e6910036c3f9d287808b173f9dbffad7a300f73" Dec 06 06:17:37 crc kubenswrapper[4809]: I1206 06:17:37.512092 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5675ff94b4-kfxhm" Dec 06 06:17:37 crc kubenswrapper[4809]: I1206 06:17:37.520177 4809 generic.go:334] "Generic (PLEG): container finished" podID="93e2d67f-9481-4ca3-83e7-4f7c0673b9d4" containerID="04ae50b6ffa0e470d79b81e343165917548537b8df78e6568beeaa62cb16d06e" exitCode=2 Dec 06 06:17:37 crc kubenswrapper[4809]: I1206 06:17:37.520208 4809 generic.go:334] "Generic (PLEG): container finished" podID="93e2d67f-9481-4ca3-83e7-4f7c0673b9d4" containerID="081257dd2f4b335328e16317962b5d8e7e6fec0da5505d2793bea03badf12dd1" exitCode=0 Dec 06 06:17:37 crc kubenswrapper[4809]: I1206 06:17:37.520256 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"93e2d67f-9481-4ca3-83e7-4f7c0673b9d4","Type":"ContainerDied","Data":"04ae50b6ffa0e470d79b81e343165917548537b8df78e6568beeaa62cb16d06e"} Dec 06 06:17:37 crc kubenswrapper[4809]: I1206 06:17:37.520302 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"93e2d67f-9481-4ca3-83e7-4f7c0673b9d4","Type":"ContainerDied","Data":"081257dd2f4b335328e16317962b5d8e7e6fec0da5505d2793bea03badf12dd1"} Dec 06 06:17:37 crc kubenswrapper[4809]: I1206 06:17:37.531517 4809 generic.go:334] "Generic (PLEG): container finished" podID="4bba6eef-372a-4b9a-a27c-76f834e3519b" containerID="79601c67c345fc5970db8138981011bc46985c80eaa23e7012e78c7dc020e526" exitCode=0 Dec 06 06:17:37 crc kubenswrapper[4809]: I1206 06:17:37.531603 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7899fdcd68-m4gx6" event={"ID":"4bba6eef-372a-4b9a-a27c-76f834e3519b","Type":"ContainerDied","Data":"79601c67c345fc5970db8138981011bc46985c80eaa23e7012e78c7dc020e526"} Dec 06 06:17:37 crc kubenswrapper[4809]: I1206 06:17:37.534313 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-6dldb" event={"ID":"8757fd43-0a69-40c8-8832-53a4acd0a912","Type":"ContainerStarted","Data":"c7b2a190e462fabd5f9d06e8ba8b95313e9ade5bf7e92b8ed03a2ecab9950af2"} Dec 06 06:17:37 crc kubenswrapper[4809]: I1206 06:17:37.534344 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-6dldb" event={"ID":"8757fd43-0a69-40c8-8832-53a4acd0a912","Type":"ContainerStarted","Data":"1d95a975f54748220b7d00d032185c87a19502fb67e16773b25d9f5beee93b5e"} Dec 06 06:17:37 crc kubenswrapper[4809]: I1206 06:17:37.537143 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"61b3cb73-1301-4073-9492-ef9f6810be3f","Type":"ContainerStarted","Data":"bab6899d52381c591757d6a96c7302eba4559ecefbf8783d11cb13c116c291f7"} Dec 06 06:17:37 crc kubenswrapper[4809]: I1206 06:17:37.539745 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7cf04523-ccb3-40bd-93d1-6dd59edf7530","Type":"ContainerStarted","Data":"c3daedaa81e8e056c9c83f7b697a72c46501badb356d2f5a02a2ba11ae0a47fe"} Dec 06 06:17:37 crc kubenswrapper[4809]: I1206 06:17:37.542712 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5675ff94b4-kfxhm"] Dec 06 06:17:37 crc kubenswrapper[4809]: I1206 06:17:37.562904 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-5675ff94b4-kfxhm"] Dec 06 06:17:37 crc kubenswrapper[4809]: I1206 06:17:37.569026 4809 scope.go:117] "RemoveContainer" containerID="594bd3483b5ff6789ef10412230d25201c6a9a66c05826c4f8915a1a7976f0d3" Dec 06 06:17:37 crc kubenswrapper[4809]: I1206 06:17:37.725899 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7899fdcd68-m4gx6" Dec 06 06:17:37 crc kubenswrapper[4809]: I1206 06:17:37.846324 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bba6eef-372a-4b9a-a27c-76f834e3519b-ovndb-tls-certs\") pod \"4bba6eef-372a-4b9a-a27c-76f834e3519b\" (UID: \"4bba6eef-372a-4b9a-a27c-76f834e3519b\") " Dec 06 06:17:37 crc kubenswrapper[4809]: I1206 06:17:37.846532 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-np4vn\" (UniqueName: \"kubernetes.io/projected/4bba6eef-372a-4b9a-a27c-76f834e3519b-kube-api-access-np4vn\") pod \"4bba6eef-372a-4b9a-a27c-76f834e3519b\" (UID: \"4bba6eef-372a-4b9a-a27c-76f834e3519b\") " Dec 06 06:17:37 crc kubenswrapper[4809]: I1206 06:17:37.846633 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4bba6eef-372a-4b9a-a27c-76f834e3519b-config\") pod \"4bba6eef-372a-4b9a-a27c-76f834e3519b\" (UID: \"4bba6eef-372a-4b9a-a27c-76f834e3519b\") " Dec 06 06:17:37 crc kubenswrapper[4809]: I1206 06:17:37.847215 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bba6eef-372a-4b9a-a27c-76f834e3519b-combined-ca-bundle\") pod \"4bba6eef-372a-4b9a-a27c-76f834e3519b\" (UID: \"4bba6eef-372a-4b9a-a27c-76f834e3519b\") " Dec 06 06:17:37 crc kubenswrapper[4809]: I1206 06:17:37.847340 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4bba6eef-372a-4b9a-a27c-76f834e3519b-httpd-config\") pod \"4bba6eef-372a-4b9a-a27c-76f834e3519b\" (UID: \"4bba6eef-372a-4b9a-a27c-76f834e3519b\") " Dec 06 06:17:37 crc kubenswrapper[4809]: I1206 06:17:37.852842 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bba6eef-372a-4b9a-a27c-76f834e3519b-kube-api-access-np4vn" (OuterVolumeSpecName: "kube-api-access-np4vn") pod "4bba6eef-372a-4b9a-a27c-76f834e3519b" (UID: "4bba6eef-372a-4b9a-a27c-76f834e3519b"). InnerVolumeSpecName "kube-api-access-np4vn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:17:37 crc kubenswrapper[4809]: I1206 06:17:37.855350 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bba6eef-372a-4b9a-a27c-76f834e3519b-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "4bba6eef-372a-4b9a-a27c-76f834e3519b" (UID: "4bba6eef-372a-4b9a-a27c-76f834e3519b"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:17:37 crc kubenswrapper[4809]: I1206 06:17:37.911744 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bba6eef-372a-4b9a-a27c-76f834e3519b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4bba6eef-372a-4b9a-a27c-76f834e3519b" (UID: "4bba6eef-372a-4b9a-a27c-76f834e3519b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:17:37 crc kubenswrapper[4809]: I1206 06:17:37.916140 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bba6eef-372a-4b9a-a27c-76f834e3519b-config" (OuterVolumeSpecName: "config") pod "4bba6eef-372a-4b9a-a27c-76f834e3519b" (UID: "4bba6eef-372a-4b9a-a27c-76f834e3519b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:17:37 crc kubenswrapper[4809]: I1206 06:17:37.951649 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-np4vn\" (UniqueName: \"kubernetes.io/projected/4bba6eef-372a-4b9a-a27c-76f834e3519b-kube-api-access-np4vn\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:37 crc kubenswrapper[4809]: I1206 06:17:37.951709 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/4bba6eef-372a-4b9a-a27c-76f834e3519b-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:37 crc kubenswrapper[4809]: I1206 06:17:37.951722 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bba6eef-372a-4b9a-a27c-76f834e3519b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:37 crc kubenswrapper[4809]: I1206 06:17:37.951735 4809 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4bba6eef-372a-4b9a-a27c-76f834e3519b-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:37 crc kubenswrapper[4809]: I1206 06:17:37.967553 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bba6eef-372a-4b9a-a27c-76f834e3519b-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "4bba6eef-372a-4b9a-a27c-76f834e3519b" (UID: "4bba6eef-372a-4b9a-a27c-76f834e3519b"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:17:38 crc kubenswrapper[4809]: I1206 06:17:38.053426 4809 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bba6eef-372a-4b9a-a27c-76f834e3519b-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:38 crc kubenswrapper[4809]: I1206 06:17:38.237248 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-55f844cf75-58svx" podUID="ecb6c22c-402b-4bdb-a791-29509985785a" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.186:5353: i/o timeout" Dec 06 06:17:38 crc kubenswrapper[4809]: I1206 06:17:38.583577 4809 generic.go:334] "Generic (PLEG): container finished" podID="8757fd43-0a69-40c8-8832-53a4acd0a912" containerID="c7b2a190e462fabd5f9d06e8ba8b95313e9ade5bf7e92b8ed03a2ecab9950af2" exitCode=0 Dec 06 06:17:38 crc kubenswrapper[4809]: I1206 06:17:38.583919 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-6dldb" event={"ID":"8757fd43-0a69-40c8-8832-53a4acd0a912","Type":"ContainerDied","Data":"c7b2a190e462fabd5f9d06e8ba8b95313e9ade5bf7e92b8ed03a2ecab9950af2"} Dec 06 06:17:38 crc kubenswrapper[4809]: I1206 06:17:38.597910 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7cf04523-ccb3-40bd-93d1-6dd59edf7530","Type":"ContainerStarted","Data":"1a542df6256ee56b5fbdc3ca23fb895140362d7cb61722388669e994937cd78d"} Dec 06 06:17:38 crc kubenswrapper[4809]: I1206 06:17:38.641402 4809 generic.go:334] "Generic (PLEG): container finished" podID="93e2d67f-9481-4ca3-83e7-4f7c0673b9d4" containerID="ce4c3b3f255d533aa317c246f5f07360cdecb16cf7ab4271977b7f91345a258f" exitCode=0 Dec 06 06:17:38 crc kubenswrapper[4809]: I1206 06:17:38.641812 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"93e2d67f-9481-4ca3-83e7-4f7c0673b9d4","Type":"ContainerDied","Data":"ce4c3b3f255d533aa317c246f5f07360cdecb16cf7ab4271977b7f91345a258f"} Dec 06 06:17:38 crc kubenswrapper[4809]: I1206 06:17:38.653359 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7899fdcd68-m4gx6" event={"ID":"4bba6eef-372a-4b9a-a27c-76f834e3519b","Type":"ContainerDied","Data":"8ede14d1ccf475536a392841b50ecdfa8c618338be8ca8321222153be7222ae0"} Dec 06 06:17:38 crc kubenswrapper[4809]: I1206 06:17:38.653421 4809 scope.go:117] "RemoveContainer" containerID="ec39c808be73c216788b8c2a0005c75163a22a80031d12660535a4f3142004ed" Dec 06 06:17:38 crc kubenswrapper[4809]: I1206 06:17:38.653525 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7899fdcd68-m4gx6" Dec 06 06:17:38 crc kubenswrapper[4809]: I1206 06:17:38.709111 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7899fdcd68-m4gx6"] Dec 06 06:17:38 crc kubenswrapper[4809]: I1206 06:17:38.722557 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-7899fdcd68-m4gx6"] Dec 06 06:17:38 crc kubenswrapper[4809]: I1206 06:17:38.793236 4809 scope.go:117] "RemoveContainer" containerID="79601c67c345fc5970db8138981011bc46985c80eaa23e7012e78c7dc020e526" Dec 06 06:17:38 crc kubenswrapper[4809]: I1206 06:17:38.866908 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:17:38 crc kubenswrapper[4809]: I1206 06:17:38.977658 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/93e2d67f-9481-4ca3-83e7-4f7c0673b9d4-sg-core-conf-yaml\") pod \"93e2d67f-9481-4ca3-83e7-4f7c0673b9d4\" (UID: \"93e2d67f-9481-4ca3-83e7-4f7c0673b9d4\") " Dec 06 06:17:38 crc kubenswrapper[4809]: I1206 06:17:38.977745 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93e2d67f-9481-4ca3-83e7-4f7c0673b9d4-scripts\") pod \"93e2d67f-9481-4ca3-83e7-4f7c0673b9d4\" (UID: \"93e2d67f-9481-4ca3-83e7-4f7c0673b9d4\") " Dec 06 06:17:38 crc kubenswrapper[4809]: I1206 06:17:38.977798 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93e2d67f-9481-4ca3-83e7-4f7c0673b9d4-config-data\") pod \"93e2d67f-9481-4ca3-83e7-4f7c0673b9d4\" (UID: \"93e2d67f-9481-4ca3-83e7-4f7c0673b9d4\") " Dec 06 06:17:38 crc kubenswrapper[4809]: I1206 06:17:38.977863 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/93e2d67f-9481-4ca3-83e7-4f7c0673b9d4-run-httpd\") pod \"93e2d67f-9481-4ca3-83e7-4f7c0673b9d4\" (UID: \"93e2d67f-9481-4ca3-83e7-4f7c0673b9d4\") " Dec 06 06:17:38 crc kubenswrapper[4809]: I1206 06:17:38.977882 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/93e2d67f-9481-4ca3-83e7-4f7c0673b9d4-log-httpd\") pod \"93e2d67f-9481-4ca3-83e7-4f7c0673b9d4\" (UID: \"93e2d67f-9481-4ca3-83e7-4f7c0673b9d4\") " Dec 06 06:17:38 crc kubenswrapper[4809]: I1206 06:17:38.977976 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93e2d67f-9481-4ca3-83e7-4f7c0673b9d4-combined-ca-bundle\") pod \"93e2d67f-9481-4ca3-83e7-4f7c0673b9d4\" (UID: \"93e2d67f-9481-4ca3-83e7-4f7c0673b9d4\") " Dec 06 06:17:38 crc kubenswrapper[4809]: I1206 06:17:38.978013 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6zc4q\" (UniqueName: \"kubernetes.io/projected/93e2d67f-9481-4ca3-83e7-4f7c0673b9d4-kube-api-access-6zc4q\") pod \"93e2d67f-9481-4ca3-83e7-4f7c0673b9d4\" (UID: \"93e2d67f-9481-4ca3-83e7-4f7c0673b9d4\") " Dec 06 06:17:38 crc kubenswrapper[4809]: I1206 06:17:38.984024 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93e2d67f-9481-4ca3-83e7-4f7c0673b9d4-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "93e2d67f-9481-4ca3-83e7-4f7c0673b9d4" (UID: "93e2d67f-9481-4ca3-83e7-4f7c0673b9d4"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:17:38 crc kubenswrapper[4809]: I1206 06:17:38.984276 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93e2d67f-9481-4ca3-83e7-4f7c0673b9d4-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "93e2d67f-9481-4ca3-83e7-4f7c0673b9d4" (UID: "93e2d67f-9481-4ca3-83e7-4f7c0673b9d4"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:17:38 crc kubenswrapper[4809]: I1206 06:17:38.987101 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93e2d67f-9481-4ca3-83e7-4f7c0673b9d4-scripts" (OuterVolumeSpecName: "scripts") pod "93e2d67f-9481-4ca3-83e7-4f7c0673b9d4" (UID: "93e2d67f-9481-4ca3-83e7-4f7c0673b9d4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:17:38 crc kubenswrapper[4809]: I1206 06:17:38.990134 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93e2d67f-9481-4ca3-83e7-4f7c0673b9d4-kube-api-access-6zc4q" (OuterVolumeSpecName: "kube-api-access-6zc4q") pod "93e2d67f-9481-4ca3-83e7-4f7c0673b9d4" (UID: "93e2d67f-9481-4ca3-83e7-4f7c0673b9d4"). InnerVolumeSpecName "kube-api-access-6zc4q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.043843 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93e2d67f-9481-4ca3-83e7-4f7c0673b9d4-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "93e2d67f-9481-4ca3-83e7-4f7c0673b9d4" (UID: "93e2d67f-9481-4ca3-83e7-4f7c0673b9d4"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.067445 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93e2d67f-9481-4ca3-83e7-4f7c0673b9d4-config-data" (OuterVolumeSpecName: "config-data") pod "93e2d67f-9481-4ca3-83e7-4f7c0673b9d4" (UID: "93e2d67f-9481-4ca3-83e7-4f7c0673b9d4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.080376 4809 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/93e2d67f-9481-4ca3-83e7-4f7c0673b9d4-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.080542 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93e2d67f-9481-4ca3-83e7-4f7c0673b9d4-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.080597 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93e2d67f-9481-4ca3-83e7-4f7c0673b9d4-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.086031 4809 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/93e2d67f-9481-4ca3-83e7-4f7c0673b9d4-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.086464 4809 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/93e2d67f-9481-4ca3-83e7-4f7c0673b9d4-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.086554 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6zc4q\" (UniqueName: \"kubernetes.io/projected/93e2d67f-9481-4ca3-83e7-4f7c0673b9d4-kube-api-access-6zc4q\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.113105 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93e2d67f-9481-4ca3-83e7-4f7c0673b9d4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "93e2d67f-9481-4ca3-83e7-4f7c0673b9d4" (UID: "93e2d67f-9481-4ca3-83e7-4f7c0673b9d4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.189520 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93e2d67f-9481-4ca3-83e7-4f7c0673b9d4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.420752 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bba6eef-372a-4b9a-a27c-76f834e3519b" path="/var/lib/kubelet/pods/4bba6eef-372a-4b9a-a27c-76f834e3519b/volumes" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.427192 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9" path="/var/lib/kubelet/pods/db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9/volumes" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.671346 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7cf04523-ccb3-40bd-93d1-6dd59edf7530","Type":"ContainerStarted","Data":"6db290f4a2b99fba5d0365f55cdced03d05a875cf48e355dd49d6d4a42fbb0e4"} Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.671829 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="7cf04523-ccb3-40bd-93d1-6dd59edf7530" containerName="cinder-api-log" containerID="cri-o://1a542df6256ee56b5fbdc3ca23fb895140362d7cb61722388669e994937cd78d" gracePeriod=30 Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.673174 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.673481 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="7cf04523-ccb3-40bd-93d1-6dd59edf7530" containerName="cinder-api" containerID="cri-o://6db290f4a2b99fba5d0365f55cdced03d05a875cf48e355dd49d6d4a42fbb0e4" gracePeriod=30 Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.678192 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"93e2d67f-9481-4ca3-83e7-4f7c0673b9d4","Type":"ContainerDied","Data":"063ec88f2d6bfed89c694189278a247c83add1e67252bd65335a8c84c9afd9f2"} Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.678234 4809 scope.go:117] "RemoveContainer" containerID="04ae50b6ffa0e470d79b81e343165917548537b8df78e6568beeaa62cb16d06e" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.678341 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.688518 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-6dldb" event={"ID":"8757fd43-0a69-40c8-8832-53a4acd0a912","Type":"ContainerStarted","Data":"f8f0dee995d02f998c3002eff13de0219a2a4ebafa905444e8aed7037afa2ce0"} Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.689249 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9776ccc5-6dldb" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.699329 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=8.699304336 podStartE2EDuration="8.699304336s" podCreationTimestamp="2025-12-06 06:17:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:17:39.697332302 +0000 UTC m=+1584.586315244" watchObservedRunningTime="2025-12-06 06:17:39.699304336 +0000 UTC m=+1584.588287278" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.745300 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9776ccc5-6dldb" podStartSLOduration=9.745280437 podStartE2EDuration="9.745280437s" podCreationTimestamp="2025-12-06 06:17:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:17:39.729548361 +0000 UTC m=+1584.618531343" watchObservedRunningTime="2025-12-06 06:17:39.745280437 +0000 UTC m=+1584.634263379" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.747789 4809 scope.go:117] "RemoveContainer" containerID="ce4c3b3f255d533aa317c246f5f07360cdecb16cf7ab4271977b7f91345a258f" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.799887 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.818012 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.822272 4809 scope.go:117] "RemoveContainer" containerID="081257dd2f4b335328e16317962b5d8e7e6fec0da5505d2793bea03badf12dd1" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.843030 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:17:39 crc kubenswrapper[4809]: E1206 06:17:39.843649 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bba6eef-372a-4b9a-a27c-76f834e3519b" containerName="neutron-httpd" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.843669 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bba6eef-372a-4b9a-a27c-76f834e3519b" containerName="neutron-httpd" Dec 06 06:17:39 crc kubenswrapper[4809]: E1206 06:17:39.843695 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecb6c22c-402b-4bdb-a791-29509985785a" containerName="dnsmasq-dns" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.843734 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecb6c22c-402b-4bdb-a791-29509985785a" containerName="dnsmasq-dns" Dec 06 06:17:39 crc kubenswrapper[4809]: E1206 06:17:39.843750 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecb6c22c-402b-4bdb-a791-29509985785a" containerName="init" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.843759 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecb6c22c-402b-4bdb-a791-29509985785a" containerName="init" Dec 06 06:17:39 crc kubenswrapper[4809]: E1206 06:17:39.843786 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9" containerName="barbican-api" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.843793 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9" containerName="barbican-api" Dec 06 06:17:39 crc kubenswrapper[4809]: E1206 06:17:39.843809 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93e2d67f-9481-4ca3-83e7-4f7c0673b9d4" containerName="ceilometer-notification-agent" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.843816 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="93e2d67f-9481-4ca3-83e7-4f7c0673b9d4" containerName="ceilometer-notification-agent" Dec 06 06:17:39 crc kubenswrapper[4809]: E1206 06:17:39.843826 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93e2d67f-9481-4ca3-83e7-4f7c0673b9d4" containerName="sg-core" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.843832 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="93e2d67f-9481-4ca3-83e7-4f7c0673b9d4" containerName="sg-core" Dec 06 06:17:39 crc kubenswrapper[4809]: E1206 06:17:39.843846 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93e2d67f-9481-4ca3-83e7-4f7c0673b9d4" containerName="ceilometer-central-agent" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.843853 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="93e2d67f-9481-4ca3-83e7-4f7c0673b9d4" containerName="ceilometer-central-agent" Dec 06 06:17:39 crc kubenswrapper[4809]: E1206 06:17:39.843865 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9" containerName="barbican-api-log" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.843872 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9" containerName="barbican-api-log" Dec 06 06:17:39 crc kubenswrapper[4809]: E1206 06:17:39.843887 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bba6eef-372a-4b9a-a27c-76f834e3519b" containerName="neutron-api" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.843894 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bba6eef-372a-4b9a-a27c-76f834e3519b" containerName="neutron-api" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.844194 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="93e2d67f-9481-4ca3-83e7-4f7c0673b9d4" containerName="ceilometer-notification-agent" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.844210 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9" containerName="barbican-api-log" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.844227 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecb6c22c-402b-4bdb-a791-29509985785a" containerName="dnsmasq-dns" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.844251 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bba6eef-372a-4b9a-a27c-76f834e3519b" containerName="neutron-api" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.844265 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="db2e7d63-181d-43f7-8cdd-7bfb8f95b6d9" containerName="barbican-api" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.844283 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="93e2d67f-9481-4ca3-83e7-4f7c0673b9d4" containerName="ceilometer-central-agent" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.844300 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="93e2d67f-9481-4ca3-83e7-4f7c0673b9d4" containerName="sg-core" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.844318 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bba6eef-372a-4b9a-a27c-76f834e3519b" containerName="neutron-httpd" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.846839 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.849500 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.851008 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.878162 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.909540 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bdfabc8f-27d1-4859-b309-800105752ea2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bdfabc8f-27d1-4859-b309-800105752ea2\") " pod="openstack/ceilometer-0" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.909598 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bdfabc8f-27d1-4859-b309-800105752ea2-log-httpd\") pod \"ceilometer-0\" (UID: \"bdfabc8f-27d1-4859-b309-800105752ea2\") " pod="openstack/ceilometer-0" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.909634 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdfabc8f-27d1-4859-b309-800105752ea2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bdfabc8f-27d1-4859-b309-800105752ea2\") " pod="openstack/ceilometer-0" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.909696 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bdfabc8f-27d1-4859-b309-800105752ea2-run-httpd\") pod \"ceilometer-0\" (UID: \"bdfabc8f-27d1-4859-b309-800105752ea2\") " pod="openstack/ceilometer-0" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.909771 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdfabc8f-27d1-4859-b309-800105752ea2-config-data\") pod \"ceilometer-0\" (UID: \"bdfabc8f-27d1-4859-b309-800105752ea2\") " pod="openstack/ceilometer-0" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.910029 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdfabc8f-27d1-4859-b309-800105752ea2-scripts\") pod \"ceilometer-0\" (UID: \"bdfabc8f-27d1-4859-b309-800105752ea2\") " pod="openstack/ceilometer-0" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.910072 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ndln\" (UniqueName: \"kubernetes.io/projected/bdfabc8f-27d1-4859-b309-800105752ea2-kube-api-access-5ndln\") pod \"ceilometer-0\" (UID: \"bdfabc8f-27d1-4859-b309-800105752ea2\") " pod="openstack/ceilometer-0" Dec 06 06:17:39 crc kubenswrapper[4809]: I1206 06:17:39.984317 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-6bb4ccfdc4-9jqw5" Dec 06 06:17:40 crc kubenswrapper[4809]: I1206 06:17:40.024925 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdfabc8f-27d1-4859-b309-800105752ea2-config-data\") pod \"ceilometer-0\" (UID: \"bdfabc8f-27d1-4859-b309-800105752ea2\") " pod="openstack/ceilometer-0" Dec 06 06:17:40 crc kubenswrapper[4809]: I1206 06:17:40.025037 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdfabc8f-27d1-4859-b309-800105752ea2-scripts\") pod \"ceilometer-0\" (UID: \"bdfabc8f-27d1-4859-b309-800105752ea2\") " pod="openstack/ceilometer-0" Dec 06 06:17:40 crc kubenswrapper[4809]: I1206 06:17:40.025075 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ndln\" (UniqueName: \"kubernetes.io/projected/bdfabc8f-27d1-4859-b309-800105752ea2-kube-api-access-5ndln\") pod \"ceilometer-0\" (UID: \"bdfabc8f-27d1-4859-b309-800105752ea2\") " pod="openstack/ceilometer-0" Dec 06 06:17:40 crc kubenswrapper[4809]: I1206 06:17:40.025200 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bdfabc8f-27d1-4859-b309-800105752ea2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bdfabc8f-27d1-4859-b309-800105752ea2\") " pod="openstack/ceilometer-0" Dec 06 06:17:40 crc kubenswrapper[4809]: I1206 06:17:40.025238 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bdfabc8f-27d1-4859-b309-800105752ea2-log-httpd\") pod \"ceilometer-0\" (UID: \"bdfabc8f-27d1-4859-b309-800105752ea2\") " pod="openstack/ceilometer-0" Dec 06 06:17:40 crc kubenswrapper[4809]: I1206 06:17:40.025273 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdfabc8f-27d1-4859-b309-800105752ea2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bdfabc8f-27d1-4859-b309-800105752ea2\") " pod="openstack/ceilometer-0" Dec 06 06:17:40 crc kubenswrapper[4809]: I1206 06:17:40.025329 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bdfabc8f-27d1-4859-b309-800105752ea2-run-httpd\") pod \"ceilometer-0\" (UID: \"bdfabc8f-27d1-4859-b309-800105752ea2\") " pod="openstack/ceilometer-0" Dec 06 06:17:40 crc kubenswrapper[4809]: I1206 06:17:40.025834 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bdfabc8f-27d1-4859-b309-800105752ea2-run-httpd\") pod \"ceilometer-0\" (UID: \"bdfabc8f-27d1-4859-b309-800105752ea2\") " pod="openstack/ceilometer-0" Dec 06 06:17:40 crc kubenswrapper[4809]: I1206 06:17:40.029042 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdfabc8f-27d1-4859-b309-800105752ea2-scripts\") pod \"ceilometer-0\" (UID: \"bdfabc8f-27d1-4859-b309-800105752ea2\") " pod="openstack/ceilometer-0" Dec 06 06:17:40 crc kubenswrapper[4809]: I1206 06:17:40.031723 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bdfabc8f-27d1-4859-b309-800105752ea2-log-httpd\") pod \"ceilometer-0\" (UID: \"bdfabc8f-27d1-4859-b309-800105752ea2\") " pod="openstack/ceilometer-0" Dec 06 06:17:40 crc kubenswrapper[4809]: I1206 06:17:40.038087 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdfabc8f-27d1-4859-b309-800105752ea2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bdfabc8f-27d1-4859-b309-800105752ea2\") " pod="openstack/ceilometer-0" Dec 06 06:17:40 crc kubenswrapper[4809]: I1206 06:17:40.044005 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bdfabc8f-27d1-4859-b309-800105752ea2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bdfabc8f-27d1-4859-b309-800105752ea2\") " pod="openstack/ceilometer-0" Dec 06 06:17:40 crc kubenswrapper[4809]: I1206 06:17:40.044571 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdfabc8f-27d1-4859-b309-800105752ea2-config-data\") pod \"ceilometer-0\" (UID: \"bdfabc8f-27d1-4859-b309-800105752ea2\") " pod="openstack/ceilometer-0" Dec 06 06:17:40 crc kubenswrapper[4809]: I1206 06:17:40.053745 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ndln\" (UniqueName: \"kubernetes.io/projected/bdfabc8f-27d1-4859-b309-800105752ea2-kube-api-access-5ndln\") pod \"ceilometer-0\" (UID: \"bdfabc8f-27d1-4859-b309-800105752ea2\") " pod="openstack/ceilometer-0" Dec 06 06:17:40 crc kubenswrapper[4809]: I1206 06:17:40.204469 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:17:40 crc kubenswrapper[4809]: I1206 06:17:40.711497 4809 generic.go:334] "Generic (PLEG): container finished" podID="7cf04523-ccb3-40bd-93d1-6dd59edf7530" containerID="6db290f4a2b99fba5d0365f55cdced03d05a875cf48e355dd49d6d4a42fbb0e4" exitCode=0 Dec 06 06:17:40 crc kubenswrapper[4809]: I1206 06:17:40.711780 4809 generic.go:334] "Generic (PLEG): container finished" podID="7cf04523-ccb3-40bd-93d1-6dd59edf7530" containerID="1a542df6256ee56b5fbdc3ca23fb895140362d7cb61722388669e994937cd78d" exitCode=143 Dec 06 06:17:40 crc kubenswrapper[4809]: I1206 06:17:40.711841 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7cf04523-ccb3-40bd-93d1-6dd59edf7530","Type":"ContainerDied","Data":"6db290f4a2b99fba5d0365f55cdced03d05a875cf48e355dd49d6d4a42fbb0e4"} Dec 06 06:17:40 crc kubenswrapper[4809]: I1206 06:17:40.711876 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7cf04523-ccb3-40bd-93d1-6dd59edf7530","Type":"ContainerDied","Data":"1a542df6256ee56b5fbdc3ca23fb895140362d7cb61722388669e994937cd78d"} Dec 06 06:17:40 crc kubenswrapper[4809]: I1206 06:17:40.711897 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7cf04523-ccb3-40bd-93d1-6dd59edf7530","Type":"ContainerDied","Data":"c3daedaa81e8e056c9c83f7b697a72c46501badb356d2f5a02a2ba11ae0a47fe"} Dec 06 06:17:40 crc kubenswrapper[4809]: I1206 06:17:40.711911 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c3daedaa81e8e056c9c83f7b697a72c46501badb356d2f5a02a2ba11ae0a47fe" Dec 06 06:17:40 crc kubenswrapper[4809]: I1206 06:17:40.720036 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"61b3cb73-1301-4073-9492-ef9f6810be3f","Type":"ContainerStarted","Data":"8b94911ee4fe5e9d341dbaf959bfe7aba30b4a12c818bd27741e70e6a6818015"} Dec 06 06:17:40 crc kubenswrapper[4809]: I1206 06:17:40.759362 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 06 06:17:40 crc kubenswrapper[4809]: I1206 06:17:40.844806 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cf04523-ccb3-40bd-93d1-6dd59edf7530-combined-ca-bundle\") pod \"7cf04523-ccb3-40bd-93d1-6dd59edf7530\" (UID: \"7cf04523-ccb3-40bd-93d1-6dd59edf7530\") " Dec 06 06:17:40 crc kubenswrapper[4809]: I1206 06:17:40.844879 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7cf04523-ccb3-40bd-93d1-6dd59edf7530-scripts\") pod \"7cf04523-ccb3-40bd-93d1-6dd59edf7530\" (UID: \"7cf04523-ccb3-40bd-93d1-6dd59edf7530\") " Dec 06 06:17:40 crc kubenswrapper[4809]: I1206 06:17:40.844958 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9zlt\" (UniqueName: \"kubernetes.io/projected/7cf04523-ccb3-40bd-93d1-6dd59edf7530-kube-api-access-h9zlt\") pod \"7cf04523-ccb3-40bd-93d1-6dd59edf7530\" (UID: \"7cf04523-ccb3-40bd-93d1-6dd59edf7530\") " Dec 06 06:17:40 crc kubenswrapper[4809]: I1206 06:17:40.844993 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7cf04523-ccb3-40bd-93d1-6dd59edf7530-etc-machine-id\") pod \"7cf04523-ccb3-40bd-93d1-6dd59edf7530\" (UID: \"7cf04523-ccb3-40bd-93d1-6dd59edf7530\") " Dec 06 06:17:40 crc kubenswrapper[4809]: I1206 06:17:40.845064 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cf04523-ccb3-40bd-93d1-6dd59edf7530-config-data\") pod \"7cf04523-ccb3-40bd-93d1-6dd59edf7530\" (UID: \"7cf04523-ccb3-40bd-93d1-6dd59edf7530\") " Dec 06 06:17:40 crc kubenswrapper[4809]: I1206 06:17:40.845101 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7cf04523-ccb3-40bd-93d1-6dd59edf7530-config-data-custom\") pod \"7cf04523-ccb3-40bd-93d1-6dd59edf7530\" (UID: \"7cf04523-ccb3-40bd-93d1-6dd59edf7530\") " Dec 06 06:17:40 crc kubenswrapper[4809]: I1206 06:17:40.845144 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7cf04523-ccb3-40bd-93d1-6dd59edf7530-logs\") pod \"7cf04523-ccb3-40bd-93d1-6dd59edf7530\" (UID: \"7cf04523-ccb3-40bd-93d1-6dd59edf7530\") " Dec 06 06:17:40 crc kubenswrapper[4809]: I1206 06:17:40.848104 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7cf04523-ccb3-40bd-93d1-6dd59edf7530-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "7cf04523-ccb3-40bd-93d1-6dd59edf7530" (UID: "7cf04523-ccb3-40bd-93d1-6dd59edf7530"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:17:40 crc kubenswrapper[4809]: I1206 06:17:40.848492 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7cf04523-ccb3-40bd-93d1-6dd59edf7530-logs" (OuterVolumeSpecName: "logs") pod "7cf04523-ccb3-40bd-93d1-6dd59edf7530" (UID: "7cf04523-ccb3-40bd-93d1-6dd59edf7530"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:17:40 crc kubenswrapper[4809]: I1206 06:17:40.854312 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7cf04523-ccb3-40bd-93d1-6dd59edf7530-scripts" (OuterVolumeSpecName: "scripts") pod "7cf04523-ccb3-40bd-93d1-6dd59edf7530" (UID: "7cf04523-ccb3-40bd-93d1-6dd59edf7530"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:17:40 crc kubenswrapper[4809]: I1206 06:17:40.854500 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7cf04523-ccb3-40bd-93d1-6dd59edf7530-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "7cf04523-ccb3-40bd-93d1-6dd59edf7530" (UID: "7cf04523-ccb3-40bd-93d1-6dd59edf7530"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:17:40 crc kubenswrapper[4809]: I1206 06:17:40.857875 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cf04523-ccb3-40bd-93d1-6dd59edf7530-kube-api-access-h9zlt" (OuterVolumeSpecName: "kube-api-access-h9zlt") pod "7cf04523-ccb3-40bd-93d1-6dd59edf7530" (UID: "7cf04523-ccb3-40bd-93d1-6dd59edf7530"). InnerVolumeSpecName "kube-api-access-h9zlt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:17:40 crc kubenswrapper[4809]: I1206 06:17:40.877154 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:17:40 crc kubenswrapper[4809]: W1206 06:17:40.880804 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbdfabc8f_27d1_4859_b309_800105752ea2.slice/crio-d9c979911b0a02af6e503d80c63e45602c873e87f33c6a5c998c6132d8436998 WatchSource:0}: Error finding container d9c979911b0a02af6e503d80c63e45602c873e87f33c6a5c998c6132d8436998: Status 404 returned error can't find the container with id d9c979911b0a02af6e503d80c63e45602c873e87f33c6a5c998c6132d8436998 Dec 06 06:17:40 crc kubenswrapper[4809]: I1206 06:17:40.905292 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7cf04523-ccb3-40bd-93d1-6dd59edf7530-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7cf04523-ccb3-40bd-93d1-6dd59edf7530" (UID: "7cf04523-ccb3-40bd-93d1-6dd59edf7530"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:17:40 crc kubenswrapper[4809]: I1206 06:17:40.930356 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7cf04523-ccb3-40bd-93d1-6dd59edf7530-config-data" (OuterVolumeSpecName: "config-data") pod "7cf04523-ccb3-40bd-93d1-6dd59edf7530" (UID: "7cf04523-ccb3-40bd-93d1-6dd59edf7530"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:17:40 crc kubenswrapper[4809]: I1206 06:17:40.948111 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cf04523-ccb3-40bd-93d1-6dd59edf7530-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:40 crc kubenswrapper[4809]: I1206 06:17:40.948159 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7cf04523-ccb3-40bd-93d1-6dd59edf7530-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:40 crc kubenswrapper[4809]: I1206 06:17:40.948172 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9zlt\" (UniqueName: \"kubernetes.io/projected/7cf04523-ccb3-40bd-93d1-6dd59edf7530-kube-api-access-h9zlt\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:40 crc kubenswrapper[4809]: I1206 06:17:40.948186 4809 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7cf04523-ccb3-40bd-93d1-6dd59edf7530-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:40 crc kubenswrapper[4809]: I1206 06:17:40.948197 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cf04523-ccb3-40bd-93d1-6dd59edf7530-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:40 crc kubenswrapper[4809]: I1206 06:17:40.948208 4809 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7cf04523-ccb3-40bd-93d1-6dd59edf7530-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:40 crc kubenswrapper[4809]: I1206 06:17:40.948219 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7cf04523-ccb3-40bd-93d1-6dd59edf7530-logs\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:41 crc kubenswrapper[4809]: I1206 06:17:41.405298 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93e2d67f-9481-4ca3-83e7-4f7c0673b9d4" path="/var/lib/kubelet/pods/93e2d67f-9481-4ca3-83e7-4f7c0673b9d4/volumes" Dec 06 06:17:41 crc kubenswrapper[4809]: I1206 06:17:41.735661 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"61b3cb73-1301-4073-9492-ef9f6810be3f","Type":"ContainerStarted","Data":"c4906c28036b8e2bd8d1e7626a169b8573e55ab8f1c5a01c0506facd86f74b6f"} Dec 06 06:17:41 crc kubenswrapper[4809]: I1206 06:17:41.741228 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 06 06:17:41 crc kubenswrapper[4809]: I1206 06:17:41.741228 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bdfabc8f-27d1-4859-b309-800105752ea2","Type":"ContainerStarted","Data":"d9c979911b0a02af6e503d80c63e45602c873e87f33c6a5c998c6132d8436998"} Dec 06 06:17:41 crc kubenswrapper[4809]: I1206 06:17:41.761179 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=10.093240564 podStartE2EDuration="11.761151937s" podCreationTimestamp="2025-12-06 06:17:30 +0000 UTC" firstStartedPulling="2025-12-06 06:17:37.167967676 +0000 UTC m=+1582.056950618" lastFinishedPulling="2025-12-06 06:17:38.835879049 +0000 UTC m=+1583.724861991" observedRunningTime="2025-12-06 06:17:41.754824077 +0000 UTC m=+1586.643807019" watchObservedRunningTime="2025-12-06 06:17:41.761151937 +0000 UTC m=+1586.650134889" Dec 06 06:17:41 crc kubenswrapper[4809]: I1206 06:17:41.785846 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 06 06:17:41 crc kubenswrapper[4809]: I1206 06:17:41.801645 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 06 06:17:41 crc kubenswrapper[4809]: I1206 06:17:41.823991 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 06 06:17:41 crc kubenswrapper[4809]: E1206 06:17:41.824628 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cf04523-ccb3-40bd-93d1-6dd59edf7530" containerName="cinder-api" Dec 06 06:17:41 crc kubenswrapper[4809]: I1206 06:17:41.824653 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cf04523-ccb3-40bd-93d1-6dd59edf7530" containerName="cinder-api" Dec 06 06:17:41 crc kubenswrapper[4809]: E1206 06:17:41.824722 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cf04523-ccb3-40bd-93d1-6dd59edf7530" containerName="cinder-api-log" Dec 06 06:17:41 crc kubenswrapper[4809]: I1206 06:17:41.824733 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cf04523-ccb3-40bd-93d1-6dd59edf7530" containerName="cinder-api-log" Dec 06 06:17:41 crc kubenswrapper[4809]: I1206 06:17:41.825064 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cf04523-ccb3-40bd-93d1-6dd59edf7530" containerName="cinder-api-log" Dec 06 06:17:41 crc kubenswrapper[4809]: I1206 06:17:41.825093 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cf04523-ccb3-40bd-93d1-6dd59edf7530" containerName="cinder-api" Dec 06 06:17:41 crc kubenswrapper[4809]: I1206 06:17:41.826669 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 06 06:17:41 crc kubenswrapper[4809]: I1206 06:17:41.830312 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 06 06:17:41 crc kubenswrapper[4809]: I1206 06:17:41.830579 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 06 06:17:41 crc kubenswrapper[4809]: I1206 06:17:41.832332 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 06 06:17:41 crc kubenswrapper[4809]: I1206 06:17:41.842205 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 06 06:17:41 crc kubenswrapper[4809]: I1206 06:17:41.867209 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a274a8e-452b-4087-92b1-c9bb713df5b5-scripts\") pod \"cinder-api-0\" (UID: \"8a274a8e-452b-4087-92b1-c9bb713df5b5\") " pod="openstack/cinder-api-0" Dec 06 06:17:41 crc kubenswrapper[4809]: I1206 06:17:41.867259 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a274a8e-452b-4087-92b1-c9bb713df5b5-public-tls-certs\") pod \"cinder-api-0\" (UID: \"8a274a8e-452b-4087-92b1-c9bb713df5b5\") " pod="openstack/cinder-api-0" Dec 06 06:17:41 crc kubenswrapper[4809]: I1206 06:17:41.867288 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a274a8e-452b-4087-92b1-c9bb713df5b5-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"8a274a8e-452b-4087-92b1-c9bb713df5b5\") " pod="openstack/cinder-api-0" Dec 06 06:17:41 crc kubenswrapper[4809]: I1206 06:17:41.867308 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8a274a8e-452b-4087-92b1-c9bb713df5b5-config-data-custom\") pod \"cinder-api-0\" (UID: \"8a274a8e-452b-4087-92b1-c9bb713df5b5\") " pod="openstack/cinder-api-0" Dec 06 06:17:41 crc kubenswrapper[4809]: I1206 06:17:41.867373 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8a274a8e-452b-4087-92b1-c9bb713df5b5-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8a274a8e-452b-4087-92b1-c9bb713df5b5\") " pod="openstack/cinder-api-0" Dec 06 06:17:41 crc kubenswrapper[4809]: I1206 06:17:41.867389 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a274a8e-452b-4087-92b1-c9bb713df5b5-logs\") pod \"cinder-api-0\" (UID: \"8a274a8e-452b-4087-92b1-c9bb713df5b5\") " pod="openstack/cinder-api-0" Dec 06 06:17:41 crc kubenswrapper[4809]: I1206 06:17:41.867442 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a274a8e-452b-4087-92b1-c9bb713df5b5-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8a274a8e-452b-4087-92b1-c9bb713df5b5\") " pod="openstack/cinder-api-0" Dec 06 06:17:41 crc kubenswrapper[4809]: I1206 06:17:41.867475 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxfvm\" (UniqueName: \"kubernetes.io/projected/8a274a8e-452b-4087-92b1-c9bb713df5b5-kube-api-access-dxfvm\") pod \"cinder-api-0\" (UID: \"8a274a8e-452b-4087-92b1-c9bb713df5b5\") " pod="openstack/cinder-api-0" Dec 06 06:17:41 crc kubenswrapper[4809]: I1206 06:17:41.867508 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a274a8e-452b-4087-92b1-c9bb713df5b5-config-data\") pod \"cinder-api-0\" (UID: \"8a274a8e-452b-4087-92b1-c9bb713df5b5\") " pod="openstack/cinder-api-0" Dec 06 06:17:41 crc kubenswrapper[4809]: I1206 06:17:41.969352 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxfvm\" (UniqueName: \"kubernetes.io/projected/8a274a8e-452b-4087-92b1-c9bb713df5b5-kube-api-access-dxfvm\") pod \"cinder-api-0\" (UID: \"8a274a8e-452b-4087-92b1-c9bb713df5b5\") " pod="openstack/cinder-api-0" Dec 06 06:17:41 crc kubenswrapper[4809]: I1206 06:17:41.969452 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a274a8e-452b-4087-92b1-c9bb713df5b5-config-data\") pod \"cinder-api-0\" (UID: \"8a274a8e-452b-4087-92b1-c9bb713df5b5\") " pod="openstack/cinder-api-0" Dec 06 06:17:41 crc kubenswrapper[4809]: I1206 06:17:41.969610 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a274a8e-452b-4087-92b1-c9bb713df5b5-scripts\") pod \"cinder-api-0\" (UID: \"8a274a8e-452b-4087-92b1-c9bb713df5b5\") " pod="openstack/cinder-api-0" Dec 06 06:17:41 crc kubenswrapper[4809]: I1206 06:17:41.969666 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a274a8e-452b-4087-92b1-c9bb713df5b5-public-tls-certs\") pod \"cinder-api-0\" (UID: \"8a274a8e-452b-4087-92b1-c9bb713df5b5\") " pod="openstack/cinder-api-0" Dec 06 06:17:41 crc kubenswrapper[4809]: I1206 06:17:41.969692 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a274a8e-452b-4087-92b1-c9bb713df5b5-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"8a274a8e-452b-4087-92b1-c9bb713df5b5\") " pod="openstack/cinder-api-0" Dec 06 06:17:41 crc kubenswrapper[4809]: I1206 06:17:41.969737 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8a274a8e-452b-4087-92b1-c9bb713df5b5-config-data-custom\") pod \"cinder-api-0\" (UID: \"8a274a8e-452b-4087-92b1-c9bb713df5b5\") " pod="openstack/cinder-api-0" Dec 06 06:17:41 crc kubenswrapper[4809]: I1206 06:17:41.969812 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8a274a8e-452b-4087-92b1-c9bb713df5b5-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8a274a8e-452b-4087-92b1-c9bb713df5b5\") " pod="openstack/cinder-api-0" Dec 06 06:17:41 crc kubenswrapper[4809]: I1206 06:17:41.969830 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a274a8e-452b-4087-92b1-c9bb713df5b5-logs\") pod \"cinder-api-0\" (UID: \"8a274a8e-452b-4087-92b1-c9bb713df5b5\") " pod="openstack/cinder-api-0" Dec 06 06:17:41 crc kubenswrapper[4809]: I1206 06:17:41.969887 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a274a8e-452b-4087-92b1-c9bb713df5b5-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8a274a8e-452b-4087-92b1-c9bb713df5b5\") " pod="openstack/cinder-api-0" Dec 06 06:17:41 crc kubenswrapper[4809]: I1206 06:17:41.971334 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8a274a8e-452b-4087-92b1-c9bb713df5b5-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8a274a8e-452b-4087-92b1-c9bb713df5b5\") " pod="openstack/cinder-api-0" Dec 06 06:17:41 crc kubenswrapper[4809]: I1206 06:17:41.972155 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a274a8e-452b-4087-92b1-c9bb713df5b5-logs\") pod \"cinder-api-0\" (UID: \"8a274a8e-452b-4087-92b1-c9bb713df5b5\") " pod="openstack/cinder-api-0" Dec 06 06:17:41 crc kubenswrapper[4809]: I1206 06:17:41.974435 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a274a8e-452b-4087-92b1-c9bb713df5b5-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"8a274a8e-452b-4087-92b1-c9bb713df5b5\") " pod="openstack/cinder-api-0" Dec 06 06:17:41 crc kubenswrapper[4809]: I1206 06:17:41.974943 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a274a8e-452b-4087-92b1-c9bb713df5b5-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8a274a8e-452b-4087-92b1-c9bb713df5b5\") " pod="openstack/cinder-api-0" Dec 06 06:17:41 crc kubenswrapper[4809]: I1206 06:17:41.976436 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a274a8e-452b-4087-92b1-c9bb713df5b5-config-data\") pod \"cinder-api-0\" (UID: \"8a274a8e-452b-4087-92b1-c9bb713df5b5\") " pod="openstack/cinder-api-0" Dec 06 06:17:41 crc kubenswrapper[4809]: I1206 06:17:41.977642 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8a274a8e-452b-4087-92b1-c9bb713df5b5-config-data-custom\") pod \"cinder-api-0\" (UID: \"8a274a8e-452b-4087-92b1-c9bb713df5b5\") " pod="openstack/cinder-api-0" Dec 06 06:17:41 crc kubenswrapper[4809]: I1206 06:17:41.978809 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a274a8e-452b-4087-92b1-c9bb713df5b5-scripts\") pod \"cinder-api-0\" (UID: \"8a274a8e-452b-4087-92b1-c9bb713df5b5\") " pod="openstack/cinder-api-0" Dec 06 06:17:41 crc kubenswrapper[4809]: I1206 06:17:41.983243 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a274a8e-452b-4087-92b1-c9bb713df5b5-public-tls-certs\") pod \"cinder-api-0\" (UID: \"8a274a8e-452b-4087-92b1-c9bb713df5b5\") " pod="openstack/cinder-api-0" Dec 06 06:17:41 crc kubenswrapper[4809]: I1206 06:17:41.988916 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxfvm\" (UniqueName: \"kubernetes.io/projected/8a274a8e-452b-4087-92b1-c9bb713df5b5-kube-api-access-dxfvm\") pod \"cinder-api-0\" (UID: \"8a274a8e-452b-4087-92b1-c9bb713df5b5\") " pod="openstack/cinder-api-0" Dec 06 06:17:42 crc kubenswrapper[4809]: I1206 06:17:42.150290 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 06 06:17:42 crc kubenswrapper[4809]: I1206 06:17:42.712767 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 06 06:17:42 crc kubenswrapper[4809]: I1206 06:17:42.787110 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8a274a8e-452b-4087-92b1-c9bb713df5b5","Type":"ContainerStarted","Data":"b537b2d731ff3f462def5f25592d43a223f0d42aacf7499f6bfd22450b275828"} Dec 06 06:17:42 crc kubenswrapper[4809]: I1206 06:17:42.803273 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bdfabc8f-27d1-4859-b309-800105752ea2","Type":"ContainerStarted","Data":"ac843dc132759e8b65936f4165e8a3c3486a36b5a3ca6303cc78964646d90d8f"} Dec 06 06:17:43 crc kubenswrapper[4809]: I1206 06:17:43.407094 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7cf04523-ccb3-40bd-93d1-6dd59edf7530" path="/var/lib/kubelet/pods/7cf04523-ccb3-40bd-93d1-6dd59edf7530/volumes" Dec 06 06:17:43 crc kubenswrapper[4809]: I1206 06:17:43.831673 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8a274a8e-452b-4087-92b1-c9bb713df5b5","Type":"ContainerStarted","Data":"3a4e12bbceb3443f4b3b175c6fe77423d96362b204f93f833a2c06df3b17d09d"} Dec 06 06:17:43 crc kubenswrapper[4809]: I1206 06:17:43.837112 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bdfabc8f-27d1-4859-b309-800105752ea2","Type":"ContainerStarted","Data":"e36372324edd2822f6662b8a4a84968d25802a012c83f32e0e9f87d761baa86a"} Dec 06 06:17:43 crc kubenswrapper[4809]: I1206 06:17:43.923527 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 06 06:17:43 crc kubenswrapper[4809]: I1206 06:17:43.927009 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 06 06:17:43 crc kubenswrapper[4809]: I1206 06:17:43.932444 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 06 06:17:43 crc kubenswrapper[4809]: I1206 06:17:43.932676 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 06 06:17:43 crc kubenswrapper[4809]: I1206 06:17:43.942550 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-n59bs" Dec 06 06:17:43 crc kubenswrapper[4809]: I1206 06:17:43.958279 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 06 06:17:44 crc kubenswrapper[4809]: I1206 06:17:44.025924 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9bd37674-546d-4297-9d5a-5c7b61bda3ad-openstack-config-secret\") pod \"openstackclient\" (UID: \"9bd37674-546d-4297-9d5a-5c7b61bda3ad\") " pod="openstack/openstackclient" Dec 06 06:17:44 crc kubenswrapper[4809]: I1206 06:17:44.025997 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bd37674-546d-4297-9d5a-5c7b61bda3ad-combined-ca-bundle\") pod \"openstackclient\" (UID: \"9bd37674-546d-4297-9d5a-5c7b61bda3ad\") " pod="openstack/openstackclient" Dec 06 06:17:44 crc kubenswrapper[4809]: I1206 06:17:44.026137 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdmxh\" (UniqueName: \"kubernetes.io/projected/9bd37674-546d-4297-9d5a-5c7b61bda3ad-kube-api-access-cdmxh\") pod \"openstackclient\" (UID: \"9bd37674-546d-4297-9d5a-5c7b61bda3ad\") " pod="openstack/openstackclient" Dec 06 06:17:44 crc kubenswrapper[4809]: I1206 06:17:44.026204 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9bd37674-546d-4297-9d5a-5c7b61bda3ad-openstack-config\") pod \"openstackclient\" (UID: \"9bd37674-546d-4297-9d5a-5c7b61bda3ad\") " pod="openstack/openstackclient" Dec 06 06:17:44 crc kubenswrapper[4809]: I1206 06:17:44.128195 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdmxh\" (UniqueName: \"kubernetes.io/projected/9bd37674-546d-4297-9d5a-5c7b61bda3ad-kube-api-access-cdmxh\") pod \"openstackclient\" (UID: \"9bd37674-546d-4297-9d5a-5c7b61bda3ad\") " pod="openstack/openstackclient" Dec 06 06:17:44 crc kubenswrapper[4809]: I1206 06:17:44.128298 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9bd37674-546d-4297-9d5a-5c7b61bda3ad-openstack-config\") pod \"openstackclient\" (UID: \"9bd37674-546d-4297-9d5a-5c7b61bda3ad\") " pod="openstack/openstackclient" Dec 06 06:17:44 crc kubenswrapper[4809]: I1206 06:17:44.128336 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9bd37674-546d-4297-9d5a-5c7b61bda3ad-openstack-config-secret\") pod \"openstackclient\" (UID: \"9bd37674-546d-4297-9d5a-5c7b61bda3ad\") " pod="openstack/openstackclient" Dec 06 06:17:44 crc kubenswrapper[4809]: I1206 06:17:44.128368 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bd37674-546d-4297-9d5a-5c7b61bda3ad-combined-ca-bundle\") pod \"openstackclient\" (UID: \"9bd37674-546d-4297-9d5a-5c7b61bda3ad\") " pod="openstack/openstackclient" Dec 06 06:17:44 crc kubenswrapper[4809]: I1206 06:17:44.131414 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9bd37674-546d-4297-9d5a-5c7b61bda3ad-openstack-config\") pod \"openstackclient\" (UID: \"9bd37674-546d-4297-9d5a-5c7b61bda3ad\") " pod="openstack/openstackclient" Dec 06 06:17:44 crc kubenswrapper[4809]: I1206 06:17:44.146419 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9bd37674-546d-4297-9d5a-5c7b61bda3ad-openstack-config-secret\") pod \"openstackclient\" (UID: \"9bd37674-546d-4297-9d5a-5c7b61bda3ad\") " pod="openstack/openstackclient" Dec 06 06:17:44 crc kubenswrapper[4809]: I1206 06:17:44.146488 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 06 06:17:44 crc kubenswrapper[4809]: E1206 06:17:44.148205 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle kube-api-access-cdmxh], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/openstackclient" podUID="9bd37674-546d-4297-9d5a-5c7b61bda3ad" Dec 06 06:17:44 crc kubenswrapper[4809]: E1206 06:17:44.148320 4809 projected.go:194] Error preparing data for projected volume kube-api-access-cdmxh for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: User "system:node:crc" cannot create resource "serviceaccounts/token" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Dec 06 06:17:44 crc kubenswrapper[4809]: E1206 06:17:44.149297 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9bd37674-546d-4297-9d5a-5c7b61bda3ad-kube-api-access-cdmxh podName:9bd37674-546d-4297-9d5a-5c7b61bda3ad nodeName:}" failed. No retries permitted until 2025-12-06 06:17:44.649272303 +0000 UTC m=+1589.538255265 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cdmxh" (UniqueName: "kubernetes.io/projected/9bd37674-546d-4297-9d5a-5c7b61bda3ad-kube-api-access-cdmxh") pod "openstackclient" (UID: "9bd37674-546d-4297-9d5a-5c7b61bda3ad") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: User "system:node:crc" cannot create resource "serviceaccounts/token" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Dec 06 06:17:44 crc kubenswrapper[4809]: I1206 06:17:44.167540 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 06 06:17:44 crc kubenswrapper[4809]: I1206 06:17:44.173970 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bd37674-546d-4297-9d5a-5c7b61bda3ad-combined-ca-bundle\") pod \"openstackclient\" (UID: \"9bd37674-546d-4297-9d5a-5c7b61bda3ad\") " pod="openstack/openstackclient" Dec 06 06:17:44 crc kubenswrapper[4809]: I1206 06:17:44.227841 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 06 06:17:44 crc kubenswrapper[4809]: I1206 06:17:44.260197 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 06 06:17:44 crc kubenswrapper[4809]: I1206 06:17:44.295876 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 06 06:17:44 crc kubenswrapper[4809]: I1206 06:17:44.384845 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c6578f3-bf7b-4892-abc9-5f4f145c8d36-combined-ca-bundle\") pod \"openstackclient\" (UID: \"6c6578f3-bf7b-4892-abc9-5f4f145c8d36\") " pod="openstack/openstackclient" Dec 06 06:17:44 crc kubenswrapper[4809]: I1206 06:17:44.385178 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxz42\" (UniqueName: \"kubernetes.io/projected/6c6578f3-bf7b-4892-abc9-5f4f145c8d36-kube-api-access-hxz42\") pod \"openstackclient\" (UID: \"6c6578f3-bf7b-4892-abc9-5f4f145c8d36\") " pod="openstack/openstackclient" Dec 06 06:17:44 crc kubenswrapper[4809]: I1206 06:17:44.385280 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6c6578f3-bf7b-4892-abc9-5f4f145c8d36-openstack-config-secret\") pod \"openstackclient\" (UID: \"6c6578f3-bf7b-4892-abc9-5f4f145c8d36\") " pod="openstack/openstackclient" Dec 06 06:17:44 crc kubenswrapper[4809]: I1206 06:17:44.385344 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6c6578f3-bf7b-4892-abc9-5f4f145c8d36-openstack-config\") pod \"openstackclient\" (UID: \"6c6578f3-bf7b-4892-abc9-5f4f145c8d36\") " pod="openstack/openstackclient" Dec 06 06:17:44 crc kubenswrapper[4809]: I1206 06:17:44.490184 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxz42\" (UniqueName: \"kubernetes.io/projected/6c6578f3-bf7b-4892-abc9-5f4f145c8d36-kube-api-access-hxz42\") pod \"openstackclient\" (UID: \"6c6578f3-bf7b-4892-abc9-5f4f145c8d36\") " pod="openstack/openstackclient" Dec 06 06:17:44 crc kubenswrapper[4809]: I1206 06:17:44.490471 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6c6578f3-bf7b-4892-abc9-5f4f145c8d36-openstack-config-secret\") pod \"openstackclient\" (UID: \"6c6578f3-bf7b-4892-abc9-5f4f145c8d36\") " pod="openstack/openstackclient" Dec 06 06:17:44 crc kubenswrapper[4809]: I1206 06:17:44.490530 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6c6578f3-bf7b-4892-abc9-5f4f145c8d36-openstack-config\") pod \"openstackclient\" (UID: \"6c6578f3-bf7b-4892-abc9-5f4f145c8d36\") " pod="openstack/openstackclient" Dec 06 06:17:44 crc kubenswrapper[4809]: I1206 06:17:44.490558 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c6578f3-bf7b-4892-abc9-5f4f145c8d36-combined-ca-bundle\") pod \"openstackclient\" (UID: \"6c6578f3-bf7b-4892-abc9-5f4f145c8d36\") " pod="openstack/openstackclient" Dec 06 06:17:44 crc kubenswrapper[4809]: I1206 06:17:44.492126 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6c6578f3-bf7b-4892-abc9-5f4f145c8d36-openstack-config\") pod \"openstackclient\" (UID: \"6c6578f3-bf7b-4892-abc9-5f4f145c8d36\") " pod="openstack/openstackclient" Dec 06 06:17:44 crc kubenswrapper[4809]: I1206 06:17:44.494770 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6c6578f3-bf7b-4892-abc9-5f4f145c8d36-openstack-config-secret\") pod \"openstackclient\" (UID: \"6c6578f3-bf7b-4892-abc9-5f4f145c8d36\") " pod="openstack/openstackclient" Dec 06 06:17:44 crc kubenswrapper[4809]: I1206 06:17:44.495148 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c6578f3-bf7b-4892-abc9-5f4f145c8d36-combined-ca-bundle\") pod \"openstackclient\" (UID: \"6c6578f3-bf7b-4892-abc9-5f4f145c8d36\") " pod="openstack/openstackclient" Dec 06 06:17:44 crc kubenswrapper[4809]: I1206 06:17:44.514528 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxz42\" (UniqueName: \"kubernetes.io/projected/6c6578f3-bf7b-4892-abc9-5f4f145c8d36-kube-api-access-hxz42\") pod \"openstackclient\" (UID: \"6c6578f3-bf7b-4892-abc9-5f4f145c8d36\") " pod="openstack/openstackclient" Dec 06 06:17:44 crc kubenswrapper[4809]: I1206 06:17:44.652243 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 06 06:17:44 crc kubenswrapper[4809]: I1206 06:17:44.695466 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdmxh\" (UniqueName: \"kubernetes.io/projected/9bd37674-546d-4297-9d5a-5c7b61bda3ad-kube-api-access-cdmxh\") pod \"openstackclient\" (UID: \"9bd37674-546d-4297-9d5a-5c7b61bda3ad\") " pod="openstack/openstackclient" Dec 06 06:17:44 crc kubenswrapper[4809]: E1206 06:17:44.697643 4809 projected.go:194] Error preparing data for projected volume kube-api-access-cdmxh for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (9bd37674-546d-4297-9d5a-5c7b61bda3ad) does not match the UID in record. The object might have been deleted and then recreated Dec 06 06:17:44 crc kubenswrapper[4809]: E1206 06:17:44.697699 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9bd37674-546d-4297-9d5a-5c7b61bda3ad-kube-api-access-cdmxh podName:9bd37674-546d-4297-9d5a-5c7b61bda3ad nodeName:}" failed. No retries permitted until 2025-12-06 06:17:45.69768396 +0000 UTC m=+1590.586666902 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cdmxh" (UniqueName: "kubernetes.io/projected/9bd37674-546d-4297-9d5a-5c7b61bda3ad-kube-api-access-cdmxh") pod "openstackclient" (UID: "9bd37674-546d-4297-9d5a-5c7b61bda3ad") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (9bd37674-546d-4297-9d5a-5c7b61bda3ad) does not match the UID in record. The object might have been deleted and then recreated Dec 06 06:17:44 crc kubenswrapper[4809]: I1206 06:17:44.852101 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8a274a8e-452b-4087-92b1-c9bb713df5b5","Type":"ContainerStarted","Data":"59033f810b973b1755d7eab6030bb510d6e012f44fed794768bfc3d51875c243"} Dec 06 06:17:44 crc kubenswrapper[4809]: I1206 06:17:44.852561 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 06 06:17:44 crc kubenswrapper[4809]: I1206 06:17:44.860628 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 06 06:17:44 crc kubenswrapper[4809]: I1206 06:17:44.861720 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bdfabc8f-27d1-4859-b309-800105752ea2","Type":"ContainerStarted","Data":"5c5b288b360303a12ddd1cdc9a809c39f23fa448c90a28fffa55c8fe6b34e0c9"} Dec 06 06:17:44 crc kubenswrapper[4809]: I1206 06:17:44.891118 4809 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="9bd37674-546d-4297-9d5a-5c7b61bda3ad" podUID="6c6578f3-bf7b-4892-abc9-5f4f145c8d36" Dec 06 06:17:44 crc kubenswrapper[4809]: I1206 06:17:44.893967 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 06 06:17:45 crc kubenswrapper[4809]: I1206 06:17:45.005249 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9bd37674-546d-4297-9d5a-5c7b61bda3ad-openstack-config-secret\") pod \"9bd37674-546d-4297-9d5a-5c7b61bda3ad\" (UID: \"9bd37674-546d-4297-9d5a-5c7b61bda3ad\") " Dec 06 06:17:45 crc kubenswrapper[4809]: I1206 06:17:45.006464 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9bd37674-546d-4297-9d5a-5c7b61bda3ad-openstack-config\") pod \"9bd37674-546d-4297-9d5a-5c7b61bda3ad\" (UID: \"9bd37674-546d-4297-9d5a-5c7b61bda3ad\") " Dec 06 06:17:45 crc kubenswrapper[4809]: I1206 06:17:45.006550 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bd37674-546d-4297-9d5a-5c7b61bda3ad-combined-ca-bundle\") pod \"9bd37674-546d-4297-9d5a-5c7b61bda3ad\" (UID: \"9bd37674-546d-4297-9d5a-5c7b61bda3ad\") " Dec 06 06:17:45 crc kubenswrapper[4809]: I1206 06:17:45.007162 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cdmxh\" (UniqueName: \"kubernetes.io/projected/9bd37674-546d-4297-9d5a-5c7b61bda3ad-kube-api-access-cdmxh\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:45 crc kubenswrapper[4809]: I1206 06:17:45.008072 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9bd37674-546d-4297-9d5a-5c7b61bda3ad-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "9bd37674-546d-4297-9d5a-5c7b61bda3ad" (UID: "9bd37674-546d-4297-9d5a-5c7b61bda3ad"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:17:45 crc kubenswrapper[4809]: I1206 06:17:45.015958 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9bd37674-546d-4297-9d5a-5c7b61bda3ad-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9bd37674-546d-4297-9d5a-5c7b61bda3ad" (UID: "9bd37674-546d-4297-9d5a-5c7b61bda3ad"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:17:45 crc kubenswrapper[4809]: I1206 06:17:45.015984 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9bd37674-546d-4297-9d5a-5c7b61bda3ad-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "9bd37674-546d-4297-9d5a-5c7b61bda3ad" (UID: "9bd37674-546d-4297-9d5a-5c7b61bda3ad"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:17:45 crc kubenswrapper[4809]: I1206 06:17:45.109562 4809 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9bd37674-546d-4297-9d5a-5c7b61bda3ad-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:45 crc kubenswrapper[4809]: I1206 06:17:45.109604 4809 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9bd37674-546d-4297-9d5a-5c7b61bda3ad-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:45 crc kubenswrapper[4809]: I1206 06:17:45.109618 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bd37674-546d-4297-9d5a-5c7b61bda3ad-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:45 crc kubenswrapper[4809]: I1206 06:17:45.163881 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.163838418 podStartE2EDuration="4.163838418s" podCreationTimestamp="2025-12-06 06:17:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:17:44.882422885 +0000 UTC m=+1589.771405837" watchObservedRunningTime="2025-12-06 06:17:45.163838418 +0000 UTC m=+1590.052821370" Dec 06 06:17:45 crc kubenswrapper[4809]: I1206 06:17:45.167354 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 06 06:17:45 crc kubenswrapper[4809]: I1206 06:17:45.402548 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9bd37674-546d-4297-9d5a-5c7b61bda3ad" path="/var/lib/kubelet/pods/9bd37674-546d-4297-9d5a-5c7b61bda3ad/volumes" Dec 06 06:17:45 crc kubenswrapper[4809]: I1206 06:17:45.881412 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 06 06:17:45 crc kubenswrapper[4809]: I1206 06:17:45.882458 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"6c6578f3-bf7b-4892-abc9-5f4f145c8d36","Type":"ContainerStarted","Data":"cf6a40f4f024d96973a9a687562638a4baca021976a4399bfd50fb177ef0920b"} Dec 06 06:17:45 crc kubenswrapper[4809]: I1206 06:17:45.892018 4809 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="9bd37674-546d-4297-9d5a-5c7b61bda3ad" podUID="6c6578f3-bf7b-4892-abc9-5f4f145c8d36" Dec 06 06:17:46 crc kubenswrapper[4809]: I1206 06:17:46.156376 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 06 06:17:46 crc kubenswrapper[4809]: I1206 06:17:46.334124 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c9776ccc5-6dldb" Dec 06 06:17:46 crc kubenswrapper[4809]: I1206 06:17:46.428372 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-k4rfq"] Dec 06 06:17:46 crc kubenswrapper[4809]: I1206 06:17:46.428608 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-85ff748b95-k4rfq" podUID="90f11553-7c39-4768-8669-85112803eb21" containerName="dnsmasq-dns" containerID="cri-o://6d46c7bf2cc29fb06ab2974a2f57728ed4bc42abd66266594d85cf362760e738" gracePeriod=10 Dec 06 06:17:46 crc kubenswrapper[4809]: I1206 06:17:46.531705 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 06 06:17:46 crc kubenswrapper[4809]: I1206 06:17:46.953066 4809 generic.go:334] "Generic (PLEG): container finished" podID="90f11553-7c39-4768-8669-85112803eb21" containerID="6d46c7bf2cc29fb06ab2974a2f57728ed4bc42abd66266594d85cf362760e738" exitCode=0 Dec 06 06:17:46 crc kubenswrapper[4809]: I1206 06:17:46.955375 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-k4rfq" event={"ID":"90f11553-7c39-4768-8669-85112803eb21","Type":"ContainerDied","Data":"6d46c7bf2cc29fb06ab2974a2f57728ed4bc42abd66266594d85cf362760e738"} Dec 06 06:17:47 crc kubenswrapper[4809]: I1206 06:17:47.021777 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 06:17:47 crc kubenswrapper[4809]: I1206 06:17:47.105361 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-k4rfq" Dec 06 06:17:47 crc kubenswrapper[4809]: I1206 06:17:47.264430 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/90f11553-7c39-4768-8669-85112803eb21-ovsdbserver-sb\") pod \"90f11553-7c39-4768-8669-85112803eb21\" (UID: \"90f11553-7c39-4768-8669-85112803eb21\") " Dec 06 06:17:47 crc kubenswrapper[4809]: I1206 06:17:47.264856 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tgdx2\" (UniqueName: \"kubernetes.io/projected/90f11553-7c39-4768-8669-85112803eb21-kube-api-access-tgdx2\") pod \"90f11553-7c39-4768-8669-85112803eb21\" (UID: \"90f11553-7c39-4768-8669-85112803eb21\") " Dec 06 06:17:47 crc kubenswrapper[4809]: I1206 06:17:47.265055 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/90f11553-7c39-4768-8669-85112803eb21-dns-svc\") pod \"90f11553-7c39-4768-8669-85112803eb21\" (UID: \"90f11553-7c39-4768-8669-85112803eb21\") " Dec 06 06:17:47 crc kubenswrapper[4809]: I1206 06:17:47.265209 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/90f11553-7c39-4768-8669-85112803eb21-ovsdbserver-nb\") pod \"90f11553-7c39-4768-8669-85112803eb21\" (UID: \"90f11553-7c39-4768-8669-85112803eb21\") " Dec 06 06:17:47 crc kubenswrapper[4809]: I1206 06:17:47.265241 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90f11553-7c39-4768-8669-85112803eb21-config\") pod \"90f11553-7c39-4768-8669-85112803eb21\" (UID: \"90f11553-7c39-4768-8669-85112803eb21\") " Dec 06 06:17:47 crc kubenswrapper[4809]: I1206 06:17:47.265262 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/90f11553-7c39-4768-8669-85112803eb21-dns-swift-storage-0\") pod \"90f11553-7c39-4768-8669-85112803eb21\" (UID: \"90f11553-7c39-4768-8669-85112803eb21\") " Dec 06 06:17:47 crc kubenswrapper[4809]: I1206 06:17:47.291977 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90f11553-7c39-4768-8669-85112803eb21-kube-api-access-tgdx2" (OuterVolumeSpecName: "kube-api-access-tgdx2") pod "90f11553-7c39-4768-8669-85112803eb21" (UID: "90f11553-7c39-4768-8669-85112803eb21"). InnerVolumeSpecName "kube-api-access-tgdx2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:17:47 crc kubenswrapper[4809]: I1206 06:17:47.351093 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90f11553-7c39-4768-8669-85112803eb21-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "90f11553-7c39-4768-8669-85112803eb21" (UID: "90f11553-7c39-4768-8669-85112803eb21"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:17:47 crc kubenswrapper[4809]: I1206 06:17:47.368508 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90f11553-7c39-4768-8669-85112803eb21-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "90f11553-7c39-4768-8669-85112803eb21" (UID: "90f11553-7c39-4768-8669-85112803eb21"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:17:47 crc kubenswrapper[4809]: I1206 06:17:47.371126 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/90f11553-7c39-4768-8669-85112803eb21-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:47 crc kubenswrapper[4809]: I1206 06:17:47.371161 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tgdx2\" (UniqueName: \"kubernetes.io/projected/90f11553-7c39-4768-8669-85112803eb21-kube-api-access-tgdx2\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:47 crc kubenswrapper[4809]: I1206 06:17:47.371180 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/90f11553-7c39-4768-8669-85112803eb21-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:47 crc kubenswrapper[4809]: I1206 06:17:47.392528 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90f11553-7c39-4768-8669-85112803eb21-config" (OuterVolumeSpecName: "config") pod "90f11553-7c39-4768-8669-85112803eb21" (UID: "90f11553-7c39-4768-8669-85112803eb21"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:17:47 crc kubenswrapper[4809]: I1206 06:17:47.424802 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90f11553-7c39-4768-8669-85112803eb21-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "90f11553-7c39-4768-8669-85112803eb21" (UID: "90f11553-7c39-4768-8669-85112803eb21"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:17:47 crc kubenswrapper[4809]: I1206 06:17:47.443105 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90f11553-7c39-4768-8669-85112803eb21-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "90f11553-7c39-4768-8669-85112803eb21" (UID: "90f11553-7c39-4768-8669-85112803eb21"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:17:47 crc kubenswrapper[4809]: I1206 06:17:47.474629 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/90f11553-7c39-4768-8669-85112803eb21-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:47 crc kubenswrapper[4809]: I1206 06:17:47.474989 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90f11553-7c39-4768-8669-85112803eb21-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:47 crc kubenswrapper[4809]: I1206 06:17:47.475063 4809 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/90f11553-7c39-4768-8669-85112803eb21-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:47 crc kubenswrapper[4809]: I1206 06:17:47.971946 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-k4rfq" event={"ID":"90f11553-7c39-4768-8669-85112803eb21","Type":"ContainerDied","Data":"3cc3ba849aa2b6e3870642cabdb432ad61b3f823251d281d314e51d0e445ad86"} Dec 06 06:17:47 crc kubenswrapper[4809]: I1206 06:17:47.971981 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-k4rfq" Dec 06 06:17:47 crc kubenswrapper[4809]: I1206 06:17:47.972005 4809 scope.go:117] "RemoveContainer" containerID="6d46c7bf2cc29fb06ab2974a2f57728ed4bc42abd66266594d85cf362760e738" Dec 06 06:17:47 crc kubenswrapper[4809]: I1206 06:17:47.975729 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bdfabc8f-27d1-4859-b309-800105752ea2","Type":"ContainerStarted","Data":"06cd048b9712e9ee61382dc3d9de5071b45e0e65d356c1b9ba02459297769163"} Dec 06 06:17:47 crc kubenswrapper[4809]: I1206 06:17:47.975882 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="61b3cb73-1301-4073-9492-ef9f6810be3f" containerName="cinder-scheduler" containerID="cri-o://8b94911ee4fe5e9d341dbaf959bfe7aba30b4a12c818bd27741e70e6a6818015" gracePeriod=30 Dec 06 06:17:47 crc kubenswrapper[4809]: I1206 06:17:47.975912 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="61b3cb73-1301-4073-9492-ef9f6810be3f" containerName="probe" containerID="cri-o://c4906c28036b8e2bd8d1e7626a169b8573e55ab8f1c5a01c0506facd86f74b6f" gracePeriod=30 Dec 06 06:17:48 crc kubenswrapper[4809]: I1206 06:17:48.012286 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-k4rfq"] Dec 06 06:17:48 crc kubenswrapper[4809]: I1206 06:17:48.034026 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-k4rfq"] Dec 06 06:17:48 crc kubenswrapper[4809]: I1206 06:17:48.046131 4809 scope.go:117] "RemoveContainer" containerID="7f39a7c141b936ffb87141024c5ebc8375a90519fdb2000f95a93b6f885d25a0" Dec 06 06:17:48 crc kubenswrapper[4809]: I1206 06:17:48.050504 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.664627705 podStartE2EDuration="9.050482113s" podCreationTimestamp="2025-12-06 06:17:39 +0000 UTC" firstStartedPulling="2025-12-06 06:17:40.886426217 +0000 UTC m=+1585.775409159" lastFinishedPulling="2025-12-06 06:17:46.272280625 +0000 UTC m=+1591.161263567" observedRunningTime="2025-12-06 06:17:48.014526314 +0000 UTC m=+1592.903509256" watchObservedRunningTime="2025-12-06 06:17:48.050482113 +0000 UTC m=+1592.939465055" Dec 06 06:17:48 crc kubenswrapper[4809]: I1206 06:17:48.339800 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5fc7d5fdc6-llxdt" Dec 06 06:17:48 crc kubenswrapper[4809]: I1206 06:17:48.349870 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5fc7d5fdc6-llxdt" Dec 06 06:17:48 crc kubenswrapper[4809]: I1206 06:17:48.998213 4809 generic.go:334] "Generic (PLEG): container finished" podID="61b3cb73-1301-4073-9492-ef9f6810be3f" containerID="c4906c28036b8e2bd8d1e7626a169b8573e55ab8f1c5a01c0506facd86f74b6f" exitCode=0 Dec 06 06:17:48 crc kubenswrapper[4809]: I1206 06:17:48.998484 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"61b3cb73-1301-4073-9492-ef9f6810be3f","Type":"ContainerDied","Data":"c4906c28036b8e2bd8d1e7626a169b8573e55ab8f1c5a01c0506facd86f74b6f"} Dec 06 06:17:49 crc kubenswrapper[4809]: I1206 06:17:49.001562 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 06:17:49 crc kubenswrapper[4809]: I1206 06:17:49.484645 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90f11553-7c39-4768-8669-85112803eb21" path="/var/lib/kubelet/pods/90f11553-7c39-4768-8669-85112803eb21/volumes" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.032243 4809 generic.go:334] "Generic (PLEG): container finished" podID="61b3cb73-1301-4073-9492-ef9f6810be3f" containerID="8b94911ee4fe5e9d341dbaf959bfe7aba30b4a12c818bd27741e70e6a6818015" exitCode=0 Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.032814 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"61b3cb73-1301-4073-9492-ef9f6810be3f","Type":"ContainerDied","Data":"8b94911ee4fe5e9d341dbaf959bfe7aba30b4a12c818bd27741e70e6a6818015"} Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.033080 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"61b3cb73-1301-4073-9492-ef9f6810be3f","Type":"ContainerDied","Data":"bab6899d52381c591757d6a96c7302eba4559ecefbf8783d11cb13c116c291f7"} Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.033135 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bab6899d52381c591757d6a96c7302eba4559ecefbf8783d11cb13c116c291f7" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.060540 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.184906 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/61b3cb73-1301-4073-9492-ef9f6810be3f-config-data-custom\") pod \"61b3cb73-1301-4073-9492-ef9f6810be3f\" (UID: \"61b3cb73-1301-4073-9492-ef9f6810be3f\") " Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.185278 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61b3cb73-1301-4073-9492-ef9f6810be3f-combined-ca-bundle\") pod \"61b3cb73-1301-4073-9492-ef9f6810be3f\" (UID: \"61b3cb73-1301-4073-9492-ef9f6810be3f\") " Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.185308 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61b3cb73-1301-4073-9492-ef9f6810be3f-config-data\") pod \"61b3cb73-1301-4073-9492-ef9f6810be3f\" (UID: \"61b3cb73-1301-4073-9492-ef9f6810be3f\") " Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.185334 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/61b3cb73-1301-4073-9492-ef9f6810be3f-scripts\") pod \"61b3cb73-1301-4073-9492-ef9f6810be3f\" (UID: \"61b3cb73-1301-4073-9492-ef9f6810be3f\") " Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.186182 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/61b3cb73-1301-4073-9492-ef9f6810be3f-etc-machine-id\") pod \"61b3cb73-1301-4073-9492-ef9f6810be3f\" (UID: \"61b3cb73-1301-4073-9492-ef9f6810be3f\") " Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.186284 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/61b3cb73-1301-4073-9492-ef9f6810be3f-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "61b3cb73-1301-4073-9492-ef9f6810be3f" (UID: "61b3cb73-1301-4073-9492-ef9f6810be3f"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.186492 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c6jm8\" (UniqueName: \"kubernetes.io/projected/61b3cb73-1301-4073-9492-ef9f6810be3f-kube-api-access-c6jm8\") pod \"61b3cb73-1301-4073-9492-ef9f6810be3f\" (UID: \"61b3cb73-1301-4073-9492-ef9f6810be3f\") " Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.187724 4809 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/61b3cb73-1301-4073-9492-ef9f6810be3f-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.191468 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61b3cb73-1301-4073-9492-ef9f6810be3f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "61b3cb73-1301-4073-9492-ef9f6810be3f" (UID: "61b3cb73-1301-4073-9492-ef9f6810be3f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.199226 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61b3cb73-1301-4073-9492-ef9f6810be3f-scripts" (OuterVolumeSpecName: "scripts") pod "61b3cb73-1301-4073-9492-ef9f6810be3f" (UID: "61b3cb73-1301-4073-9492-ef9f6810be3f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.199257 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61b3cb73-1301-4073-9492-ef9f6810be3f-kube-api-access-c6jm8" (OuterVolumeSpecName: "kube-api-access-c6jm8") pod "61b3cb73-1301-4073-9492-ef9f6810be3f" (UID: "61b3cb73-1301-4073-9492-ef9f6810be3f"). InnerVolumeSpecName "kube-api-access-c6jm8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.278736 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61b3cb73-1301-4073-9492-ef9f6810be3f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "61b3cb73-1301-4073-9492-ef9f6810be3f" (UID: "61b3cb73-1301-4073-9492-ef9f6810be3f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.290205 4809 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/61b3cb73-1301-4073-9492-ef9f6810be3f-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.290235 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61b3cb73-1301-4073-9492-ef9f6810be3f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.290245 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/61b3cb73-1301-4073-9492-ef9f6810be3f-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.290257 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c6jm8\" (UniqueName: \"kubernetes.io/projected/61b3cb73-1301-4073-9492-ef9f6810be3f-kube-api-access-c6jm8\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.377504 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61b3cb73-1301-4073-9492-ef9f6810be3f-config-data" (OuterVolumeSpecName: "config-data") pod "61b3cb73-1301-4073-9492-ef9f6810be3f" (UID: "61b3cb73-1301-4073-9492-ef9f6810be3f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.392680 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61b3cb73-1301-4073-9492-ef9f6810be3f-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.651060 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-58bb4f66d7-bdk65"] Dec 06 06:17:51 crc kubenswrapper[4809]: E1206 06:17:51.651678 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61b3cb73-1301-4073-9492-ef9f6810be3f" containerName="probe" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.651699 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="61b3cb73-1301-4073-9492-ef9f6810be3f" containerName="probe" Dec 06 06:17:51 crc kubenswrapper[4809]: E1206 06:17:51.651718 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61b3cb73-1301-4073-9492-ef9f6810be3f" containerName="cinder-scheduler" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.651727 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="61b3cb73-1301-4073-9492-ef9f6810be3f" containerName="cinder-scheduler" Dec 06 06:17:51 crc kubenswrapper[4809]: E1206 06:17:51.651747 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90f11553-7c39-4768-8669-85112803eb21" containerName="dnsmasq-dns" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.651755 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="90f11553-7c39-4768-8669-85112803eb21" containerName="dnsmasq-dns" Dec 06 06:17:51 crc kubenswrapper[4809]: E1206 06:17:51.651787 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90f11553-7c39-4768-8669-85112803eb21" containerName="init" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.651794 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="90f11553-7c39-4768-8669-85112803eb21" containerName="init" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.652079 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="90f11553-7c39-4768-8669-85112803eb21" containerName="dnsmasq-dns" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.652106 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="61b3cb73-1301-4073-9492-ef9f6810be3f" containerName="cinder-scheduler" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.652142 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="61b3cb73-1301-4073-9492-ef9f6810be3f" containerName="probe" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.653576 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-58bb4f66d7-bdk65" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.656188 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.656499 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.656732 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.678504 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-58bb4f66d7-bdk65"] Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.803493 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d496af20-e3cf-4ce6-86f5-df841a9ba42c-combined-ca-bundle\") pod \"swift-proxy-58bb4f66d7-bdk65\" (UID: \"d496af20-e3cf-4ce6-86f5-df841a9ba42c\") " pod="openstack/swift-proxy-58bb4f66d7-bdk65" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.803584 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d496af20-e3cf-4ce6-86f5-df841a9ba42c-log-httpd\") pod \"swift-proxy-58bb4f66d7-bdk65\" (UID: \"d496af20-e3cf-4ce6-86f5-df841a9ba42c\") " pod="openstack/swift-proxy-58bb4f66d7-bdk65" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.803611 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d496af20-e3cf-4ce6-86f5-df841a9ba42c-internal-tls-certs\") pod \"swift-proxy-58bb4f66d7-bdk65\" (UID: \"d496af20-e3cf-4ce6-86f5-df841a9ba42c\") " pod="openstack/swift-proxy-58bb4f66d7-bdk65" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.803653 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d496af20-e3cf-4ce6-86f5-df841a9ba42c-run-httpd\") pod \"swift-proxy-58bb4f66d7-bdk65\" (UID: \"d496af20-e3cf-4ce6-86f5-df841a9ba42c\") " pod="openstack/swift-proxy-58bb4f66d7-bdk65" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.803924 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d496af20-e3cf-4ce6-86f5-df841a9ba42c-etc-swift\") pod \"swift-proxy-58bb4f66d7-bdk65\" (UID: \"d496af20-e3cf-4ce6-86f5-df841a9ba42c\") " pod="openstack/swift-proxy-58bb4f66d7-bdk65" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.804017 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjpx6\" (UniqueName: \"kubernetes.io/projected/d496af20-e3cf-4ce6-86f5-df841a9ba42c-kube-api-access-gjpx6\") pod \"swift-proxy-58bb4f66d7-bdk65\" (UID: \"d496af20-e3cf-4ce6-86f5-df841a9ba42c\") " pod="openstack/swift-proxy-58bb4f66d7-bdk65" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.804186 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d496af20-e3cf-4ce6-86f5-df841a9ba42c-public-tls-certs\") pod \"swift-proxy-58bb4f66d7-bdk65\" (UID: \"d496af20-e3cf-4ce6-86f5-df841a9ba42c\") " pod="openstack/swift-proxy-58bb4f66d7-bdk65" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.804225 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d496af20-e3cf-4ce6-86f5-df841a9ba42c-config-data\") pod \"swift-proxy-58bb4f66d7-bdk65\" (UID: \"d496af20-e3cf-4ce6-86f5-df841a9ba42c\") " pod="openstack/swift-proxy-58bb4f66d7-bdk65" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.906219 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d496af20-e3cf-4ce6-86f5-df841a9ba42c-etc-swift\") pod \"swift-proxy-58bb4f66d7-bdk65\" (UID: \"d496af20-e3cf-4ce6-86f5-df841a9ba42c\") " pod="openstack/swift-proxy-58bb4f66d7-bdk65" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.906306 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjpx6\" (UniqueName: \"kubernetes.io/projected/d496af20-e3cf-4ce6-86f5-df841a9ba42c-kube-api-access-gjpx6\") pod \"swift-proxy-58bb4f66d7-bdk65\" (UID: \"d496af20-e3cf-4ce6-86f5-df841a9ba42c\") " pod="openstack/swift-proxy-58bb4f66d7-bdk65" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.906436 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d496af20-e3cf-4ce6-86f5-df841a9ba42c-public-tls-certs\") pod \"swift-proxy-58bb4f66d7-bdk65\" (UID: \"d496af20-e3cf-4ce6-86f5-df841a9ba42c\") " pod="openstack/swift-proxy-58bb4f66d7-bdk65" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.906467 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d496af20-e3cf-4ce6-86f5-df841a9ba42c-config-data\") pod \"swift-proxy-58bb4f66d7-bdk65\" (UID: \"d496af20-e3cf-4ce6-86f5-df841a9ba42c\") " pod="openstack/swift-proxy-58bb4f66d7-bdk65" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.907133 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d496af20-e3cf-4ce6-86f5-df841a9ba42c-combined-ca-bundle\") pod \"swift-proxy-58bb4f66d7-bdk65\" (UID: \"d496af20-e3cf-4ce6-86f5-df841a9ba42c\") " pod="openstack/swift-proxy-58bb4f66d7-bdk65" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.907600 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d496af20-e3cf-4ce6-86f5-df841a9ba42c-log-httpd\") pod \"swift-proxy-58bb4f66d7-bdk65\" (UID: \"d496af20-e3cf-4ce6-86f5-df841a9ba42c\") " pod="openstack/swift-proxy-58bb4f66d7-bdk65" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.907649 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d496af20-e3cf-4ce6-86f5-df841a9ba42c-internal-tls-certs\") pod \"swift-proxy-58bb4f66d7-bdk65\" (UID: \"d496af20-e3cf-4ce6-86f5-df841a9ba42c\") " pod="openstack/swift-proxy-58bb4f66d7-bdk65" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.907767 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d496af20-e3cf-4ce6-86f5-df841a9ba42c-run-httpd\") pod \"swift-proxy-58bb4f66d7-bdk65\" (UID: \"d496af20-e3cf-4ce6-86f5-df841a9ba42c\") " pod="openstack/swift-proxy-58bb4f66d7-bdk65" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.908394 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d496af20-e3cf-4ce6-86f5-df841a9ba42c-log-httpd\") pod \"swift-proxy-58bb4f66d7-bdk65\" (UID: \"d496af20-e3cf-4ce6-86f5-df841a9ba42c\") " pod="openstack/swift-proxy-58bb4f66d7-bdk65" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.908466 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d496af20-e3cf-4ce6-86f5-df841a9ba42c-run-httpd\") pod \"swift-proxy-58bb4f66d7-bdk65\" (UID: \"d496af20-e3cf-4ce6-86f5-df841a9ba42c\") " pod="openstack/swift-proxy-58bb4f66d7-bdk65" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.911759 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d496af20-e3cf-4ce6-86f5-df841a9ba42c-internal-tls-certs\") pod \"swift-proxy-58bb4f66d7-bdk65\" (UID: \"d496af20-e3cf-4ce6-86f5-df841a9ba42c\") " pod="openstack/swift-proxy-58bb4f66d7-bdk65" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.911985 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d496af20-e3cf-4ce6-86f5-df841a9ba42c-combined-ca-bundle\") pod \"swift-proxy-58bb4f66d7-bdk65\" (UID: \"d496af20-e3cf-4ce6-86f5-df841a9ba42c\") " pod="openstack/swift-proxy-58bb4f66d7-bdk65" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.912677 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d496af20-e3cf-4ce6-86f5-df841a9ba42c-public-tls-certs\") pod \"swift-proxy-58bb4f66d7-bdk65\" (UID: \"d496af20-e3cf-4ce6-86f5-df841a9ba42c\") " pod="openstack/swift-proxy-58bb4f66d7-bdk65" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.913099 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d496af20-e3cf-4ce6-86f5-df841a9ba42c-etc-swift\") pod \"swift-proxy-58bb4f66d7-bdk65\" (UID: \"d496af20-e3cf-4ce6-86f5-df841a9ba42c\") " pod="openstack/swift-proxy-58bb4f66d7-bdk65" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.917050 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d496af20-e3cf-4ce6-86f5-df841a9ba42c-config-data\") pod \"swift-proxy-58bb4f66d7-bdk65\" (UID: \"d496af20-e3cf-4ce6-86f5-df841a9ba42c\") " pod="openstack/swift-proxy-58bb4f66d7-bdk65" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.928702 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjpx6\" (UniqueName: \"kubernetes.io/projected/d496af20-e3cf-4ce6-86f5-df841a9ba42c-kube-api-access-gjpx6\") pod \"swift-proxy-58bb4f66d7-bdk65\" (UID: \"d496af20-e3cf-4ce6-86f5-df841a9ba42c\") " pod="openstack/swift-proxy-58bb4f66d7-bdk65" Dec 06 06:17:51 crc kubenswrapper[4809]: I1206 06:17:51.977881 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-58bb4f66d7-bdk65" Dec 06 06:17:52 crc kubenswrapper[4809]: I1206 06:17:52.044029 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 06 06:17:52 crc kubenswrapper[4809]: I1206 06:17:52.162955 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 06:17:52 crc kubenswrapper[4809]: I1206 06:17:52.204850 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 06:17:52 crc kubenswrapper[4809]: I1206 06:17:52.218875 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 06:17:52 crc kubenswrapper[4809]: I1206 06:17:52.221065 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 06 06:17:52 crc kubenswrapper[4809]: I1206 06:17:52.226871 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 06 06:17:52 crc kubenswrapper[4809]: I1206 06:17:52.239901 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 06:17:52 crc kubenswrapper[4809]: I1206 06:17:52.320535 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5twj6\" (UniqueName: \"kubernetes.io/projected/47aec31f-059f-4f3c-baef-4f9406e8cd4c-kube-api-access-5twj6\") pod \"cinder-scheduler-0\" (UID: \"47aec31f-059f-4f3c-baef-4f9406e8cd4c\") " pod="openstack/cinder-scheduler-0" Dec 06 06:17:52 crc kubenswrapper[4809]: I1206 06:17:52.320609 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/47aec31f-059f-4f3c-baef-4f9406e8cd4c-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"47aec31f-059f-4f3c-baef-4f9406e8cd4c\") " pod="openstack/cinder-scheduler-0" Dec 06 06:17:52 crc kubenswrapper[4809]: I1206 06:17:52.320755 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47aec31f-059f-4f3c-baef-4f9406e8cd4c-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"47aec31f-059f-4f3c-baef-4f9406e8cd4c\") " pod="openstack/cinder-scheduler-0" Dec 06 06:17:52 crc kubenswrapper[4809]: I1206 06:17:52.320802 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47aec31f-059f-4f3c-baef-4f9406e8cd4c-scripts\") pod \"cinder-scheduler-0\" (UID: \"47aec31f-059f-4f3c-baef-4f9406e8cd4c\") " pod="openstack/cinder-scheduler-0" Dec 06 06:17:52 crc kubenswrapper[4809]: I1206 06:17:52.320824 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/47aec31f-059f-4f3c-baef-4f9406e8cd4c-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"47aec31f-059f-4f3c-baef-4f9406e8cd4c\") " pod="openstack/cinder-scheduler-0" Dec 06 06:17:52 crc kubenswrapper[4809]: I1206 06:17:52.320905 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47aec31f-059f-4f3c-baef-4f9406e8cd4c-config-data\") pod \"cinder-scheduler-0\" (UID: \"47aec31f-059f-4f3c-baef-4f9406e8cd4c\") " pod="openstack/cinder-scheduler-0" Dec 06 06:17:52 crc kubenswrapper[4809]: I1206 06:17:52.423869 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47aec31f-059f-4f3c-baef-4f9406e8cd4c-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"47aec31f-059f-4f3c-baef-4f9406e8cd4c\") " pod="openstack/cinder-scheduler-0" Dec 06 06:17:52 crc kubenswrapper[4809]: I1206 06:17:52.426898 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47aec31f-059f-4f3c-baef-4f9406e8cd4c-scripts\") pod \"cinder-scheduler-0\" (UID: \"47aec31f-059f-4f3c-baef-4f9406e8cd4c\") " pod="openstack/cinder-scheduler-0" Dec 06 06:17:52 crc kubenswrapper[4809]: I1206 06:17:52.427042 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/47aec31f-059f-4f3c-baef-4f9406e8cd4c-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"47aec31f-059f-4f3c-baef-4f9406e8cd4c\") " pod="openstack/cinder-scheduler-0" Dec 06 06:17:52 crc kubenswrapper[4809]: I1206 06:17:52.427295 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47aec31f-059f-4f3c-baef-4f9406e8cd4c-config-data\") pod \"cinder-scheduler-0\" (UID: \"47aec31f-059f-4f3c-baef-4f9406e8cd4c\") " pod="openstack/cinder-scheduler-0" Dec 06 06:17:52 crc kubenswrapper[4809]: I1206 06:17:52.427439 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5twj6\" (UniqueName: \"kubernetes.io/projected/47aec31f-059f-4f3c-baef-4f9406e8cd4c-kube-api-access-5twj6\") pod \"cinder-scheduler-0\" (UID: \"47aec31f-059f-4f3c-baef-4f9406e8cd4c\") " pod="openstack/cinder-scheduler-0" Dec 06 06:17:52 crc kubenswrapper[4809]: I1206 06:17:52.427547 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/47aec31f-059f-4f3c-baef-4f9406e8cd4c-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"47aec31f-059f-4f3c-baef-4f9406e8cd4c\") " pod="openstack/cinder-scheduler-0" Dec 06 06:17:52 crc kubenswrapper[4809]: I1206 06:17:52.428025 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/47aec31f-059f-4f3c-baef-4f9406e8cd4c-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"47aec31f-059f-4f3c-baef-4f9406e8cd4c\") " pod="openstack/cinder-scheduler-0" Dec 06 06:17:52 crc kubenswrapper[4809]: I1206 06:17:52.441854 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47aec31f-059f-4f3c-baef-4f9406e8cd4c-scripts\") pod \"cinder-scheduler-0\" (UID: \"47aec31f-059f-4f3c-baef-4f9406e8cd4c\") " pod="openstack/cinder-scheduler-0" Dec 06 06:17:52 crc kubenswrapper[4809]: I1206 06:17:52.442852 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47aec31f-059f-4f3c-baef-4f9406e8cd4c-config-data\") pod \"cinder-scheduler-0\" (UID: \"47aec31f-059f-4f3c-baef-4f9406e8cd4c\") " pod="openstack/cinder-scheduler-0" Dec 06 06:17:52 crc kubenswrapper[4809]: I1206 06:17:52.443730 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47aec31f-059f-4f3c-baef-4f9406e8cd4c-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"47aec31f-059f-4f3c-baef-4f9406e8cd4c\") " pod="openstack/cinder-scheduler-0" Dec 06 06:17:52 crc kubenswrapper[4809]: I1206 06:17:52.450829 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/47aec31f-059f-4f3c-baef-4f9406e8cd4c-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"47aec31f-059f-4f3c-baef-4f9406e8cd4c\") " pod="openstack/cinder-scheduler-0" Dec 06 06:17:52 crc kubenswrapper[4809]: I1206 06:17:52.473955 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5twj6\" (UniqueName: \"kubernetes.io/projected/47aec31f-059f-4f3c-baef-4f9406e8cd4c-kube-api-access-5twj6\") pod \"cinder-scheduler-0\" (UID: \"47aec31f-059f-4f3c-baef-4f9406e8cd4c\") " pod="openstack/cinder-scheduler-0" Dec 06 06:17:52 crc kubenswrapper[4809]: I1206 06:17:52.569853 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 06 06:17:52 crc kubenswrapper[4809]: I1206 06:17:52.779518 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-58bb4f66d7-bdk65"] Dec 06 06:17:53 crc kubenswrapper[4809]: I1206 06:17:53.067634 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-58bb4f66d7-bdk65" event={"ID":"d496af20-e3cf-4ce6-86f5-df841a9ba42c","Type":"ContainerStarted","Data":"f35d3f0cd7307f70bb828110fbe4378cdfa13237d9bcaf05edc4556a4f97d736"} Dec 06 06:17:53 crc kubenswrapper[4809]: I1206 06:17:53.134040 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 06:17:53 crc kubenswrapper[4809]: I1206 06:17:53.409265 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61b3cb73-1301-4073-9492-ef9f6810be3f" path="/var/lib/kubelet/pods/61b3cb73-1301-4073-9492-ef9f6810be3f/volumes" Dec 06 06:17:54 crc kubenswrapper[4809]: I1206 06:17:54.083447 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"47aec31f-059f-4f3c-baef-4f9406e8cd4c","Type":"ContainerStarted","Data":"7951799bf8adbd07048cc9bf3327f3c1d3356f35e33f8b3a3a088903a642497b"} Dec 06 06:17:54 crc kubenswrapper[4809]: I1206 06:17:54.085526 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-58bb4f66d7-bdk65" event={"ID":"d496af20-e3cf-4ce6-86f5-df841a9ba42c","Type":"ContainerStarted","Data":"b5f685e9337843390eedf1b81336c03082254debeee5e851bb1839c872a52d71"} Dec 06 06:17:55 crc kubenswrapper[4809]: I1206 06:17:55.103832 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-58bb4f66d7-bdk65" event={"ID":"d496af20-e3cf-4ce6-86f5-df841a9ba42c","Type":"ContainerStarted","Data":"657dc56f7faeb650483198f101e96b8f02a625d6dcabd48cabc7381f79e9a0e8"} Dec 06 06:17:55 crc kubenswrapper[4809]: I1206 06:17:55.104724 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-58bb4f66d7-bdk65" Dec 06 06:17:55 crc kubenswrapper[4809]: I1206 06:17:55.104751 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-58bb4f66d7-bdk65" Dec 06 06:17:55 crc kubenswrapper[4809]: I1206 06:17:55.115227 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"47aec31f-059f-4f3c-baef-4f9406e8cd4c","Type":"ContainerStarted","Data":"fefd478b5d0c758d8e4f6cc5ea01120bb084cf57e583dab89c1587610263b6bf"} Dec 06 06:17:55 crc kubenswrapper[4809]: I1206 06:17:55.139854 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-58bb4f66d7-bdk65" podStartSLOduration=4.139834646 podStartE2EDuration="4.139834646s" podCreationTimestamp="2025-12-06 06:17:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:17:55.127367 +0000 UTC m=+1600.016349942" watchObservedRunningTime="2025-12-06 06:17:55.139834646 +0000 UTC m=+1600.028817588" Dec 06 06:17:55 crc kubenswrapper[4809]: I1206 06:17:55.153120 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.153107504 podStartE2EDuration="3.153107504s" podCreationTimestamp="2025-12-06 06:17:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:17:55.151166232 +0000 UTC m=+1600.040149174" watchObservedRunningTime="2025-12-06 06:17:55.153107504 +0000 UTC m=+1600.042090446" Dec 06 06:17:55 crc kubenswrapper[4809]: I1206 06:17:55.422625 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:17:55 crc kubenswrapper[4809]: I1206 06:17:55.422869 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bdfabc8f-27d1-4859-b309-800105752ea2" containerName="ceilometer-central-agent" containerID="cri-o://ac843dc132759e8b65936f4165e8a3c3486a36b5a3ca6303cc78964646d90d8f" gracePeriod=30 Dec 06 06:17:55 crc kubenswrapper[4809]: I1206 06:17:55.423366 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bdfabc8f-27d1-4859-b309-800105752ea2" containerName="sg-core" containerID="cri-o://5c5b288b360303a12ddd1cdc9a809c39f23fa448c90a28fffa55c8fe6b34e0c9" gracePeriod=30 Dec 06 06:17:55 crc kubenswrapper[4809]: I1206 06:17:55.423489 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bdfabc8f-27d1-4859-b309-800105752ea2" containerName="proxy-httpd" containerID="cri-o://06cd048b9712e9ee61382dc3d9de5071b45e0e65d356c1b9ba02459297769163" gracePeriod=30 Dec 06 06:17:55 crc kubenswrapper[4809]: I1206 06:17:55.423535 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bdfabc8f-27d1-4859-b309-800105752ea2" containerName="ceilometer-notification-agent" containerID="cri-o://e36372324edd2822f6662b8a4a84968d25802a012c83f32e0e9f87d761baa86a" gracePeriod=30 Dec 06 06:17:55 crc kubenswrapper[4809]: I1206 06:17:55.793587 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 06 06:17:56 crc kubenswrapper[4809]: I1206 06:17:56.164264 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"47aec31f-059f-4f3c-baef-4f9406e8cd4c","Type":"ContainerStarted","Data":"39ae6c2a3d32d59536e8755e78cbbbc5eec76b6e9424314138dd1c1905e788a3"} Dec 06 06:17:56 crc kubenswrapper[4809]: I1206 06:17:56.203616 4809 generic.go:334] "Generic (PLEG): container finished" podID="bdfabc8f-27d1-4859-b309-800105752ea2" containerID="06cd048b9712e9ee61382dc3d9de5071b45e0e65d356c1b9ba02459297769163" exitCode=0 Dec 06 06:17:56 crc kubenswrapper[4809]: I1206 06:17:56.203659 4809 generic.go:334] "Generic (PLEG): container finished" podID="bdfabc8f-27d1-4859-b309-800105752ea2" containerID="5c5b288b360303a12ddd1cdc9a809c39f23fa448c90a28fffa55c8fe6b34e0c9" exitCode=2 Dec 06 06:17:56 crc kubenswrapper[4809]: I1206 06:17:56.203666 4809 generic.go:334] "Generic (PLEG): container finished" podID="bdfabc8f-27d1-4859-b309-800105752ea2" containerID="ac843dc132759e8b65936f4165e8a3c3486a36b5a3ca6303cc78964646d90d8f" exitCode=0 Dec 06 06:17:56 crc kubenswrapper[4809]: I1206 06:17:56.203984 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bdfabc8f-27d1-4859-b309-800105752ea2","Type":"ContainerDied","Data":"06cd048b9712e9ee61382dc3d9de5071b45e0e65d356c1b9ba02459297769163"} Dec 06 06:17:56 crc kubenswrapper[4809]: I1206 06:17:56.204016 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bdfabc8f-27d1-4859-b309-800105752ea2","Type":"ContainerDied","Data":"5c5b288b360303a12ddd1cdc9a809c39f23fa448c90a28fffa55c8fe6b34e0c9"} Dec 06 06:17:56 crc kubenswrapper[4809]: I1206 06:17:56.204030 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bdfabc8f-27d1-4859-b309-800105752ea2","Type":"ContainerDied","Data":"ac843dc132759e8b65936f4165e8a3c3486a36b5a3ca6303cc78964646d90d8f"} Dec 06 06:17:57 crc kubenswrapper[4809]: I1206 06:17:57.240151 4809 generic.go:334] "Generic (PLEG): container finished" podID="bdfabc8f-27d1-4859-b309-800105752ea2" containerID="e36372324edd2822f6662b8a4a84968d25802a012c83f32e0e9f87d761baa86a" exitCode=0 Dec 06 06:17:57 crc kubenswrapper[4809]: I1206 06:17:57.241657 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bdfabc8f-27d1-4859-b309-800105752ea2","Type":"ContainerDied","Data":"e36372324edd2822f6662b8a4a84968d25802a012c83f32e0e9f87d761baa86a"} Dec 06 06:17:57 crc kubenswrapper[4809]: I1206 06:17:57.570513 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 06 06:18:01 crc kubenswrapper[4809]: I1206 06:18:01.639925 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-7c94c94748-57b42"] Dec 06 06:18:01 crc kubenswrapper[4809]: I1206 06:18:01.643176 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7c94c94748-57b42" Dec 06 06:18:01 crc kubenswrapper[4809]: I1206 06:18:01.652157 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-km9wl" Dec 06 06:18:01 crc kubenswrapper[4809]: I1206 06:18:01.652519 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Dec 06 06:18:01 crc kubenswrapper[4809]: I1206 06:18:01.653538 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Dec 06 06:18:01 crc kubenswrapper[4809]: I1206 06:18:01.676493 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-7c94c94748-57b42"] Dec 06 06:18:01 crc kubenswrapper[4809]: I1206 06:18:01.697648 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brq7t\" (UniqueName: \"kubernetes.io/projected/515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a-kube-api-access-brq7t\") pod \"heat-engine-7c94c94748-57b42\" (UID: \"515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a\") " pod="openstack/heat-engine-7c94c94748-57b42" Dec 06 06:18:01 crc kubenswrapper[4809]: I1206 06:18:01.697769 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a-combined-ca-bundle\") pod \"heat-engine-7c94c94748-57b42\" (UID: \"515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a\") " pod="openstack/heat-engine-7c94c94748-57b42" Dec 06 06:18:01 crc kubenswrapper[4809]: I1206 06:18:01.697786 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a-config-data\") pod \"heat-engine-7c94c94748-57b42\" (UID: \"515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a\") " pod="openstack/heat-engine-7c94c94748-57b42" Dec 06 06:18:01 crc kubenswrapper[4809]: I1206 06:18:01.697887 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a-config-data-custom\") pod \"heat-engine-7c94c94748-57b42\" (UID: \"515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a\") " pod="openstack/heat-engine-7c94c94748-57b42" Dec 06 06:18:01 crc kubenswrapper[4809]: I1206 06:18:01.783333 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-5drs7"] Dec 06 06:18:01 crc kubenswrapper[4809]: I1206 06:18:01.786916 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-5drs7" Dec 06 06:18:01 crc kubenswrapper[4809]: I1206 06:18:01.800889 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a-config-data-custom\") pod \"heat-engine-7c94c94748-57b42\" (UID: \"515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a\") " pod="openstack/heat-engine-7c94c94748-57b42" Dec 06 06:18:01 crc kubenswrapper[4809]: I1206 06:18:01.801071 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brq7t\" (UniqueName: \"kubernetes.io/projected/515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a-kube-api-access-brq7t\") pod \"heat-engine-7c94c94748-57b42\" (UID: \"515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a\") " pod="openstack/heat-engine-7c94c94748-57b42" Dec 06 06:18:01 crc kubenswrapper[4809]: I1206 06:18:01.801546 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a-combined-ca-bundle\") pod \"heat-engine-7c94c94748-57b42\" (UID: \"515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a\") " pod="openstack/heat-engine-7c94c94748-57b42" Dec 06 06:18:01 crc kubenswrapper[4809]: I1206 06:18:01.801586 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a-config-data\") pod \"heat-engine-7c94c94748-57b42\" (UID: \"515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a\") " pod="openstack/heat-engine-7c94c94748-57b42" Dec 06 06:18:01 crc kubenswrapper[4809]: I1206 06:18:01.818269 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a-config-data-custom\") pod \"heat-engine-7c94c94748-57b42\" (UID: \"515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a\") " pod="openstack/heat-engine-7c94c94748-57b42" Dec 06 06:18:01 crc kubenswrapper[4809]: I1206 06:18:01.821750 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a-combined-ca-bundle\") pod \"heat-engine-7c94c94748-57b42\" (UID: \"515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a\") " pod="openstack/heat-engine-7c94c94748-57b42" Dec 06 06:18:01 crc kubenswrapper[4809]: I1206 06:18:01.834049 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-7d5f5dd46-d69vn"] Dec 06 06:18:01 crc kubenswrapper[4809]: I1206 06:18:01.844347 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a-config-data\") pod \"heat-engine-7c94c94748-57b42\" (UID: \"515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a\") " pod="openstack/heat-engine-7c94c94748-57b42" Dec 06 06:18:01 crc kubenswrapper[4809]: I1206 06:18:01.851876 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brq7t\" (UniqueName: \"kubernetes.io/projected/515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a-kube-api-access-brq7t\") pod \"heat-engine-7c94c94748-57b42\" (UID: \"515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a\") " pod="openstack/heat-engine-7c94c94748-57b42" Dec 06 06:18:01 crc kubenswrapper[4809]: I1206 06:18:01.852542 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7d5f5dd46-d69vn" Dec 06 06:18:01 crc kubenswrapper[4809]: I1206 06:18:01.868056 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Dec 06 06:18:01 crc kubenswrapper[4809]: I1206 06:18:01.901305 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-5drs7"] Dec 06 06:18:01 crc kubenswrapper[4809]: I1206 06:18:01.904185 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3eed0435-bed9-4713-a316-acf245fd8872-config-data\") pod \"heat-cfnapi-7d5f5dd46-d69vn\" (UID: \"3eed0435-bed9-4713-a316-acf245fd8872\") " pod="openstack/heat-cfnapi-7d5f5dd46-d69vn" Dec 06 06:18:01 crc kubenswrapper[4809]: I1206 06:18:01.904321 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0710c66a-2ed5-4bf6-8f12-e3754d122d08-dns-svc\") pod \"dnsmasq-dns-7756b9d78c-5drs7\" (UID: \"0710c66a-2ed5-4bf6-8f12-e3754d122d08\") " pod="openstack/dnsmasq-dns-7756b9d78c-5drs7" Dec 06 06:18:01 crc kubenswrapper[4809]: I1206 06:18:01.904413 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3eed0435-bed9-4713-a316-acf245fd8872-combined-ca-bundle\") pod \"heat-cfnapi-7d5f5dd46-d69vn\" (UID: \"3eed0435-bed9-4713-a316-acf245fd8872\") " pod="openstack/heat-cfnapi-7d5f5dd46-d69vn" Dec 06 06:18:01 crc kubenswrapper[4809]: I1206 06:18:01.904462 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3eed0435-bed9-4713-a316-acf245fd8872-config-data-custom\") pod \"heat-cfnapi-7d5f5dd46-d69vn\" (UID: \"3eed0435-bed9-4713-a316-acf245fd8872\") " pod="openstack/heat-cfnapi-7d5f5dd46-d69vn" Dec 06 06:18:01 crc kubenswrapper[4809]: I1206 06:18:01.904488 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0710c66a-2ed5-4bf6-8f12-e3754d122d08-config\") pod \"dnsmasq-dns-7756b9d78c-5drs7\" (UID: \"0710c66a-2ed5-4bf6-8f12-e3754d122d08\") " pod="openstack/dnsmasq-dns-7756b9d78c-5drs7" Dec 06 06:18:01 crc kubenswrapper[4809]: I1206 06:18:01.904515 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmsrn\" (UniqueName: \"kubernetes.io/projected/3eed0435-bed9-4713-a316-acf245fd8872-kube-api-access-nmsrn\") pod \"heat-cfnapi-7d5f5dd46-d69vn\" (UID: \"3eed0435-bed9-4713-a316-acf245fd8872\") " pod="openstack/heat-cfnapi-7d5f5dd46-d69vn" Dec 06 06:18:01 crc kubenswrapper[4809]: I1206 06:18:01.904543 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0710c66a-2ed5-4bf6-8f12-e3754d122d08-ovsdbserver-nb\") pod \"dnsmasq-dns-7756b9d78c-5drs7\" (UID: \"0710c66a-2ed5-4bf6-8f12-e3754d122d08\") " pod="openstack/dnsmasq-dns-7756b9d78c-5drs7" Dec 06 06:18:01 crc kubenswrapper[4809]: I1206 06:18:01.904602 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0710c66a-2ed5-4bf6-8f12-e3754d122d08-ovsdbserver-sb\") pod \"dnsmasq-dns-7756b9d78c-5drs7\" (UID: \"0710c66a-2ed5-4bf6-8f12-e3754d122d08\") " pod="openstack/dnsmasq-dns-7756b9d78c-5drs7" Dec 06 06:18:01 crc kubenswrapper[4809]: I1206 06:18:01.904658 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0710c66a-2ed5-4bf6-8f12-e3754d122d08-dns-swift-storage-0\") pod \"dnsmasq-dns-7756b9d78c-5drs7\" (UID: \"0710c66a-2ed5-4bf6-8f12-e3754d122d08\") " pod="openstack/dnsmasq-dns-7756b9d78c-5drs7" Dec 06 06:18:01 crc kubenswrapper[4809]: I1206 06:18:01.904688 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lsdv\" (UniqueName: \"kubernetes.io/projected/0710c66a-2ed5-4bf6-8f12-e3754d122d08-kube-api-access-4lsdv\") pod \"dnsmasq-dns-7756b9d78c-5drs7\" (UID: \"0710c66a-2ed5-4bf6-8f12-e3754d122d08\") " pod="openstack/dnsmasq-dns-7756b9d78c-5drs7" Dec 06 06:18:01 crc kubenswrapper[4809]: I1206 06:18:01.933283 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-7d5f5dd46-d69vn"] Dec 06 06:18:01 crc kubenswrapper[4809]: I1206 06:18:01.968081 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-d46db5664-p2pfk"] Dec 06 06:18:01 crc kubenswrapper[4809]: I1206 06:18:01.972417 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-d46db5664-p2pfk" Dec 06 06:18:01 crc kubenswrapper[4809]: I1206 06:18:01.976659 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Dec 06 06:18:01 crc kubenswrapper[4809]: I1206 06:18:01.977253 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7c94c94748-57b42" Dec 06 06:18:01 crc kubenswrapper[4809]: I1206 06:18:01.997529 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-d46db5664-p2pfk"] Dec 06 06:18:01 crc kubenswrapper[4809]: I1206 06:18:01.997650 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-58bb4f66d7-bdk65" Dec 06 06:18:02 crc kubenswrapper[4809]: I1206 06:18:02.000191 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-58bb4f66d7-bdk65" Dec 06 06:18:02 crc kubenswrapper[4809]: I1206 06:18:02.007921 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/413fcc56-206e-402f-a6f9-a5f9a4cab527-config-data\") pod \"heat-api-d46db5664-p2pfk\" (UID: \"413fcc56-206e-402f-a6f9-a5f9a4cab527\") " pod="openstack/heat-api-d46db5664-p2pfk" Dec 06 06:18:02 crc kubenswrapper[4809]: I1206 06:18:02.008011 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3eed0435-bed9-4713-a316-acf245fd8872-combined-ca-bundle\") pod \"heat-cfnapi-7d5f5dd46-d69vn\" (UID: \"3eed0435-bed9-4713-a316-acf245fd8872\") " pod="openstack/heat-cfnapi-7d5f5dd46-d69vn" Dec 06 06:18:02 crc kubenswrapper[4809]: I1206 06:18:02.008038 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/413fcc56-206e-402f-a6f9-a5f9a4cab527-combined-ca-bundle\") pod \"heat-api-d46db5664-p2pfk\" (UID: \"413fcc56-206e-402f-a6f9-a5f9a4cab527\") " pod="openstack/heat-api-d46db5664-p2pfk" Dec 06 06:18:02 crc kubenswrapper[4809]: I1206 06:18:02.008080 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3eed0435-bed9-4713-a316-acf245fd8872-config-data-custom\") pod \"heat-cfnapi-7d5f5dd46-d69vn\" (UID: \"3eed0435-bed9-4713-a316-acf245fd8872\") " pod="openstack/heat-cfnapi-7d5f5dd46-d69vn" Dec 06 06:18:02 crc kubenswrapper[4809]: I1206 06:18:02.008109 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0710c66a-2ed5-4bf6-8f12-e3754d122d08-config\") pod \"dnsmasq-dns-7756b9d78c-5drs7\" (UID: \"0710c66a-2ed5-4bf6-8f12-e3754d122d08\") " pod="openstack/dnsmasq-dns-7756b9d78c-5drs7" Dec 06 06:18:02 crc kubenswrapper[4809]: I1206 06:18:02.008141 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmsrn\" (UniqueName: \"kubernetes.io/projected/3eed0435-bed9-4713-a316-acf245fd8872-kube-api-access-nmsrn\") pod \"heat-cfnapi-7d5f5dd46-d69vn\" (UID: \"3eed0435-bed9-4713-a316-acf245fd8872\") " pod="openstack/heat-cfnapi-7d5f5dd46-d69vn" Dec 06 06:18:02 crc kubenswrapper[4809]: I1206 06:18:02.008168 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0710c66a-2ed5-4bf6-8f12-e3754d122d08-ovsdbserver-nb\") pod \"dnsmasq-dns-7756b9d78c-5drs7\" (UID: \"0710c66a-2ed5-4bf6-8f12-e3754d122d08\") " pod="openstack/dnsmasq-dns-7756b9d78c-5drs7" Dec 06 06:18:02 crc kubenswrapper[4809]: I1206 06:18:02.008211 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgk4h\" (UniqueName: \"kubernetes.io/projected/413fcc56-206e-402f-a6f9-a5f9a4cab527-kube-api-access-sgk4h\") pod \"heat-api-d46db5664-p2pfk\" (UID: \"413fcc56-206e-402f-a6f9-a5f9a4cab527\") " pod="openstack/heat-api-d46db5664-p2pfk" Dec 06 06:18:02 crc kubenswrapper[4809]: I1206 06:18:02.008250 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0710c66a-2ed5-4bf6-8f12-e3754d122d08-ovsdbserver-sb\") pod \"dnsmasq-dns-7756b9d78c-5drs7\" (UID: \"0710c66a-2ed5-4bf6-8f12-e3754d122d08\") " pod="openstack/dnsmasq-dns-7756b9d78c-5drs7" Dec 06 06:18:02 crc kubenswrapper[4809]: I1206 06:18:02.008309 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0710c66a-2ed5-4bf6-8f12-e3754d122d08-dns-swift-storage-0\") pod \"dnsmasq-dns-7756b9d78c-5drs7\" (UID: \"0710c66a-2ed5-4bf6-8f12-e3754d122d08\") " pod="openstack/dnsmasq-dns-7756b9d78c-5drs7" Dec 06 06:18:02 crc kubenswrapper[4809]: I1206 06:18:02.008342 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lsdv\" (UniqueName: \"kubernetes.io/projected/0710c66a-2ed5-4bf6-8f12-e3754d122d08-kube-api-access-4lsdv\") pod \"dnsmasq-dns-7756b9d78c-5drs7\" (UID: \"0710c66a-2ed5-4bf6-8f12-e3754d122d08\") " pod="openstack/dnsmasq-dns-7756b9d78c-5drs7" Dec 06 06:18:02 crc kubenswrapper[4809]: I1206 06:18:02.008471 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3eed0435-bed9-4713-a316-acf245fd8872-config-data\") pod \"heat-cfnapi-7d5f5dd46-d69vn\" (UID: \"3eed0435-bed9-4713-a316-acf245fd8872\") " pod="openstack/heat-cfnapi-7d5f5dd46-d69vn" Dec 06 06:18:02 crc kubenswrapper[4809]: I1206 06:18:02.008538 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/413fcc56-206e-402f-a6f9-a5f9a4cab527-config-data-custom\") pod \"heat-api-d46db5664-p2pfk\" (UID: \"413fcc56-206e-402f-a6f9-a5f9a4cab527\") " pod="openstack/heat-api-d46db5664-p2pfk" Dec 06 06:18:02 crc kubenswrapper[4809]: I1206 06:18:02.008596 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0710c66a-2ed5-4bf6-8f12-e3754d122d08-dns-svc\") pod \"dnsmasq-dns-7756b9d78c-5drs7\" (UID: \"0710c66a-2ed5-4bf6-8f12-e3754d122d08\") " pod="openstack/dnsmasq-dns-7756b9d78c-5drs7" Dec 06 06:18:02 crc kubenswrapper[4809]: I1206 06:18:02.013723 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0710c66a-2ed5-4bf6-8f12-e3754d122d08-config\") pod \"dnsmasq-dns-7756b9d78c-5drs7\" (UID: \"0710c66a-2ed5-4bf6-8f12-e3754d122d08\") " pod="openstack/dnsmasq-dns-7756b9d78c-5drs7" Dec 06 06:18:02 crc kubenswrapper[4809]: I1206 06:18:02.014441 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0710c66a-2ed5-4bf6-8f12-e3754d122d08-dns-swift-storage-0\") pod \"dnsmasq-dns-7756b9d78c-5drs7\" (UID: \"0710c66a-2ed5-4bf6-8f12-e3754d122d08\") " pod="openstack/dnsmasq-dns-7756b9d78c-5drs7" Dec 06 06:18:02 crc kubenswrapper[4809]: I1206 06:18:02.015296 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0710c66a-2ed5-4bf6-8f12-e3754d122d08-ovsdbserver-sb\") pod \"dnsmasq-dns-7756b9d78c-5drs7\" (UID: \"0710c66a-2ed5-4bf6-8f12-e3754d122d08\") " pod="openstack/dnsmasq-dns-7756b9d78c-5drs7" Dec 06 06:18:02 crc kubenswrapper[4809]: I1206 06:18:02.015667 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0710c66a-2ed5-4bf6-8f12-e3754d122d08-ovsdbserver-nb\") pod \"dnsmasq-dns-7756b9d78c-5drs7\" (UID: \"0710c66a-2ed5-4bf6-8f12-e3754d122d08\") " pod="openstack/dnsmasq-dns-7756b9d78c-5drs7" Dec 06 06:18:02 crc kubenswrapper[4809]: I1206 06:18:02.022340 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3eed0435-bed9-4713-a316-acf245fd8872-combined-ca-bundle\") pod \"heat-cfnapi-7d5f5dd46-d69vn\" (UID: \"3eed0435-bed9-4713-a316-acf245fd8872\") " pod="openstack/heat-cfnapi-7d5f5dd46-d69vn" Dec 06 06:18:02 crc kubenswrapper[4809]: I1206 06:18:02.032184 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0710c66a-2ed5-4bf6-8f12-e3754d122d08-dns-svc\") pod \"dnsmasq-dns-7756b9d78c-5drs7\" (UID: \"0710c66a-2ed5-4bf6-8f12-e3754d122d08\") " pod="openstack/dnsmasq-dns-7756b9d78c-5drs7" Dec 06 06:18:02 crc kubenswrapper[4809]: I1206 06:18:02.039330 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmsrn\" (UniqueName: \"kubernetes.io/projected/3eed0435-bed9-4713-a316-acf245fd8872-kube-api-access-nmsrn\") pod \"heat-cfnapi-7d5f5dd46-d69vn\" (UID: \"3eed0435-bed9-4713-a316-acf245fd8872\") " pod="openstack/heat-cfnapi-7d5f5dd46-d69vn" Dec 06 06:18:02 crc kubenswrapper[4809]: I1206 06:18:02.043776 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3eed0435-bed9-4713-a316-acf245fd8872-config-data\") pod \"heat-cfnapi-7d5f5dd46-d69vn\" (UID: \"3eed0435-bed9-4713-a316-acf245fd8872\") " pod="openstack/heat-cfnapi-7d5f5dd46-d69vn" Dec 06 06:18:02 crc kubenswrapper[4809]: I1206 06:18:02.051679 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3eed0435-bed9-4713-a316-acf245fd8872-config-data-custom\") pod \"heat-cfnapi-7d5f5dd46-d69vn\" (UID: \"3eed0435-bed9-4713-a316-acf245fd8872\") " pod="openstack/heat-cfnapi-7d5f5dd46-d69vn" Dec 06 06:18:02 crc kubenswrapper[4809]: I1206 06:18:02.059601 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lsdv\" (UniqueName: \"kubernetes.io/projected/0710c66a-2ed5-4bf6-8f12-e3754d122d08-kube-api-access-4lsdv\") pod \"dnsmasq-dns-7756b9d78c-5drs7\" (UID: \"0710c66a-2ed5-4bf6-8f12-e3754d122d08\") " pod="openstack/dnsmasq-dns-7756b9d78c-5drs7" Dec 06 06:18:02 crc kubenswrapper[4809]: I1206 06:18:02.113298 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/413fcc56-206e-402f-a6f9-a5f9a4cab527-config-data-custom\") pod \"heat-api-d46db5664-p2pfk\" (UID: \"413fcc56-206e-402f-a6f9-a5f9a4cab527\") " pod="openstack/heat-api-d46db5664-p2pfk" Dec 06 06:18:02 crc kubenswrapper[4809]: I1206 06:18:02.113409 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/413fcc56-206e-402f-a6f9-a5f9a4cab527-config-data\") pod \"heat-api-d46db5664-p2pfk\" (UID: \"413fcc56-206e-402f-a6f9-a5f9a4cab527\") " pod="openstack/heat-api-d46db5664-p2pfk" Dec 06 06:18:02 crc kubenswrapper[4809]: I1206 06:18:02.113439 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/413fcc56-206e-402f-a6f9-a5f9a4cab527-combined-ca-bundle\") pod \"heat-api-d46db5664-p2pfk\" (UID: \"413fcc56-206e-402f-a6f9-a5f9a4cab527\") " pod="openstack/heat-api-d46db5664-p2pfk" Dec 06 06:18:02 crc kubenswrapper[4809]: I1206 06:18:02.113498 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgk4h\" (UniqueName: \"kubernetes.io/projected/413fcc56-206e-402f-a6f9-a5f9a4cab527-kube-api-access-sgk4h\") pod \"heat-api-d46db5664-p2pfk\" (UID: \"413fcc56-206e-402f-a6f9-a5f9a4cab527\") " pod="openstack/heat-api-d46db5664-p2pfk" Dec 06 06:18:02 crc kubenswrapper[4809]: I1206 06:18:02.135615 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/413fcc56-206e-402f-a6f9-a5f9a4cab527-combined-ca-bundle\") pod \"heat-api-d46db5664-p2pfk\" (UID: \"413fcc56-206e-402f-a6f9-a5f9a4cab527\") " pod="openstack/heat-api-d46db5664-p2pfk" Dec 06 06:18:02 crc kubenswrapper[4809]: I1206 06:18:02.154197 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/413fcc56-206e-402f-a6f9-a5f9a4cab527-config-data\") pod \"heat-api-d46db5664-p2pfk\" (UID: \"413fcc56-206e-402f-a6f9-a5f9a4cab527\") " pod="openstack/heat-api-d46db5664-p2pfk" Dec 06 06:18:02 crc kubenswrapper[4809]: I1206 06:18:02.154298 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/413fcc56-206e-402f-a6f9-a5f9a4cab527-config-data-custom\") pod \"heat-api-d46db5664-p2pfk\" (UID: \"413fcc56-206e-402f-a6f9-a5f9a4cab527\") " pod="openstack/heat-api-d46db5664-p2pfk" Dec 06 06:18:02 crc kubenswrapper[4809]: I1206 06:18:02.175087 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgk4h\" (UniqueName: \"kubernetes.io/projected/413fcc56-206e-402f-a6f9-a5f9a4cab527-kube-api-access-sgk4h\") pod \"heat-api-d46db5664-p2pfk\" (UID: \"413fcc56-206e-402f-a6f9-a5f9a4cab527\") " pod="openstack/heat-api-d46db5664-p2pfk" Dec 06 06:18:02 crc kubenswrapper[4809]: I1206 06:18:02.234262 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-5drs7" Dec 06 06:18:02 crc kubenswrapper[4809]: I1206 06:18:02.258508 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7d5f5dd46-d69vn" Dec 06 06:18:02 crc kubenswrapper[4809]: I1206 06:18:02.319290 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-d46db5664-p2pfk" Dec 06 06:18:02 crc kubenswrapper[4809]: I1206 06:18:02.916705 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 06 06:18:04 crc kubenswrapper[4809]: I1206 06:18:04.501025 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:18:04 crc kubenswrapper[4809]: I1206 06:18:04.501511 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:18:06 crc kubenswrapper[4809]: E1206 06:18:06.032286 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified" Dec 06 06:18:06 crc kubenswrapper[4809]: E1206 06:18:06.032716 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:openstackclient,Image:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,Command:[/bin/sleep],Args:[infinity],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5dch676h58dh659hc5h56fh645hd6h56ch5ddh559h75h575h5fbh56bh56dh677h76h5cch67dh54dh688h6dh57fh648h55h5fh7fh65bh7fhc5h68q,ValueFrom:nil,},EnvVar{Name:OS_CLOUD,Value:default,ValueFrom:nil,},EnvVar{Name:PROMETHEUS_CA_CERT,Value:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,ValueFrom:nil,},EnvVar{Name:PROMETHEUS_HOST,Value:metric-storage-prometheus.openstack.svc,ValueFrom:nil,},EnvVar{Name:PROMETHEUS_PORT,Value:9090,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:openstack-config,ReadOnly:false,MountPath:/home/cloud-admin/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/home/cloud-admin/.config/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/home/cloud-admin/cloudrc,SubPath:cloudrc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hxz42,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42401,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42401,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstackclient_openstack(6c6578f3-bf7b-4892-abc9-5f4f145c8d36): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 06:18:06 crc kubenswrapper[4809]: E1206 06:18:06.033920 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstackclient\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstackclient" podUID="6c6578f3-bf7b-4892-abc9-5f4f145c8d36" Dec 06 06:18:06 crc kubenswrapper[4809]: E1206 06:18:06.395979 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstackclient\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified\\\"\"" pod="openstack/openstackclient" podUID="6c6578f3-bf7b-4892-abc9-5f4f145c8d36" Dec 06 06:18:06 crc kubenswrapper[4809]: I1206 06:18:06.673399 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:18:06 crc kubenswrapper[4809]: I1206 06:18:06.759990 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bdfabc8f-27d1-4859-b309-800105752ea2-sg-core-conf-yaml\") pod \"bdfabc8f-27d1-4859-b309-800105752ea2\" (UID: \"bdfabc8f-27d1-4859-b309-800105752ea2\") " Dec 06 06:18:06 crc kubenswrapper[4809]: I1206 06:18:06.760089 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdfabc8f-27d1-4859-b309-800105752ea2-scripts\") pod \"bdfabc8f-27d1-4859-b309-800105752ea2\" (UID: \"bdfabc8f-27d1-4859-b309-800105752ea2\") " Dec 06 06:18:06 crc kubenswrapper[4809]: I1206 06:18:06.760183 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5ndln\" (UniqueName: \"kubernetes.io/projected/bdfabc8f-27d1-4859-b309-800105752ea2-kube-api-access-5ndln\") pod \"bdfabc8f-27d1-4859-b309-800105752ea2\" (UID: \"bdfabc8f-27d1-4859-b309-800105752ea2\") " Dec 06 06:18:06 crc kubenswrapper[4809]: I1206 06:18:06.760249 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bdfabc8f-27d1-4859-b309-800105752ea2-run-httpd\") pod \"bdfabc8f-27d1-4859-b309-800105752ea2\" (UID: \"bdfabc8f-27d1-4859-b309-800105752ea2\") " Dec 06 06:18:06 crc kubenswrapper[4809]: I1206 06:18:06.760281 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdfabc8f-27d1-4859-b309-800105752ea2-combined-ca-bundle\") pod \"bdfabc8f-27d1-4859-b309-800105752ea2\" (UID: \"bdfabc8f-27d1-4859-b309-800105752ea2\") " Dec 06 06:18:06 crc kubenswrapper[4809]: I1206 06:18:06.760344 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdfabc8f-27d1-4859-b309-800105752ea2-config-data\") pod \"bdfabc8f-27d1-4859-b309-800105752ea2\" (UID: \"bdfabc8f-27d1-4859-b309-800105752ea2\") " Dec 06 06:18:06 crc kubenswrapper[4809]: I1206 06:18:06.760458 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bdfabc8f-27d1-4859-b309-800105752ea2-log-httpd\") pod \"bdfabc8f-27d1-4859-b309-800105752ea2\" (UID: \"bdfabc8f-27d1-4859-b309-800105752ea2\") " Dec 06 06:18:06 crc kubenswrapper[4809]: I1206 06:18:06.761234 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bdfabc8f-27d1-4859-b309-800105752ea2-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "bdfabc8f-27d1-4859-b309-800105752ea2" (UID: "bdfabc8f-27d1-4859-b309-800105752ea2"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:18:06 crc kubenswrapper[4809]: I1206 06:18:06.762231 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bdfabc8f-27d1-4859-b309-800105752ea2-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "bdfabc8f-27d1-4859-b309-800105752ea2" (UID: "bdfabc8f-27d1-4859-b309-800105752ea2"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:18:06 crc kubenswrapper[4809]: I1206 06:18:06.764721 4809 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bdfabc8f-27d1-4859-b309-800105752ea2-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:06 crc kubenswrapper[4809]: I1206 06:18:06.764750 4809 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bdfabc8f-27d1-4859-b309-800105752ea2-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:06 crc kubenswrapper[4809]: I1206 06:18:06.767464 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdfabc8f-27d1-4859-b309-800105752ea2-scripts" (OuterVolumeSpecName: "scripts") pod "bdfabc8f-27d1-4859-b309-800105752ea2" (UID: "bdfabc8f-27d1-4859-b309-800105752ea2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:18:06 crc kubenswrapper[4809]: I1206 06:18:06.771012 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdfabc8f-27d1-4859-b309-800105752ea2-kube-api-access-5ndln" (OuterVolumeSpecName: "kube-api-access-5ndln") pod "bdfabc8f-27d1-4859-b309-800105752ea2" (UID: "bdfabc8f-27d1-4859-b309-800105752ea2"). InnerVolumeSpecName "kube-api-access-5ndln". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:18:06 crc kubenswrapper[4809]: I1206 06:18:06.821861 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdfabc8f-27d1-4859-b309-800105752ea2-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "bdfabc8f-27d1-4859-b309-800105752ea2" (UID: "bdfabc8f-27d1-4859-b309-800105752ea2"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:18:06 crc kubenswrapper[4809]: I1206 06:18:06.868522 4809 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bdfabc8f-27d1-4859-b309-800105752ea2-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:06 crc kubenswrapper[4809]: I1206 06:18:06.868548 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdfabc8f-27d1-4859-b309-800105752ea2-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:06 crc kubenswrapper[4809]: I1206 06:18:06.868558 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5ndln\" (UniqueName: \"kubernetes.io/projected/bdfabc8f-27d1-4859-b309-800105752ea2-kube-api-access-5ndln\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:06 crc kubenswrapper[4809]: I1206 06:18:06.933288 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdfabc8f-27d1-4859-b309-800105752ea2-config-data" (OuterVolumeSpecName: "config-data") pod "bdfabc8f-27d1-4859-b309-800105752ea2" (UID: "bdfabc8f-27d1-4859-b309-800105752ea2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:18:06 crc kubenswrapper[4809]: I1206 06:18:06.953263 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdfabc8f-27d1-4859-b309-800105752ea2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bdfabc8f-27d1-4859-b309-800105752ea2" (UID: "bdfabc8f-27d1-4859-b309-800105752ea2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:18:06 crc kubenswrapper[4809]: I1206 06:18:06.970638 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdfabc8f-27d1-4859-b309-800105752ea2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:06 crc kubenswrapper[4809]: I1206 06:18:06.970671 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdfabc8f-27d1-4859-b309-800105752ea2-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.129395 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-d46db5664-p2pfk"] Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.146051 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-7c94c94748-57b42"] Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.164242 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-5drs7"] Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.195256 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-7d5f5dd46-d69vn"] Dec 06 06:18:07 crc kubenswrapper[4809]: W1206 06:18:07.222082 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3eed0435_bed9_4713_a316_acf245fd8872.slice/crio-329e6470ae493f53d473445a5c2e6ea11a65543d179ed049883025c0ca38ebed WatchSource:0}: Error finding container 329e6470ae493f53d473445a5c2e6ea11a65543d179ed049883025c0ca38ebed: Status 404 returned error can't find the container with id 329e6470ae493f53d473445a5c2e6ea11a65543d179ed049883025c0ca38ebed Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.482129 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7c94c94748-57b42" event={"ID":"515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a","Type":"ContainerStarted","Data":"b0902d6876565ea7f2bc92882c1c0668113eb3beb5b752b48269e172803040a3"} Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.492347 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-d46db5664-p2pfk" event={"ID":"413fcc56-206e-402f-a6f9-a5f9a4cab527","Type":"ContainerStarted","Data":"625787070e1fbbb17d0d059c34c27c6c0301e9194d9c5dccffc36da08d3e462c"} Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.539258 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.539998 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bdfabc8f-27d1-4859-b309-800105752ea2","Type":"ContainerDied","Data":"d9c979911b0a02af6e503d80c63e45602c873e87f33c6a5c998c6132d8436998"} Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.540036 4809 scope.go:117] "RemoveContainer" containerID="06cd048b9712e9ee61382dc3d9de5071b45e0e65d356c1b9ba02459297769163" Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.556512 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-5drs7" event={"ID":"0710c66a-2ed5-4bf6-8f12-e3754d122d08","Type":"ContainerStarted","Data":"ea9bfdc08ba590ca8519cf0f03d895fa18fc120fad32675a8d9fb1e12c6b72b1"} Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.565986 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7d5f5dd46-d69vn" event={"ID":"3eed0435-bed9-4713-a316-acf245fd8872","Type":"ContainerStarted","Data":"329e6470ae493f53d473445a5c2e6ea11a65543d179ed049883025c0ca38ebed"} Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.580120 4809 scope.go:117] "RemoveContainer" containerID="5c5b288b360303a12ddd1cdc9a809c39f23fa448c90a28fffa55c8fe6b34e0c9" Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.598110 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.616103 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.656049 4809 scope.go:117] "RemoveContainer" containerID="e36372324edd2822f6662b8a4a84968d25802a012c83f32e0e9f87d761baa86a" Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.656708 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:18:07 crc kubenswrapper[4809]: E1206 06:18:07.657296 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdfabc8f-27d1-4859-b309-800105752ea2" containerName="ceilometer-notification-agent" Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.657318 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdfabc8f-27d1-4859-b309-800105752ea2" containerName="ceilometer-notification-agent" Dec 06 06:18:07 crc kubenswrapper[4809]: E1206 06:18:07.657326 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdfabc8f-27d1-4859-b309-800105752ea2" containerName="sg-core" Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.657333 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdfabc8f-27d1-4859-b309-800105752ea2" containerName="sg-core" Dec 06 06:18:07 crc kubenswrapper[4809]: E1206 06:18:07.657344 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdfabc8f-27d1-4859-b309-800105752ea2" containerName="ceilometer-central-agent" Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.657350 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdfabc8f-27d1-4859-b309-800105752ea2" containerName="ceilometer-central-agent" Dec 06 06:18:07 crc kubenswrapper[4809]: E1206 06:18:07.657375 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdfabc8f-27d1-4859-b309-800105752ea2" containerName="proxy-httpd" Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.657380 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdfabc8f-27d1-4859-b309-800105752ea2" containerName="proxy-httpd" Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.657605 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdfabc8f-27d1-4859-b309-800105752ea2" containerName="sg-core" Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.657617 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdfabc8f-27d1-4859-b309-800105752ea2" containerName="ceilometer-notification-agent" Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.657638 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdfabc8f-27d1-4859-b309-800105752ea2" containerName="proxy-httpd" Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.657655 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdfabc8f-27d1-4859-b309-800105752ea2" containerName="ceilometer-central-agent" Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.673728 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.676816 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.682349 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.682647 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.710407 4809 scope.go:117] "RemoveContainer" containerID="ac843dc132759e8b65936f4165e8a3c3486a36b5a3ca6303cc78964646d90d8f" Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.791305 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53615cf8-5ea5-4d28-a4e7-1491e2b33744-config-data\") pod \"ceilometer-0\" (UID: \"53615cf8-5ea5-4d28-a4e7-1491e2b33744\") " pod="openstack/ceilometer-0" Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.791757 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53615cf8-5ea5-4d28-a4e7-1491e2b33744-scripts\") pod \"ceilometer-0\" (UID: \"53615cf8-5ea5-4d28-a4e7-1491e2b33744\") " pod="openstack/ceilometer-0" Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.791782 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/53615cf8-5ea5-4d28-a4e7-1491e2b33744-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"53615cf8-5ea5-4d28-a4e7-1491e2b33744\") " pod="openstack/ceilometer-0" Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.791884 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53615cf8-5ea5-4d28-a4e7-1491e2b33744-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"53615cf8-5ea5-4d28-a4e7-1491e2b33744\") " pod="openstack/ceilometer-0" Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.791980 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pskh5\" (UniqueName: \"kubernetes.io/projected/53615cf8-5ea5-4d28-a4e7-1491e2b33744-kube-api-access-pskh5\") pod \"ceilometer-0\" (UID: \"53615cf8-5ea5-4d28-a4e7-1491e2b33744\") " pod="openstack/ceilometer-0" Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.792229 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53615cf8-5ea5-4d28-a4e7-1491e2b33744-log-httpd\") pod \"ceilometer-0\" (UID: \"53615cf8-5ea5-4d28-a4e7-1491e2b33744\") " pod="openstack/ceilometer-0" Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.792292 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53615cf8-5ea5-4d28-a4e7-1491e2b33744-run-httpd\") pod \"ceilometer-0\" (UID: \"53615cf8-5ea5-4d28-a4e7-1491e2b33744\") " pod="openstack/ceilometer-0" Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.911433 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pskh5\" (UniqueName: \"kubernetes.io/projected/53615cf8-5ea5-4d28-a4e7-1491e2b33744-kube-api-access-pskh5\") pod \"ceilometer-0\" (UID: \"53615cf8-5ea5-4d28-a4e7-1491e2b33744\") " pod="openstack/ceilometer-0" Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.911555 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53615cf8-5ea5-4d28-a4e7-1491e2b33744-log-httpd\") pod \"ceilometer-0\" (UID: \"53615cf8-5ea5-4d28-a4e7-1491e2b33744\") " pod="openstack/ceilometer-0" Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.911585 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53615cf8-5ea5-4d28-a4e7-1491e2b33744-run-httpd\") pod \"ceilometer-0\" (UID: \"53615cf8-5ea5-4d28-a4e7-1491e2b33744\") " pod="openstack/ceilometer-0" Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.911651 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53615cf8-5ea5-4d28-a4e7-1491e2b33744-config-data\") pod \"ceilometer-0\" (UID: \"53615cf8-5ea5-4d28-a4e7-1491e2b33744\") " pod="openstack/ceilometer-0" Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.911750 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53615cf8-5ea5-4d28-a4e7-1491e2b33744-scripts\") pod \"ceilometer-0\" (UID: \"53615cf8-5ea5-4d28-a4e7-1491e2b33744\") " pod="openstack/ceilometer-0" Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.911775 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/53615cf8-5ea5-4d28-a4e7-1491e2b33744-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"53615cf8-5ea5-4d28-a4e7-1491e2b33744\") " pod="openstack/ceilometer-0" Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.911841 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53615cf8-5ea5-4d28-a4e7-1491e2b33744-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"53615cf8-5ea5-4d28-a4e7-1491e2b33744\") " pod="openstack/ceilometer-0" Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.912117 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53615cf8-5ea5-4d28-a4e7-1491e2b33744-log-httpd\") pod \"ceilometer-0\" (UID: \"53615cf8-5ea5-4d28-a4e7-1491e2b33744\") " pod="openstack/ceilometer-0" Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.912411 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53615cf8-5ea5-4d28-a4e7-1491e2b33744-run-httpd\") pod \"ceilometer-0\" (UID: \"53615cf8-5ea5-4d28-a4e7-1491e2b33744\") " pod="openstack/ceilometer-0" Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.921616 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53615cf8-5ea5-4d28-a4e7-1491e2b33744-scripts\") pod \"ceilometer-0\" (UID: \"53615cf8-5ea5-4d28-a4e7-1491e2b33744\") " pod="openstack/ceilometer-0" Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.924829 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53615cf8-5ea5-4d28-a4e7-1491e2b33744-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"53615cf8-5ea5-4d28-a4e7-1491e2b33744\") " pod="openstack/ceilometer-0" Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.927731 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53615cf8-5ea5-4d28-a4e7-1491e2b33744-config-data\") pod \"ceilometer-0\" (UID: \"53615cf8-5ea5-4d28-a4e7-1491e2b33744\") " pod="openstack/ceilometer-0" Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.929684 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/53615cf8-5ea5-4d28-a4e7-1491e2b33744-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"53615cf8-5ea5-4d28-a4e7-1491e2b33744\") " pod="openstack/ceilometer-0" Dec 06 06:18:07 crc kubenswrapper[4809]: I1206 06:18:07.964305 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pskh5\" (UniqueName: \"kubernetes.io/projected/53615cf8-5ea5-4d28-a4e7-1491e2b33744-kube-api-access-pskh5\") pod \"ceilometer-0\" (UID: \"53615cf8-5ea5-4d28-a4e7-1491e2b33744\") " pod="openstack/ceilometer-0" Dec 06 06:18:07 crc kubenswrapper[4809]: E1206 06:18:07.967801 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0710c66a_2ed5_4bf6_8f12_e3754d122d08.slice/crio-c6e22fe4c2b6cc2192719a3b4daa8bcc285d59c1befbaf185eb3d743da6af958.scope\": RecentStats: unable to find data in memory cache]" Dec 06 06:18:08 crc kubenswrapper[4809]: I1206 06:18:08.033217 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:18:08 crc kubenswrapper[4809]: I1206 06:18:08.613984 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7c94c94748-57b42" event={"ID":"515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a","Type":"ContainerStarted","Data":"1b4497e7e985b3dcecad685e6cbc0fe4e3ac63269895ab1f86489595826907ef"} Dec 06 06:18:08 crc kubenswrapper[4809]: I1206 06:18:08.616479 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-7c94c94748-57b42" Dec 06 06:18:08 crc kubenswrapper[4809]: I1206 06:18:08.620324 4809 generic.go:334] "Generic (PLEG): container finished" podID="0710c66a-2ed5-4bf6-8f12-e3754d122d08" containerID="c6e22fe4c2b6cc2192719a3b4daa8bcc285d59c1befbaf185eb3d743da6af958" exitCode=0 Dec 06 06:18:08 crc kubenswrapper[4809]: I1206 06:18:08.620368 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-5drs7" event={"ID":"0710c66a-2ed5-4bf6-8f12-e3754d122d08","Type":"ContainerDied","Data":"c6e22fe4c2b6cc2192719a3b4daa8bcc285d59c1befbaf185eb3d743da6af958"} Dec 06 06:18:08 crc kubenswrapper[4809]: I1206 06:18:08.643834 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-7c94c94748-57b42" podStartSLOduration=7.643807186 podStartE2EDuration="7.643807186s" podCreationTimestamp="2025-12-06 06:18:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:18:08.638489573 +0000 UTC m=+1613.527472525" watchObservedRunningTime="2025-12-06 06:18:08.643807186 +0000 UTC m=+1613.532790128" Dec 06 06:18:08 crc kubenswrapper[4809]: I1206 06:18:08.686564 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.417150 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bdfabc8f-27d1-4859-b309-800105752ea2" path="/var/lib/kubelet/pods/bdfabc8f-27d1-4859-b309-800105752ea2/volumes" Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.561496 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-6d8ccd7c9f-nbxnr"] Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.562941 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-6d8ccd7c9f-nbxnr" Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.587123 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-6d8ccd7c9f-nbxnr"] Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.595781 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-97f66c66d-jlz22"] Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.602339 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-97f66c66d-jlz22" Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.621641 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-c48964f4f-ktxp6"] Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.623429 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-c48964f4f-ktxp6" Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.640452 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-97f66c66d-jlz22"] Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.656030 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"53615cf8-5ea5-4d28-a4e7-1491e2b33744","Type":"ContainerStarted","Data":"d3427867fc1f8bf40890103584f0e3c254bffa7e0aae8bc3ccceb041647e1727"} Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.661508 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-c48964f4f-ktxp6"] Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.680401 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbw2p\" (UniqueName: \"kubernetes.io/projected/04d8a790-dedd-44c4-9b8f-093abb541e73-kube-api-access-bbw2p\") pod \"heat-engine-6d8ccd7c9f-nbxnr\" (UID: \"04d8a790-dedd-44c4-9b8f-093abb541e73\") " pod="openstack/heat-engine-6d8ccd7c9f-nbxnr" Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.680501 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04d8a790-dedd-44c4-9b8f-093abb541e73-combined-ca-bundle\") pod \"heat-engine-6d8ccd7c9f-nbxnr\" (UID: \"04d8a790-dedd-44c4-9b8f-093abb541e73\") " pod="openstack/heat-engine-6d8ccd7c9f-nbxnr" Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.680521 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/04d8a790-dedd-44c4-9b8f-093abb541e73-config-data-custom\") pod \"heat-engine-6d8ccd7c9f-nbxnr\" (UID: \"04d8a790-dedd-44c4-9b8f-093abb541e73\") " pod="openstack/heat-engine-6d8ccd7c9f-nbxnr" Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.680571 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04d8a790-dedd-44c4-9b8f-093abb541e73-config-data\") pod \"heat-engine-6d8ccd7c9f-nbxnr\" (UID: \"04d8a790-dedd-44c4-9b8f-093abb541e73\") " pod="openstack/heat-engine-6d8ccd7c9f-nbxnr" Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.783004 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04d8a790-dedd-44c4-9b8f-093abb541e73-combined-ca-bundle\") pod \"heat-engine-6d8ccd7c9f-nbxnr\" (UID: \"04d8a790-dedd-44c4-9b8f-093abb541e73\") " pod="openstack/heat-engine-6d8ccd7c9f-nbxnr" Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.783086 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/04d8a790-dedd-44c4-9b8f-093abb541e73-config-data-custom\") pod \"heat-engine-6d8ccd7c9f-nbxnr\" (UID: \"04d8a790-dedd-44c4-9b8f-093abb541e73\") " pod="openstack/heat-engine-6d8ccd7c9f-nbxnr" Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.783126 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0a5c1cb9-6796-43cd-8c09-68ade099a586-config-data-custom\") pod \"heat-cfnapi-c48964f4f-ktxp6\" (UID: \"0a5c1cb9-6796-43cd-8c09-68ade099a586\") " pod="openstack/heat-cfnapi-c48964f4f-ktxp6" Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.783194 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgj5f\" (UniqueName: \"kubernetes.io/projected/0a5c1cb9-6796-43cd-8c09-68ade099a586-kube-api-access-bgj5f\") pod \"heat-cfnapi-c48964f4f-ktxp6\" (UID: \"0a5c1cb9-6796-43cd-8c09-68ade099a586\") " pod="openstack/heat-cfnapi-c48964f4f-ktxp6" Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.783232 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04d8a790-dedd-44c4-9b8f-093abb541e73-config-data\") pod \"heat-engine-6d8ccd7c9f-nbxnr\" (UID: \"04d8a790-dedd-44c4-9b8f-093abb541e73\") " pod="openstack/heat-engine-6d8ccd7c9f-nbxnr" Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.783384 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a5c1cb9-6796-43cd-8c09-68ade099a586-config-data\") pod \"heat-cfnapi-c48964f4f-ktxp6\" (UID: \"0a5c1cb9-6796-43cd-8c09-68ade099a586\") " pod="openstack/heat-cfnapi-c48964f4f-ktxp6" Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.783574 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a5c1cb9-6796-43cd-8c09-68ade099a586-combined-ca-bundle\") pod \"heat-cfnapi-c48964f4f-ktxp6\" (UID: \"0a5c1cb9-6796-43cd-8c09-68ade099a586\") " pod="openstack/heat-cfnapi-c48964f4f-ktxp6" Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.783636 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85faa375-eb95-49b6-ba0b-967eeeebb2e8-config-data\") pod \"heat-api-97f66c66d-jlz22\" (UID: \"85faa375-eb95-49b6-ba0b-967eeeebb2e8\") " pod="openstack/heat-api-97f66c66d-jlz22" Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.783679 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85faa375-eb95-49b6-ba0b-967eeeebb2e8-combined-ca-bundle\") pod \"heat-api-97f66c66d-jlz22\" (UID: \"85faa375-eb95-49b6-ba0b-967eeeebb2e8\") " pod="openstack/heat-api-97f66c66d-jlz22" Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.783863 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/85faa375-eb95-49b6-ba0b-967eeeebb2e8-config-data-custom\") pod \"heat-api-97f66c66d-jlz22\" (UID: \"85faa375-eb95-49b6-ba0b-967eeeebb2e8\") " pod="openstack/heat-api-97f66c66d-jlz22" Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.784419 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbw2p\" (UniqueName: \"kubernetes.io/projected/04d8a790-dedd-44c4-9b8f-093abb541e73-kube-api-access-bbw2p\") pod \"heat-engine-6d8ccd7c9f-nbxnr\" (UID: \"04d8a790-dedd-44c4-9b8f-093abb541e73\") " pod="openstack/heat-engine-6d8ccd7c9f-nbxnr" Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.784619 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8d6v8\" (UniqueName: \"kubernetes.io/projected/85faa375-eb95-49b6-ba0b-967eeeebb2e8-kube-api-access-8d6v8\") pod \"heat-api-97f66c66d-jlz22\" (UID: \"85faa375-eb95-49b6-ba0b-967eeeebb2e8\") " pod="openstack/heat-api-97f66c66d-jlz22" Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.789177 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04d8a790-dedd-44c4-9b8f-093abb541e73-config-data\") pod \"heat-engine-6d8ccd7c9f-nbxnr\" (UID: \"04d8a790-dedd-44c4-9b8f-093abb541e73\") " pod="openstack/heat-engine-6d8ccd7c9f-nbxnr" Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.789295 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/04d8a790-dedd-44c4-9b8f-093abb541e73-config-data-custom\") pod \"heat-engine-6d8ccd7c9f-nbxnr\" (UID: \"04d8a790-dedd-44c4-9b8f-093abb541e73\") " pod="openstack/heat-engine-6d8ccd7c9f-nbxnr" Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.790140 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04d8a790-dedd-44c4-9b8f-093abb541e73-combined-ca-bundle\") pod \"heat-engine-6d8ccd7c9f-nbxnr\" (UID: \"04d8a790-dedd-44c4-9b8f-093abb541e73\") " pod="openstack/heat-engine-6d8ccd7c9f-nbxnr" Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.804616 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbw2p\" (UniqueName: \"kubernetes.io/projected/04d8a790-dedd-44c4-9b8f-093abb541e73-kube-api-access-bbw2p\") pod \"heat-engine-6d8ccd7c9f-nbxnr\" (UID: \"04d8a790-dedd-44c4-9b8f-093abb541e73\") " pod="openstack/heat-engine-6d8ccd7c9f-nbxnr" Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.881978 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-6d8ccd7c9f-nbxnr" Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.886355 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0a5c1cb9-6796-43cd-8c09-68ade099a586-config-data-custom\") pod \"heat-cfnapi-c48964f4f-ktxp6\" (UID: \"0a5c1cb9-6796-43cd-8c09-68ade099a586\") " pod="openstack/heat-cfnapi-c48964f4f-ktxp6" Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.886444 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgj5f\" (UniqueName: \"kubernetes.io/projected/0a5c1cb9-6796-43cd-8c09-68ade099a586-kube-api-access-bgj5f\") pod \"heat-cfnapi-c48964f4f-ktxp6\" (UID: \"0a5c1cb9-6796-43cd-8c09-68ade099a586\") " pod="openstack/heat-cfnapi-c48964f4f-ktxp6" Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.886508 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a5c1cb9-6796-43cd-8c09-68ade099a586-config-data\") pod \"heat-cfnapi-c48964f4f-ktxp6\" (UID: \"0a5c1cb9-6796-43cd-8c09-68ade099a586\") " pod="openstack/heat-cfnapi-c48964f4f-ktxp6" Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.886593 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a5c1cb9-6796-43cd-8c09-68ade099a586-combined-ca-bundle\") pod \"heat-cfnapi-c48964f4f-ktxp6\" (UID: \"0a5c1cb9-6796-43cd-8c09-68ade099a586\") " pod="openstack/heat-cfnapi-c48964f4f-ktxp6" Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.886643 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85faa375-eb95-49b6-ba0b-967eeeebb2e8-config-data\") pod \"heat-api-97f66c66d-jlz22\" (UID: \"85faa375-eb95-49b6-ba0b-967eeeebb2e8\") " pod="openstack/heat-api-97f66c66d-jlz22" Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.886677 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85faa375-eb95-49b6-ba0b-967eeeebb2e8-combined-ca-bundle\") pod \"heat-api-97f66c66d-jlz22\" (UID: \"85faa375-eb95-49b6-ba0b-967eeeebb2e8\") " pod="openstack/heat-api-97f66c66d-jlz22" Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.886721 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/85faa375-eb95-49b6-ba0b-967eeeebb2e8-config-data-custom\") pod \"heat-api-97f66c66d-jlz22\" (UID: \"85faa375-eb95-49b6-ba0b-967eeeebb2e8\") " pod="openstack/heat-api-97f66c66d-jlz22" Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.886830 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8d6v8\" (UniqueName: \"kubernetes.io/projected/85faa375-eb95-49b6-ba0b-967eeeebb2e8-kube-api-access-8d6v8\") pod \"heat-api-97f66c66d-jlz22\" (UID: \"85faa375-eb95-49b6-ba0b-967eeeebb2e8\") " pod="openstack/heat-api-97f66c66d-jlz22" Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.891847 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a5c1cb9-6796-43cd-8c09-68ade099a586-config-data\") pod \"heat-cfnapi-c48964f4f-ktxp6\" (UID: \"0a5c1cb9-6796-43cd-8c09-68ade099a586\") " pod="openstack/heat-cfnapi-c48964f4f-ktxp6" Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.892300 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/85faa375-eb95-49b6-ba0b-967eeeebb2e8-config-data-custom\") pod \"heat-api-97f66c66d-jlz22\" (UID: \"85faa375-eb95-49b6-ba0b-967eeeebb2e8\") " pod="openstack/heat-api-97f66c66d-jlz22" Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.892650 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0a5c1cb9-6796-43cd-8c09-68ade099a586-config-data-custom\") pod \"heat-cfnapi-c48964f4f-ktxp6\" (UID: \"0a5c1cb9-6796-43cd-8c09-68ade099a586\") " pod="openstack/heat-cfnapi-c48964f4f-ktxp6" Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.895568 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85faa375-eb95-49b6-ba0b-967eeeebb2e8-combined-ca-bundle\") pod \"heat-api-97f66c66d-jlz22\" (UID: \"85faa375-eb95-49b6-ba0b-967eeeebb2e8\") " pod="openstack/heat-api-97f66c66d-jlz22" Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.896444 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85faa375-eb95-49b6-ba0b-967eeeebb2e8-config-data\") pod \"heat-api-97f66c66d-jlz22\" (UID: \"85faa375-eb95-49b6-ba0b-967eeeebb2e8\") " pod="openstack/heat-api-97f66c66d-jlz22" Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.916495 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8d6v8\" (UniqueName: \"kubernetes.io/projected/85faa375-eb95-49b6-ba0b-967eeeebb2e8-kube-api-access-8d6v8\") pod \"heat-api-97f66c66d-jlz22\" (UID: \"85faa375-eb95-49b6-ba0b-967eeeebb2e8\") " pod="openstack/heat-api-97f66c66d-jlz22" Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.918649 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a5c1cb9-6796-43cd-8c09-68ade099a586-combined-ca-bundle\") pod \"heat-cfnapi-c48964f4f-ktxp6\" (UID: \"0a5c1cb9-6796-43cd-8c09-68ade099a586\") " pod="openstack/heat-cfnapi-c48964f4f-ktxp6" Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.930716 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgj5f\" (UniqueName: \"kubernetes.io/projected/0a5c1cb9-6796-43cd-8c09-68ade099a586-kube-api-access-bgj5f\") pod \"heat-cfnapi-c48964f4f-ktxp6\" (UID: \"0a5c1cb9-6796-43cd-8c09-68ade099a586\") " pod="openstack/heat-cfnapi-c48964f4f-ktxp6" Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.939599 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-97f66c66d-jlz22" Dec 06 06:18:09 crc kubenswrapper[4809]: I1206 06:18:09.950610 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-c48964f4f-ktxp6" Dec 06 06:18:11 crc kubenswrapper[4809]: I1206 06:18:11.090319 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-6d8ccd7c9f-nbxnr"] Dec 06 06:18:11 crc kubenswrapper[4809]: W1206 06:18:11.091597 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod04d8a790_dedd_44c4_9b8f_093abb541e73.slice/crio-c8a90f915d63b2ce34446edce039cffc57b5a7aa443396df4dc86065c69b389d WatchSource:0}: Error finding container c8a90f915d63b2ce34446edce039cffc57b5a7aa443396df4dc86065c69b389d: Status 404 returned error can't find the container with id c8a90f915d63b2ce34446edce039cffc57b5a7aa443396df4dc86065c69b389d Dec 06 06:18:11 crc kubenswrapper[4809]: I1206 06:18:11.386169 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-97f66c66d-jlz22"] Dec 06 06:18:11 crc kubenswrapper[4809]: I1206 06:18:11.428511 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-c48964f4f-ktxp6"] Dec 06 06:18:11 crc kubenswrapper[4809]: I1206 06:18:11.647666 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-d46db5664-p2pfk"] Dec 06 06:18:11 crc kubenswrapper[4809]: I1206 06:18:11.679327 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-7d5f5dd46-d69vn"] Dec 06 06:18:11 crc kubenswrapper[4809]: I1206 06:18:11.710867 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-6d8ccd7c9f-nbxnr" event={"ID":"04d8a790-dedd-44c4-9b8f-093abb541e73","Type":"ContainerStarted","Data":"c8a90f915d63b2ce34446edce039cffc57b5a7aa443396df4dc86065c69b389d"} Dec 06 06:18:11 crc kubenswrapper[4809]: I1206 06:18:11.718453 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-97f66c66d-jlz22" event={"ID":"85faa375-eb95-49b6-ba0b-967eeeebb2e8","Type":"ContainerStarted","Data":"5196b40cec8315a33e9d4b30a8977972d17c2168ecc1c198817eb75438793ac6"} Dec 06 06:18:11 crc kubenswrapper[4809]: I1206 06:18:11.720778 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-c48964f4f-ktxp6" event={"ID":"0a5c1cb9-6796-43cd-8c09-68ade099a586","Type":"ContainerStarted","Data":"88ee53f5e833ba9abf430b5608360b12bdf739e3bfb70da2ac427465fc1e150f"} Dec 06 06:18:11 crc kubenswrapper[4809]: I1206 06:18:11.725985 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-9f9dd4fbb-4pcn7"] Dec 06 06:18:11 crc kubenswrapper[4809]: I1206 06:18:11.728566 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-9f9dd4fbb-4pcn7" Dec 06 06:18:11 crc kubenswrapper[4809]: I1206 06:18:11.736324 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-internal-svc" Dec 06 06:18:11 crc kubenswrapper[4809]: I1206 06:18:11.736773 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-public-svc" Dec 06 06:18:11 crc kubenswrapper[4809]: I1206 06:18:11.736165 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-bdfbd85-ntgxq"] Dec 06 06:18:11 crc kubenswrapper[4809]: I1206 06:18:11.742158 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-bdfbd85-ntgxq" Dec 06 06:18:11 crc kubenswrapper[4809]: I1206 06:18:11.744290 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-public-svc" Dec 06 06:18:11 crc kubenswrapper[4809]: I1206 06:18:11.744583 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-internal-svc" Dec 06 06:18:11 crc kubenswrapper[4809]: I1206 06:18:11.758174 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-bdfbd85-ntgxq"] Dec 06 06:18:11 crc kubenswrapper[4809]: I1206 06:18:11.775294 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-9f9dd4fbb-4pcn7"] Dec 06 06:18:11 crc kubenswrapper[4809]: I1206 06:18:11.893107 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grs95\" (UniqueName: \"kubernetes.io/projected/e2c1d3c1-fa26-4525-bca8-3449d4535409-kube-api-access-grs95\") pod \"heat-cfnapi-9f9dd4fbb-4pcn7\" (UID: \"e2c1d3c1-fa26-4525-bca8-3449d4535409\") " pod="openstack/heat-cfnapi-9f9dd4fbb-4pcn7" Dec 06 06:18:11 crc kubenswrapper[4809]: I1206 06:18:11.893154 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e2c1d3c1-fa26-4525-bca8-3449d4535409-config-data-custom\") pod \"heat-cfnapi-9f9dd4fbb-4pcn7\" (UID: \"e2c1d3c1-fa26-4525-bca8-3449d4535409\") " pod="openstack/heat-cfnapi-9f9dd4fbb-4pcn7" Dec 06 06:18:11 crc kubenswrapper[4809]: I1206 06:18:11.893205 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2c1d3c1-fa26-4525-bca8-3449d4535409-public-tls-certs\") pod \"heat-cfnapi-9f9dd4fbb-4pcn7\" (UID: \"e2c1d3c1-fa26-4525-bca8-3449d4535409\") " pod="openstack/heat-cfnapi-9f9dd4fbb-4pcn7" Dec 06 06:18:11 crc kubenswrapper[4809]: I1206 06:18:11.893250 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d76c911-21ec-4f29-9707-e8101ed3aed9-config-data\") pod \"heat-api-bdfbd85-ntgxq\" (UID: \"4d76c911-21ec-4f29-9707-e8101ed3aed9\") " pod="openstack/heat-api-bdfbd85-ntgxq" Dec 06 06:18:11 crc kubenswrapper[4809]: I1206 06:18:11.893268 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4d76c911-21ec-4f29-9707-e8101ed3aed9-config-data-custom\") pod \"heat-api-bdfbd85-ntgxq\" (UID: \"4d76c911-21ec-4f29-9707-e8101ed3aed9\") " pod="openstack/heat-api-bdfbd85-ntgxq" Dec 06 06:18:11 crc kubenswrapper[4809]: I1206 06:18:11.893287 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d76c911-21ec-4f29-9707-e8101ed3aed9-public-tls-certs\") pod \"heat-api-bdfbd85-ntgxq\" (UID: \"4d76c911-21ec-4f29-9707-e8101ed3aed9\") " pod="openstack/heat-api-bdfbd85-ntgxq" Dec 06 06:18:11 crc kubenswrapper[4809]: I1206 06:18:11.893312 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d76c911-21ec-4f29-9707-e8101ed3aed9-combined-ca-bundle\") pod \"heat-api-bdfbd85-ntgxq\" (UID: \"4d76c911-21ec-4f29-9707-e8101ed3aed9\") " pod="openstack/heat-api-bdfbd85-ntgxq" Dec 06 06:18:11 crc kubenswrapper[4809]: I1206 06:18:11.893382 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2c1d3c1-fa26-4525-bca8-3449d4535409-combined-ca-bundle\") pod \"heat-cfnapi-9f9dd4fbb-4pcn7\" (UID: \"e2c1d3c1-fa26-4525-bca8-3449d4535409\") " pod="openstack/heat-cfnapi-9f9dd4fbb-4pcn7" Dec 06 06:18:11 crc kubenswrapper[4809]: I1206 06:18:11.893424 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2c1d3c1-fa26-4525-bca8-3449d4535409-internal-tls-certs\") pod \"heat-cfnapi-9f9dd4fbb-4pcn7\" (UID: \"e2c1d3c1-fa26-4525-bca8-3449d4535409\") " pod="openstack/heat-cfnapi-9f9dd4fbb-4pcn7" Dec 06 06:18:11 crc kubenswrapper[4809]: I1206 06:18:11.893486 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5z7dq\" (UniqueName: \"kubernetes.io/projected/4d76c911-21ec-4f29-9707-e8101ed3aed9-kube-api-access-5z7dq\") pod \"heat-api-bdfbd85-ntgxq\" (UID: \"4d76c911-21ec-4f29-9707-e8101ed3aed9\") " pod="openstack/heat-api-bdfbd85-ntgxq" Dec 06 06:18:11 crc kubenswrapper[4809]: I1206 06:18:11.896486 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2c1d3c1-fa26-4525-bca8-3449d4535409-config-data\") pod \"heat-cfnapi-9f9dd4fbb-4pcn7\" (UID: \"e2c1d3c1-fa26-4525-bca8-3449d4535409\") " pod="openstack/heat-cfnapi-9f9dd4fbb-4pcn7" Dec 06 06:18:11 crc kubenswrapper[4809]: I1206 06:18:11.896659 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d76c911-21ec-4f29-9707-e8101ed3aed9-internal-tls-certs\") pod \"heat-api-bdfbd85-ntgxq\" (UID: \"4d76c911-21ec-4f29-9707-e8101ed3aed9\") " pod="openstack/heat-api-bdfbd85-ntgxq" Dec 06 06:18:11 crc kubenswrapper[4809]: I1206 06:18:11.998819 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2c1d3c1-fa26-4525-bca8-3449d4535409-config-data\") pod \"heat-cfnapi-9f9dd4fbb-4pcn7\" (UID: \"e2c1d3c1-fa26-4525-bca8-3449d4535409\") " pod="openstack/heat-cfnapi-9f9dd4fbb-4pcn7" Dec 06 06:18:11 crc kubenswrapper[4809]: I1206 06:18:11.998881 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d76c911-21ec-4f29-9707-e8101ed3aed9-internal-tls-certs\") pod \"heat-api-bdfbd85-ntgxq\" (UID: \"4d76c911-21ec-4f29-9707-e8101ed3aed9\") " pod="openstack/heat-api-bdfbd85-ntgxq" Dec 06 06:18:11 crc kubenswrapper[4809]: I1206 06:18:11.998915 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grs95\" (UniqueName: \"kubernetes.io/projected/e2c1d3c1-fa26-4525-bca8-3449d4535409-kube-api-access-grs95\") pod \"heat-cfnapi-9f9dd4fbb-4pcn7\" (UID: \"e2c1d3c1-fa26-4525-bca8-3449d4535409\") " pod="openstack/heat-cfnapi-9f9dd4fbb-4pcn7" Dec 06 06:18:11 crc kubenswrapper[4809]: I1206 06:18:11.998953 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e2c1d3c1-fa26-4525-bca8-3449d4535409-config-data-custom\") pod \"heat-cfnapi-9f9dd4fbb-4pcn7\" (UID: \"e2c1d3c1-fa26-4525-bca8-3449d4535409\") " pod="openstack/heat-cfnapi-9f9dd4fbb-4pcn7" Dec 06 06:18:11 crc kubenswrapper[4809]: I1206 06:18:11.998984 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2c1d3c1-fa26-4525-bca8-3449d4535409-public-tls-certs\") pod \"heat-cfnapi-9f9dd4fbb-4pcn7\" (UID: \"e2c1d3c1-fa26-4525-bca8-3449d4535409\") " pod="openstack/heat-cfnapi-9f9dd4fbb-4pcn7" Dec 06 06:18:11 crc kubenswrapper[4809]: I1206 06:18:11.999026 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d76c911-21ec-4f29-9707-e8101ed3aed9-config-data\") pod \"heat-api-bdfbd85-ntgxq\" (UID: \"4d76c911-21ec-4f29-9707-e8101ed3aed9\") " pod="openstack/heat-api-bdfbd85-ntgxq" Dec 06 06:18:11 crc kubenswrapper[4809]: I1206 06:18:11.999050 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4d76c911-21ec-4f29-9707-e8101ed3aed9-config-data-custom\") pod \"heat-api-bdfbd85-ntgxq\" (UID: \"4d76c911-21ec-4f29-9707-e8101ed3aed9\") " pod="openstack/heat-api-bdfbd85-ntgxq" Dec 06 06:18:11 crc kubenswrapper[4809]: I1206 06:18:11.999067 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d76c911-21ec-4f29-9707-e8101ed3aed9-public-tls-certs\") pod \"heat-api-bdfbd85-ntgxq\" (UID: \"4d76c911-21ec-4f29-9707-e8101ed3aed9\") " pod="openstack/heat-api-bdfbd85-ntgxq" Dec 06 06:18:11 crc kubenswrapper[4809]: I1206 06:18:11.999085 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d76c911-21ec-4f29-9707-e8101ed3aed9-combined-ca-bundle\") pod \"heat-api-bdfbd85-ntgxq\" (UID: \"4d76c911-21ec-4f29-9707-e8101ed3aed9\") " pod="openstack/heat-api-bdfbd85-ntgxq" Dec 06 06:18:11 crc kubenswrapper[4809]: I1206 06:18:11.999144 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2c1d3c1-fa26-4525-bca8-3449d4535409-combined-ca-bundle\") pod \"heat-cfnapi-9f9dd4fbb-4pcn7\" (UID: \"e2c1d3c1-fa26-4525-bca8-3449d4535409\") " pod="openstack/heat-cfnapi-9f9dd4fbb-4pcn7" Dec 06 06:18:11 crc kubenswrapper[4809]: I1206 06:18:11.999179 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2c1d3c1-fa26-4525-bca8-3449d4535409-internal-tls-certs\") pod \"heat-cfnapi-9f9dd4fbb-4pcn7\" (UID: \"e2c1d3c1-fa26-4525-bca8-3449d4535409\") " pod="openstack/heat-cfnapi-9f9dd4fbb-4pcn7" Dec 06 06:18:11 crc kubenswrapper[4809]: I1206 06:18:11.999197 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5z7dq\" (UniqueName: \"kubernetes.io/projected/4d76c911-21ec-4f29-9707-e8101ed3aed9-kube-api-access-5z7dq\") pod \"heat-api-bdfbd85-ntgxq\" (UID: \"4d76c911-21ec-4f29-9707-e8101ed3aed9\") " pod="openstack/heat-api-bdfbd85-ntgxq" Dec 06 06:18:12 crc kubenswrapper[4809]: I1206 06:18:12.006497 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d76c911-21ec-4f29-9707-e8101ed3aed9-combined-ca-bundle\") pod \"heat-api-bdfbd85-ntgxq\" (UID: \"4d76c911-21ec-4f29-9707-e8101ed3aed9\") " pod="openstack/heat-api-bdfbd85-ntgxq" Dec 06 06:18:12 crc kubenswrapper[4809]: I1206 06:18:12.010378 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4d76c911-21ec-4f29-9707-e8101ed3aed9-config-data-custom\") pod \"heat-api-bdfbd85-ntgxq\" (UID: \"4d76c911-21ec-4f29-9707-e8101ed3aed9\") " pod="openstack/heat-api-bdfbd85-ntgxq" Dec 06 06:18:12 crc kubenswrapper[4809]: I1206 06:18:12.013733 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e2c1d3c1-fa26-4525-bca8-3449d4535409-config-data-custom\") pod \"heat-cfnapi-9f9dd4fbb-4pcn7\" (UID: \"e2c1d3c1-fa26-4525-bca8-3449d4535409\") " pod="openstack/heat-cfnapi-9f9dd4fbb-4pcn7" Dec 06 06:18:12 crc kubenswrapper[4809]: I1206 06:18:12.014923 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d76c911-21ec-4f29-9707-e8101ed3aed9-public-tls-certs\") pod \"heat-api-bdfbd85-ntgxq\" (UID: \"4d76c911-21ec-4f29-9707-e8101ed3aed9\") " pod="openstack/heat-api-bdfbd85-ntgxq" Dec 06 06:18:12 crc kubenswrapper[4809]: I1206 06:18:12.015156 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2c1d3c1-fa26-4525-bca8-3449d4535409-internal-tls-certs\") pod \"heat-cfnapi-9f9dd4fbb-4pcn7\" (UID: \"e2c1d3c1-fa26-4525-bca8-3449d4535409\") " pod="openstack/heat-cfnapi-9f9dd4fbb-4pcn7" Dec 06 06:18:12 crc kubenswrapper[4809]: I1206 06:18:12.015697 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d76c911-21ec-4f29-9707-e8101ed3aed9-internal-tls-certs\") pod \"heat-api-bdfbd85-ntgxq\" (UID: \"4d76c911-21ec-4f29-9707-e8101ed3aed9\") " pod="openstack/heat-api-bdfbd85-ntgxq" Dec 06 06:18:12 crc kubenswrapper[4809]: I1206 06:18:12.015965 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2c1d3c1-fa26-4525-bca8-3449d4535409-combined-ca-bundle\") pod \"heat-cfnapi-9f9dd4fbb-4pcn7\" (UID: \"e2c1d3c1-fa26-4525-bca8-3449d4535409\") " pod="openstack/heat-cfnapi-9f9dd4fbb-4pcn7" Dec 06 06:18:12 crc kubenswrapper[4809]: I1206 06:18:12.016122 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d76c911-21ec-4f29-9707-e8101ed3aed9-config-data\") pod \"heat-api-bdfbd85-ntgxq\" (UID: \"4d76c911-21ec-4f29-9707-e8101ed3aed9\") " pod="openstack/heat-api-bdfbd85-ntgxq" Dec 06 06:18:12 crc kubenswrapper[4809]: I1206 06:18:12.016618 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2c1d3c1-fa26-4525-bca8-3449d4535409-config-data\") pod \"heat-cfnapi-9f9dd4fbb-4pcn7\" (UID: \"e2c1d3c1-fa26-4525-bca8-3449d4535409\") " pod="openstack/heat-cfnapi-9f9dd4fbb-4pcn7" Dec 06 06:18:12 crc kubenswrapper[4809]: I1206 06:18:12.018539 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2c1d3c1-fa26-4525-bca8-3449d4535409-public-tls-certs\") pod \"heat-cfnapi-9f9dd4fbb-4pcn7\" (UID: \"e2c1d3c1-fa26-4525-bca8-3449d4535409\") " pod="openstack/heat-cfnapi-9f9dd4fbb-4pcn7" Dec 06 06:18:12 crc kubenswrapper[4809]: I1206 06:18:12.019917 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grs95\" (UniqueName: \"kubernetes.io/projected/e2c1d3c1-fa26-4525-bca8-3449d4535409-kube-api-access-grs95\") pod \"heat-cfnapi-9f9dd4fbb-4pcn7\" (UID: \"e2c1d3c1-fa26-4525-bca8-3449d4535409\") " pod="openstack/heat-cfnapi-9f9dd4fbb-4pcn7" Dec 06 06:18:12 crc kubenswrapper[4809]: I1206 06:18:12.021948 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5z7dq\" (UniqueName: \"kubernetes.io/projected/4d76c911-21ec-4f29-9707-e8101ed3aed9-kube-api-access-5z7dq\") pod \"heat-api-bdfbd85-ntgxq\" (UID: \"4d76c911-21ec-4f29-9707-e8101ed3aed9\") " pod="openstack/heat-api-bdfbd85-ntgxq" Dec 06 06:18:12 crc kubenswrapper[4809]: I1206 06:18:12.105463 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-9f9dd4fbb-4pcn7" Dec 06 06:18:12 crc kubenswrapper[4809]: I1206 06:18:12.119679 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-bdfbd85-ntgxq" Dec 06 06:18:12 crc kubenswrapper[4809]: W1206 06:18:12.613244 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d76c911_21ec_4f29_9707_e8101ed3aed9.slice/crio-e77f2c6fca8113852b1528ba7eddcfbf74cc021cd24983b020fb02b1080a99e9 WatchSource:0}: Error finding container e77f2c6fca8113852b1528ba7eddcfbf74cc021cd24983b020fb02b1080a99e9: Status 404 returned error can't find the container with id e77f2c6fca8113852b1528ba7eddcfbf74cc021cd24983b020fb02b1080a99e9 Dec 06 06:18:12 crc kubenswrapper[4809]: I1206 06:18:12.618892 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-bdfbd85-ntgxq"] Dec 06 06:18:12 crc kubenswrapper[4809]: I1206 06:18:12.694497 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:18:12 crc kubenswrapper[4809]: I1206 06:18:12.731551 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-bdfbd85-ntgxq" event={"ID":"4d76c911-21ec-4f29-9707-e8101ed3aed9","Type":"ContainerStarted","Data":"e77f2c6fca8113852b1528ba7eddcfbf74cc021cd24983b020fb02b1080a99e9"} Dec 06 06:18:13 crc kubenswrapper[4809]: W1206 06:18:13.278035 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode2c1d3c1_fa26_4525_bca8_3449d4535409.slice/crio-ae065b96655f46ceab54eae0d579044734ee0202e9f32454371dc81dae893293 WatchSource:0}: Error finding container ae065b96655f46ceab54eae0d579044734ee0202e9f32454371dc81dae893293: Status 404 returned error can't find the container with id ae065b96655f46ceab54eae0d579044734ee0202e9f32454371dc81dae893293 Dec 06 06:18:13 crc kubenswrapper[4809]: I1206 06:18:13.278646 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-9f9dd4fbb-4pcn7"] Dec 06 06:18:13 crc kubenswrapper[4809]: I1206 06:18:13.761755 4809 generic.go:334] "Generic (PLEG): container finished" podID="0a5c1cb9-6796-43cd-8c09-68ade099a586" containerID="003cfca1ebe709bf7acf0d167058b9edafdc3284967d347cc711bffe010fa582" exitCode=1 Dec 06 06:18:13 crc kubenswrapper[4809]: I1206 06:18:13.761834 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-c48964f4f-ktxp6" event={"ID":"0a5c1cb9-6796-43cd-8c09-68ade099a586","Type":"ContainerDied","Data":"003cfca1ebe709bf7acf0d167058b9edafdc3284967d347cc711bffe010fa582"} Dec 06 06:18:13 crc kubenswrapper[4809]: I1206 06:18:13.762427 4809 scope.go:117] "RemoveContainer" containerID="003cfca1ebe709bf7acf0d167058b9edafdc3284967d347cc711bffe010fa582" Dec 06 06:18:13 crc kubenswrapper[4809]: I1206 06:18:13.778369 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"53615cf8-5ea5-4d28-a4e7-1491e2b33744","Type":"ContainerStarted","Data":"9dc198feff550848f012dfdf8ff0459b0a049a7da9c085bb6ac08abdb3e3c283"} Dec 06 06:18:13 crc kubenswrapper[4809]: I1206 06:18:13.794587 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7d5f5dd46-d69vn" event={"ID":"3eed0435-bed9-4713-a316-acf245fd8872","Type":"ContainerStarted","Data":"c40144d5c1e82b4670872a06349157ead7e0cc0a0a9e35d48d2825dbd81033a3"} Dec 06 06:18:13 crc kubenswrapper[4809]: I1206 06:18:13.794737 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-7d5f5dd46-d69vn" podUID="3eed0435-bed9-4713-a316-acf245fd8872" containerName="heat-cfnapi" containerID="cri-o://c40144d5c1e82b4670872a06349157ead7e0cc0a0a9e35d48d2825dbd81033a3" gracePeriod=60 Dec 06 06:18:13 crc kubenswrapper[4809]: I1206 06:18:13.795103 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-7d5f5dd46-d69vn" Dec 06 06:18:13 crc kubenswrapper[4809]: I1206 06:18:13.801168 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-9f9dd4fbb-4pcn7" event={"ID":"e2c1d3c1-fa26-4525-bca8-3449d4535409","Type":"ContainerStarted","Data":"ae065b96655f46ceab54eae0d579044734ee0202e9f32454371dc81dae893293"} Dec 06 06:18:13 crc kubenswrapper[4809]: I1206 06:18:13.812140 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-97f66c66d-jlz22" event={"ID":"85faa375-eb95-49b6-ba0b-967eeeebb2e8","Type":"ContainerStarted","Data":"d6aae92b638fa40b2193bcc2adaf0beb360877b71d2ee4b2d3d12d01a286f321"} Dec 06 06:18:13 crc kubenswrapper[4809]: I1206 06:18:13.828275 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-97f66c66d-jlz22" Dec 06 06:18:13 crc kubenswrapper[4809]: I1206 06:18:13.831491 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-d46db5664-p2pfk" event={"ID":"413fcc56-206e-402f-a6f9-a5f9a4cab527","Type":"ContainerStarted","Data":"74edf042562ca73e9cce347689aa9fd4799c4252d80a5e6cd21a53daa3db21bd"} Dec 06 06:18:13 crc kubenswrapper[4809]: I1206 06:18:13.831713 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-d46db5664-p2pfk" podUID="413fcc56-206e-402f-a6f9-a5f9a4cab527" containerName="heat-api" containerID="cri-o://74edf042562ca73e9cce347689aa9fd4799c4252d80a5e6cd21a53daa3db21bd" gracePeriod=60 Dec 06 06:18:13 crc kubenswrapper[4809]: I1206 06:18:13.831921 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-d46db5664-p2pfk" Dec 06 06:18:13 crc kubenswrapper[4809]: I1206 06:18:13.840382 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-5drs7" event={"ID":"0710c66a-2ed5-4bf6-8f12-e3754d122d08","Type":"ContainerStarted","Data":"c7206cbf87c3ef1a5a62aa4aa4e23858e953abea02fc17523eee8ffb3a20dcd2"} Dec 06 06:18:13 crc kubenswrapper[4809]: I1206 06:18:13.842076 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7756b9d78c-5drs7" Dec 06 06:18:13 crc kubenswrapper[4809]: I1206 06:18:13.848098 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-6d8ccd7c9f-nbxnr" event={"ID":"04d8a790-dedd-44c4-9b8f-093abb541e73","Type":"ContainerStarted","Data":"3178492e382f3c8585308c4ee39f1caea0bb67c6f97da76aea6237e1669dc416"} Dec 06 06:18:13 crc kubenswrapper[4809]: I1206 06:18:13.849035 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-6d8ccd7c9f-nbxnr" Dec 06 06:18:13 crc kubenswrapper[4809]: I1206 06:18:13.850408 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-bdfbd85-ntgxq" event={"ID":"4d76c911-21ec-4f29-9707-e8101ed3aed9","Type":"ContainerStarted","Data":"fc6a1d95e1293610b382f626e7c6ef59525c05b478a9c00131b4ad52b7b351a1"} Dec 06 06:18:13 crc kubenswrapper[4809]: I1206 06:18:13.852535 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-bdfbd85-ntgxq" Dec 06 06:18:13 crc kubenswrapper[4809]: I1206 06:18:13.904675 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-7d5f5dd46-d69vn" podStartSLOduration=9.669797461 podStartE2EDuration="12.904652853s" podCreationTimestamp="2025-12-06 06:18:01 +0000 UTC" firstStartedPulling="2025-12-06 06:18:07.225331044 +0000 UTC m=+1612.114313986" lastFinishedPulling="2025-12-06 06:18:10.460186436 +0000 UTC m=+1615.349169378" observedRunningTime="2025-12-06 06:18:13.857228553 +0000 UTC m=+1618.746211495" watchObservedRunningTime="2025-12-06 06:18:13.904652853 +0000 UTC m=+1618.793635795" Dec 06 06:18:13 crc kubenswrapper[4809]: I1206 06:18:13.940641 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-bdfbd85-ntgxq" podStartSLOduration=2.940621893 podStartE2EDuration="2.940621893s" podCreationTimestamp="2025-12-06 06:18:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:18:13.909726259 +0000 UTC m=+1618.798709211" watchObservedRunningTime="2025-12-06 06:18:13.940621893 +0000 UTC m=+1618.829604835" Dec 06 06:18:13 crc kubenswrapper[4809]: I1206 06:18:13.951951 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-6d8ccd7c9f-nbxnr" podStartSLOduration=4.951920797 podStartE2EDuration="4.951920797s" podCreationTimestamp="2025-12-06 06:18:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:18:13.937249222 +0000 UTC m=+1618.826232164" watchObservedRunningTime="2025-12-06 06:18:13.951920797 +0000 UTC m=+1618.840903739" Dec 06 06:18:13 crc kubenswrapper[4809]: I1206 06:18:13.970513 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-d46db5664-p2pfk" podStartSLOduration=9.709163303 podStartE2EDuration="12.970496019s" podCreationTimestamp="2025-12-06 06:18:01 +0000 UTC" firstStartedPulling="2025-12-06 06:18:07.13512226 +0000 UTC m=+1612.024105202" lastFinishedPulling="2025-12-06 06:18:10.396454976 +0000 UTC m=+1615.285437918" observedRunningTime="2025-12-06 06:18:13.955910235 +0000 UTC m=+1618.844893187" watchObservedRunningTime="2025-12-06 06:18:13.970496019 +0000 UTC m=+1618.859478961" Dec 06 06:18:13 crc kubenswrapper[4809]: I1206 06:18:13.980910 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7756b9d78c-5drs7" podStartSLOduration=12.980896189 podStartE2EDuration="12.980896189s" podCreationTimestamp="2025-12-06 06:18:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:18:13.975573546 +0000 UTC m=+1618.864556488" watchObservedRunningTime="2025-12-06 06:18:13.980896189 +0000 UTC m=+1618.869879131" Dec 06 06:18:13 crc kubenswrapper[4809]: I1206 06:18:13.996922 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-97f66c66d-jlz22" podStartSLOduration=4.996904771 podStartE2EDuration="4.996904771s" podCreationTimestamp="2025-12-06 06:18:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:18:13.990447237 +0000 UTC m=+1618.879430179" watchObservedRunningTime="2025-12-06 06:18:13.996904771 +0000 UTC m=+1618.885887713" Dec 06 06:18:14 crc kubenswrapper[4809]: I1206 06:18:14.880374 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-9f9dd4fbb-4pcn7" event={"ID":"e2c1d3c1-fa26-4525-bca8-3449d4535409","Type":"ContainerStarted","Data":"4502d90eeaeae3ae19d45c87f3b707c6d81d5e4600ae45736dbc55681e40ee56"} Dec 06 06:18:14 crc kubenswrapper[4809]: I1206 06:18:14.885697 4809 generic.go:334] "Generic (PLEG): container finished" podID="85faa375-eb95-49b6-ba0b-967eeeebb2e8" containerID="d6aae92b638fa40b2193bcc2adaf0beb360877b71d2ee4b2d3d12d01a286f321" exitCode=1 Dec 06 06:18:14 crc kubenswrapper[4809]: I1206 06:18:14.886260 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-97f66c66d-jlz22" event={"ID":"85faa375-eb95-49b6-ba0b-967eeeebb2e8","Type":"ContainerDied","Data":"d6aae92b638fa40b2193bcc2adaf0beb360877b71d2ee4b2d3d12d01a286f321"} Dec 06 06:18:14 crc kubenswrapper[4809]: I1206 06:18:14.886600 4809 scope.go:117] "RemoveContainer" containerID="d6aae92b638fa40b2193bcc2adaf0beb360877b71d2ee4b2d3d12d01a286f321" Dec 06 06:18:14 crc kubenswrapper[4809]: I1206 06:18:14.895682 4809 generic.go:334] "Generic (PLEG): container finished" podID="413fcc56-206e-402f-a6f9-a5f9a4cab527" containerID="74edf042562ca73e9cce347689aa9fd4799c4252d80a5e6cd21a53daa3db21bd" exitCode=0 Dec 06 06:18:14 crc kubenswrapper[4809]: I1206 06:18:14.895776 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-d46db5664-p2pfk" event={"ID":"413fcc56-206e-402f-a6f9-a5f9a4cab527","Type":"ContainerDied","Data":"74edf042562ca73e9cce347689aa9fd4799c4252d80a5e6cd21a53daa3db21bd"} Dec 06 06:18:14 crc kubenswrapper[4809]: I1206 06:18:14.905410 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-9f9dd4fbb-4pcn7" podStartSLOduration=3.9053874029999998 podStartE2EDuration="3.905387403s" podCreationTimestamp="2025-12-06 06:18:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:18:14.899745021 +0000 UTC m=+1619.788727963" watchObservedRunningTime="2025-12-06 06:18:14.905387403 +0000 UTC m=+1619.794370335" Dec 06 06:18:14 crc kubenswrapper[4809]: I1206 06:18:14.928005 4809 generic.go:334] "Generic (PLEG): container finished" podID="0a5c1cb9-6796-43cd-8c09-68ade099a586" containerID="1d0263157b9458dfb5dfb10941e0c8b0e1d1278bb04e62b0c7baa14522ce3dd1" exitCode=1 Dec 06 06:18:14 crc kubenswrapper[4809]: I1206 06:18:14.928065 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-c48964f4f-ktxp6" event={"ID":"0a5c1cb9-6796-43cd-8c09-68ade099a586","Type":"ContainerDied","Data":"1d0263157b9458dfb5dfb10941e0c8b0e1d1278bb04e62b0c7baa14522ce3dd1"} Dec 06 06:18:14 crc kubenswrapper[4809]: I1206 06:18:14.928102 4809 scope.go:117] "RemoveContainer" containerID="003cfca1ebe709bf7acf0d167058b9edafdc3284967d347cc711bffe010fa582" Dec 06 06:18:14 crc kubenswrapper[4809]: I1206 06:18:14.928810 4809 scope.go:117] "RemoveContainer" containerID="1d0263157b9458dfb5dfb10941e0c8b0e1d1278bb04e62b0c7baa14522ce3dd1" Dec 06 06:18:14 crc kubenswrapper[4809]: E1206 06:18:14.929133 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-c48964f4f-ktxp6_openstack(0a5c1cb9-6796-43cd-8c09-68ade099a586)\"" pod="openstack/heat-cfnapi-c48964f4f-ktxp6" podUID="0a5c1cb9-6796-43cd-8c09-68ade099a586" Dec 06 06:18:14 crc kubenswrapper[4809]: I1206 06:18:14.940183 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-api-97f66c66d-jlz22" Dec 06 06:18:14 crc kubenswrapper[4809]: I1206 06:18:14.943678 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"53615cf8-5ea5-4d28-a4e7-1491e2b33744","Type":"ContainerStarted","Data":"999d97cab57c383062b33ab0f30fd1a595134ddd3aaef5109e965c3b0cca4b63"} Dec 06 06:18:14 crc kubenswrapper[4809]: I1206 06:18:14.945881 4809 generic.go:334] "Generic (PLEG): container finished" podID="3eed0435-bed9-4713-a316-acf245fd8872" containerID="c40144d5c1e82b4670872a06349157ead7e0cc0a0a9e35d48d2825dbd81033a3" exitCode=0 Dec 06 06:18:14 crc kubenswrapper[4809]: I1206 06:18:14.947282 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7d5f5dd46-d69vn" event={"ID":"3eed0435-bed9-4713-a316-acf245fd8872","Type":"ContainerDied","Data":"c40144d5c1e82b4670872a06349157ead7e0cc0a0a9e35d48d2825dbd81033a3"} Dec 06 06:18:14 crc kubenswrapper[4809]: I1206 06:18:14.951756 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-cfnapi-c48964f4f-ktxp6" Dec 06 06:18:14 crc kubenswrapper[4809]: I1206 06:18:14.955301 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-c48964f4f-ktxp6" Dec 06 06:18:15 crc kubenswrapper[4809]: I1206 06:18:15.294969 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-d46db5664-p2pfk" Dec 06 06:18:15 crc kubenswrapper[4809]: I1206 06:18:15.365849 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7d5f5dd46-d69vn" Dec 06 06:18:15 crc kubenswrapper[4809]: I1206 06:18:15.435233 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/413fcc56-206e-402f-a6f9-a5f9a4cab527-config-data-custom\") pod \"413fcc56-206e-402f-a6f9-a5f9a4cab527\" (UID: \"413fcc56-206e-402f-a6f9-a5f9a4cab527\") " Dec 06 06:18:15 crc kubenswrapper[4809]: I1206 06:18:15.435284 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/413fcc56-206e-402f-a6f9-a5f9a4cab527-config-data\") pod \"413fcc56-206e-402f-a6f9-a5f9a4cab527\" (UID: \"413fcc56-206e-402f-a6f9-a5f9a4cab527\") " Dec 06 06:18:15 crc kubenswrapper[4809]: I1206 06:18:15.435311 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sgk4h\" (UniqueName: \"kubernetes.io/projected/413fcc56-206e-402f-a6f9-a5f9a4cab527-kube-api-access-sgk4h\") pod \"413fcc56-206e-402f-a6f9-a5f9a4cab527\" (UID: \"413fcc56-206e-402f-a6f9-a5f9a4cab527\") " Dec 06 06:18:15 crc kubenswrapper[4809]: I1206 06:18:15.435564 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/413fcc56-206e-402f-a6f9-a5f9a4cab527-combined-ca-bundle\") pod \"413fcc56-206e-402f-a6f9-a5f9a4cab527\" (UID: \"413fcc56-206e-402f-a6f9-a5f9a4cab527\") " Dec 06 06:18:15 crc kubenswrapper[4809]: I1206 06:18:15.451620 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/413fcc56-206e-402f-a6f9-a5f9a4cab527-kube-api-access-sgk4h" (OuterVolumeSpecName: "kube-api-access-sgk4h") pod "413fcc56-206e-402f-a6f9-a5f9a4cab527" (UID: "413fcc56-206e-402f-a6f9-a5f9a4cab527"). InnerVolumeSpecName "kube-api-access-sgk4h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:18:15 crc kubenswrapper[4809]: I1206 06:18:15.451668 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/413fcc56-206e-402f-a6f9-a5f9a4cab527-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "413fcc56-206e-402f-a6f9-a5f9a4cab527" (UID: "413fcc56-206e-402f-a6f9-a5f9a4cab527"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:18:15 crc kubenswrapper[4809]: I1206 06:18:15.511191 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/413fcc56-206e-402f-a6f9-a5f9a4cab527-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "413fcc56-206e-402f-a6f9-a5f9a4cab527" (UID: "413fcc56-206e-402f-a6f9-a5f9a4cab527"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:18:15 crc kubenswrapper[4809]: I1206 06:18:15.537097 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3eed0435-bed9-4713-a316-acf245fd8872-config-data\") pod \"3eed0435-bed9-4713-a316-acf245fd8872\" (UID: \"3eed0435-bed9-4713-a316-acf245fd8872\") " Dec 06 06:18:15 crc kubenswrapper[4809]: I1206 06:18:15.537190 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3eed0435-bed9-4713-a316-acf245fd8872-combined-ca-bundle\") pod \"3eed0435-bed9-4713-a316-acf245fd8872\" (UID: \"3eed0435-bed9-4713-a316-acf245fd8872\") " Dec 06 06:18:15 crc kubenswrapper[4809]: I1206 06:18:15.537304 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nmsrn\" (UniqueName: \"kubernetes.io/projected/3eed0435-bed9-4713-a316-acf245fd8872-kube-api-access-nmsrn\") pod \"3eed0435-bed9-4713-a316-acf245fd8872\" (UID: \"3eed0435-bed9-4713-a316-acf245fd8872\") " Dec 06 06:18:15 crc kubenswrapper[4809]: I1206 06:18:15.537338 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3eed0435-bed9-4713-a316-acf245fd8872-config-data-custom\") pod \"3eed0435-bed9-4713-a316-acf245fd8872\" (UID: \"3eed0435-bed9-4713-a316-acf245fd8872\") " Dec 06 06:18:15 crc kubenswrapper[4809]: I1206 06:18:15.537948 4809 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/413fcc56-206e-402f-a6f9-a5f9a4cab527-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:15 crc kubenswrapper[4809]: I1206 06:18:15.537965 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sgk4h\" (UniqueName: \"kubernetes.io/projected/413fcc56-206e-402f-a6f9-a5f9a4cab527-kube-api-access-sgk4h\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:15 crc kubenswrapper[4809]: I1206 06:18:15.537975 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/413fcc56-206e-402f-a6f9-a5f9a4cab527-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:15 crc kubenswrapper[4809]: I1206 06:18:15.556040 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3eed0435-bed9-4713-a316-acf245fd8872-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "3eed0435-bed9-4713-a316-acf245fd8872" (UID: "3eed0435-bed9-4713-a316-acf245fd8872"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:18:15 crc kubenswrapper[4809]: I1206 06:18:15.557400 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3eed0435-bed9-4713-a316-acf245fd8872-kube-api-access-nmsrn" (OuterVolumeSpecName: "kube-api-access-nmsrn") pod "3eed0435-bed9-4713-a316-acf245fd8872" (UID: "3eed0435-bed9-4713-a316-acf245fd8872"). InnerVolumeSpecName "kube-api-access-nmsrn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:18:15 crc kubenswrapper[4809]: I1206 06:18:15.603038 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/413fcc56-206e-402f-a6f9-a5f9a4cab527-config-data" (OuterVolumeSpecName: "config-data") pod "413fcc56-206e-402f-a6f9-a5f9a4cab527" (UID: "413fcc56-206e-402f-a6f9-a5f9a4cab527"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:18:15 crc kubenswrapper[4809]: I1206 06:18:15.627095 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3eed0435-bed9-4713-a316-acf245fd8872-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3eed0435-bed9-4713-a316-acf245fd8872" (UID: "3eed0435-bed9-4713-a316-acf245fd8872"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:18:15 crc kubenswrapper[4809]: I1206 06:18:15.640300 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3eed0435-bed9-4713-a316-acf245fd8872-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:15 crc kubenswrapper[4809]: I1206 06:18:15.640328 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nmsrn\" (UniqueName: \"kubernetes.io/projected/3eed0435-bed9-4713-a316-acf245fd8872-kube-api-access-nmsrn\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:15 crc kubenswrapper[4809]: I1206 06:18:15.640340 4809 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3eed0435-bed9-4713-a316-acf245fd8872-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:15 crc kubenswrapper[4809]: I1206 06:18:15.640353 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/413fcc56-206e-402f-a6f9-a5f9a4cab527-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:15 crc kubenswrapper[4809]: I1206 06:18:15.680826 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3eed0435-bed9-4713-a316-acf245fd8872-config-data" (OuterVolumeSpecName: "config-data") pod "3eed0435-bed9-4713-a316-acf245fd8872" (UID: "3eed0435-bed9-4713-a316-acf245fd8872"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:18:15 crc kubenswrapper[4809]: I1206 06:18:15.743324 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3eed0435-bed9-4713-a316-acf245fd8872-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:15 crc kubenswrapper[4809]: I1206 06:18:15.884895 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-bmh2p"] Dec 06 06:18:15 crc kubenswrapper[4809]: E1206 06:18:15.888263 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="413fcc56-206e-402f-a6f9-a5f9a4cab527" containerName="heat-api" Dec 06 06:18:15 crc kubenswrapper[4809]: I1206 06:18:15.888376 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="413fcc56-206e-402f-a6f9-a5f9a4cab527" containerName="heat-api" Dec 06 06:18:15 crc kubenswrapper[4809]: E1206 06:18:15.888449 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3eed0435-bed9-4713-a316-acf245fd8872" containerName="heat-cfnapi" Dec 06 06:18:15 crc kubenswrapper[4809]: I1206 06:18:15.888501 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3eed0435-bed9-4713-a316-acf245fd8872" containerName="heat-cfnapi" Dec 06 06:18:15 crc kubenswrapper[4809]: I1206 06:18:15.888797 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="413fcc56-206e-402f-a6f9-a5f9a4cab527" containerName="heat-api" Dec 06 06:18:15 crc kubenswrapper[4809]: I1206 06:18:15.888867 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="3eed0435-bed9-4713-a316-acf245fd8872" containerName="heat-cfnapi" Dec 06 06:18:15 crc kubenswrapper[4809]: I1206 06:18:15.889875 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-bmh2p" Dec 06 06:18:15 crc kubenswrapper[4809]: I1206 06:18:15.898896 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-bmh2p"] Dec 06 06:18:15 crc kubenswrapper[4809]: I1206 06:18:15.993919 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-sd9ds"] Dec 06 06:18:15 crc kubenswrapper[4809]: I1206 06:18:15.995868 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-sd9ds" Dec 06 06:18:15 crc kubenswrapper[4809]: I1206 06:18:15.997257 4809 scope.go:117] "RemoveContainer" containerID="1d0263157b9458dfb5dfb10941e0c8b0e1d1278bb04e62b0c7baa14522ce3dd1" Dec 06 06:18:15 crc kubenswrapper[4809]: E1206 06:18:15.997594 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-c48964f4f-ktxp6_openstack(0a5c1cb9-6796-43cd-8c09-68ade099a586)\"" pod="openstack/heat-cfnapi-c48964f4f-ktxp6" podUID="0a5c1cb9-6796-43cd-8c09-68ade099a586" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.034472 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"53615cf8-5ea5-4d28-a4e7-1491e2b33744","Type":"ContainerStarted","Data":"89b31f5e6af406027c0bbe907db9671a164bbfd100c01beee0a9a23b402a0893"} Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.039185 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-sd9ds"] Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.045109 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7d5f5dd46-d69vn" event={"ID":"3eed0435-bed9-4713-a316-acf245fd8872","Type":"ContainerDied","Data":"329e6470ae493f53d473445a5c2e6ea11a65543d179ed049883025c0ca38ebed"} Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.045325 4809 scope.go:117] "RemoveContainer" containerID="c40144d5c1e82b4670872a06349157ead7e0cc0a0a9e35d48d2825dbd81033a3" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.045604 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7d5f5dd46-d69vn" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.055776 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsgpz\" (UniqueName: \"kubernetes.io/projected/ee554177-cc9d-4969-9085-e507976f0f28-kube-api-access-hsgpz\") pod \"nova-api-db-create-bmh2p\" (UID: \"ee554177-cc9d-4969-9085-e507976f0f28\") " pod="openstack/nova-api-db-create-bmh2p" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.055858 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee554177-cc9d-4969-9085-e507976f0f28-operator-scripts\") pod \"nova-api-db-create-bmh2p\" (UID: \"ee554177-cc9d-4969-9085-e507976f0f28\") " pod="openstack/nova-api-db-create-bmh2p" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.083351 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-97f66c66d-jlz22" event={"ID":"85faa375-eb95-49b6-ba0b-967eeeebb2e8","Type":"ContainerStarted","Data":"c5a34db92db6dc08e7a02ec45c3ae41092245c47d6f77e2fde124fd9d8fbdc19"} Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.087177 4809 scope.go:117] "RemoveContainer" containerID="c5a34db92db6dc08e7a02ec45c3ae41092245c47d6f77e2fde124fd9d8fbdc19" Dec 06 06:18:16 crc kubenswrapper[4809]: E1206 06:18:16.087687 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-97f66c66d-jlz22_openstack(85faa375-eb95-49b6-ba0b-967eeeebb2e8)\"" pod="openstack/heat-api-97f66c66d-jlz22" podUID="85faa375-eb95-49b6-ba0b-967eeeebb2e8" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.088467 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-d46db5664-p2pfk" event={"ID":"413fcc56-206e-402f-a6f9-a5f9a4cab527","Type":"ContainerDied","Data":"625787070e1fbbb17d0d059c34c27c6c0301e9194d9c5dccffc36da08d3e462c"} Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.088514 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-9f9dd4fbb-4pcn7" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.088565 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-d46db5664-p2pfk" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.160776 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee554177-cc9d-4969-9085-e507976f0f28-operator-scripts\") pod \"nova-api-db-create-bmh2p\" (UID: \"ee554177-cc9d-4969-9085-e507976f0f28\") " pod="openstack/nova-api-db-create-bmh2p" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.167058 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7z7g\" (UniqueName: \"kubernetes.io/projected/c6b76b1b-c52d-464c-9b94-1e67d459f3b3-kube-api-access-l7z7g\") pod \"nova-cell0-db-create-sd9ds\" (UID: \"c6b76b1b-c52d-464c-9b94-1e67d459f3b3\") " pod="openstack/nova-cell0-db-create-sd9ds" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.167122 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsgpz\" (UniqueName: \"kubernetes.io/projected/ee554177-cc9d-4969-9085-e507976f0f28-kube-api-access-hsgpz\") pod \"nova-api-db-create-bmh2p\" (UID: \"ee554177-cc9d-4969-9085-e507976f0f28\") " pod="openstack/nova-api-db-create-bmh2p" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.167152 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c6b76b1b-c52d-464c-9b94-1e67d459f3b3-operator-scripts\") pod \"nova-cell0-db-create-sd9ds\" (UID: \"c6b76b1b-c52d-464c-9b94-1e67d459f3b3\") " pod="openstack/nova-cell0-db-create-sd9ds" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.163574 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-4733-account-create-update-z8gxx"] Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.168989 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-4733-account-create-update-z8gxx" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.166171 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee554177-cc9d-4969-9085-e507976f0f28-operator-scripts\") pod \"nova-api-db-create-bmh2p\" (UID: \"ee554177-cc9d-4969-9085-e507976f0f28\") " pod="openstack/nova-api-db-create-bmh2p" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.174047 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-4733-account-create-update-z8gxx"] Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.176860 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.191236 4809 scope.go:117] "RemoveContainer" containerID="74edf042562ca73e9cce347689aa9fd4799c4252d80a5e6cd21a53daa3db21bd" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.206908 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsgpz\" (UniqueName: \"kubernetes.io/projected/ee554177-cc9d-4969-9085-e507976f0f28-kube-api-access-hsgpz\") pod \"nova-api-db-create-bmh2p\" (UID: \"ee554177-cc9d-4969-9085-e507976f0f28\") " pod="openstack/nova-api-db-create-bmh2p" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.216041 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-bmh2p" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.258997 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-7d5f5dd46-d69vn"] Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.273214 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-7d5f5dd46-d69vn"] Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.274085 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7z7g\" (UniqueName: \"kubernetes.io/projected/c6b76b1b-c52d-464c-9b94-1e67d459f3b3-kube-api-access-l7z7g\") pod \"nova-cell0-db-create-sd9ds\" (UID: \"c6b76b1b-c52d-464c-9b94-1e67d459f3b3\") " pod="openstack/nova-cell0-db-create-sd9ds" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.274159 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c6b76b1b-c52d-464c-9b94-1e67d459f3b3-operator-scripts\") pod \"nova-cell0-db-create-sd9ds\" (UID: \"c6b76b1b-c52d-464c-9b94-1e67d459f3b3\") " pod="openstack/nova-cell0-db-create-sd9ds" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.275246 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c6b76b1b-c52d-464c-9b94-1e67d459f3b3-operator-scripts\") pod \"nova-cell0-db-create-sd9ds\" (UID: \"c6b76b1b-c52d-464c-9b94-1e67d459f3b3\") " pod="openstack/nova-cell0-db-create-sd9ds" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.289246 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-d46db5664-p2pfk"] Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.314500 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-d46db5664-p2pfk"] Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.348468 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7z7g\" (UniqueName: \"kubernetes.io/projected/c6b76b1b-c52d-464c-9b94-1e67d459f3b3-kube-api-access-l7z7g\") pod \"nova-cell0-db-create-sd9ds\" (UID: \"c6b76b1b-c52d-464c-9b94-1e67d459f3b3\") " pod="openstack/nova-cell0-db-create-sd9ds" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.352382 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-fb6zj"] Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.354383 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-fb6zj" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.388384 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-356a-account-create-update-6mtbv"] Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.390021 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-fb6zj"] Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.390100 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-356a-account-create-update-6mtbv" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.391133 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m84mp\" (UniqueName: \"kubernetes.io/projected/a0866cc1-8f06-4514-9939-5a4cfadc72b3-kube-api-access-m84mp\") pod \"nova-api-4733-account-create-update-z8gxx\" (UID: \"a0866cc1-8f06-4514-9939-5a4cfadc72b3\") " pod="openstack/nova-api-4733-account-create-update-z8gxx" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.391381 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a0866cc1-8f06-4514-9939-5a4cfadc72b3-operator-scripts\") pod \"nova-api-4733-account-create-update-z8gxx\" (UID: \"a0866cc1-8f06-4514-9939-5a4cfadc72b3\") " pod="openstack/nova-api-4733-account-create-update-z8gxx" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.392373 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.392574 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-sd9ds" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.410038 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-356a-account-create-update-6mtbv"] Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.497856 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m84mp\" (UniqueName: \"kubernetes.io/projected/a0866cc1-8f06-4514-9939-5a4cfadc72b3-kube-api-access-m84mp\") pod \"nova-api-4733-account-create-update-z8gxx\" (UID: \"a0866cc1-8f06-4514-9939-5a4cfadc72b3\") " pod="openstack/nova-api-4733-account-create-update-z8gxx" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.497975 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/412c5e46-e0e4-4b8a-af43-616019f9b619-operator-scripts\") pod \"nova-cell1-db-create-fb6zj\" (UID: \"412c5e46-e0e4-4b8a-af43-616019f9b619\") " pod="openstack/nova-cell1-db-create-fb6zj" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.498037 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5w7j\" (UniqueName: \"kubernetes.io/projected/5984ffba-4d1d-4976-b97f-606e2f557b03-kube-api-access-k5w7j\") pod \"nova-cell0-356a-account-create-update-6mtbv\" (UID: \"5984ffba-4d1d-4976-b97f-606e2f557b03\") " pod="openstack/nova-cell0-356a-account-create-update-6mtbv" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.498082 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5984ffba-4d1d-4976-b97f-606e2f557b03-operator-scripts\") pod \"nova-cell0-356a-account-create-update-6mtbv\" (UID: \"5984ffba-4d1d-4976-b97f-606e2f557b03\") " pod="openstack/nova-cell0-356a-account-create-update-6mtbv" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.498204 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrh89\" (UniqueName: \"kubernetes.io/projected/412c5e46-e0e4-4b8a-af43-616019f9b619-kube-api-access-xrh89\") pod \"nova-cell1-db-create-fb6zj\" (UID: \"412c5e46-e0e4-4b8a-af43-616019f9b619\") " pod="openstack/nova-cell1-db-create-fb6zj" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.498232 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a0866cc1-8f06-4514-9939-5a4cfadc72b3-operator-scripts\") pod \"nova-api-4733-account-create-update-z8gxx\" (UID: \"a0866cc1-8f06-4514-9939-5a4cfadc72b3\") " pod="openstack/nova-api-4733-account-create-update-z8gxx" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.499113 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a0866cc1-8f06-4514-9939-5a4cfadc72b3-operator-scripts\") pod \"nova-api-4733-account-create-update-z8gxx\" (UID: \"a0866cc1-8f06-4514-9939-5a4cfadc72b3\") " pod="openstack/nova-api-4733-account-create-update-z8gxx" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.556789 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m84mp\" (UniqueName: \"kubernetes.io/projected/a0866cc1-8f06-4514-9939-5a4cfadc72b3-kube-api-access-m84mp\") pod \"nova-api-4733-account-create-update-z8gxx\" (UID: \"a0866cc1-8f06-4514-9939-5a4cfadc72b3\") " pod="openstack/nova-api-4733-account-create-update-z8gxx" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.578591 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-8841-account-create-update-qhld2"] Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.583908 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-8841-account-create-update-qhld2" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.604870 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-8841-account-create-update-qhld2"] Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.616407 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27c8j\" (UniqueName: \"kubernetes.io/projected/7494a97b-56a3-45e8-a189-48eea155d3b3-kube-api-access-27c8j\") pod \"nova-cell1-8841-account-create-update-qhld2\" (UID: \"7494a97b-56a3-45e8-a189-48eea155d3b3\") " pod="openstack/nova-cell1-8841-account-create-update-qhld2" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.616540 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrh89\" (UniqueName: \"kubernetes.io/projected/412c5e46-e0e4-4b8a-af43-616019f9b619-kube-api-access-xrh89\") pod \"nova-cell1-db-create-fb6zj\" (UID: \"412c5e46-e0e4-4b8a-af43-616019f9b619\") " pod="openstack/nova-cell1-db-create-fb6zj" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.616802 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7494a97b-56a3-45e8-a189-48eea155d3b3-operator-scripts\") pod \"nova-cell1-8841-account-create-update-qhld2\" (UID: \"7494a97b-56a3-45e8-a189-48eea155d3b3\") " pod="openstack/nova-cell1-8841-account-create-update-qhld2" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.616916 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/412c5e46-e0e4-4b8a-af43-616019f9b619-operator-scripts\") pod \"nova-cell1-db-create-fb6zj\" (UID: \"412c5e46-e0e4-4b8a-af43-616019f9b619\") " pod="openstack/nova-cell1-db-create-fb6zj" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.616991 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5w7j\" (UniqueName: \"kubernetes.io/projected/5984ffba-4d1d-4976-b97f-606e2f557b03-kube-api-access-k5w7j\") pod \"nova-cell0-356a-account-create-update-6mtbv\" (UID: \"5984ffba-4d1d-4976-b97f-606e2f557b03\") " pod="openstack/nova-cell0-356a-account-create-update-6mtbv" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.617082 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5984ffba-4d1d-4976-b97f-606e2f557b03-operator-scripts\") pod \"nova-cell0-356a-account-create-update-6mtbv\" (UID: \"5984ffba-4d1d-4976-b97f-606e2f557b03\") " pod="openstack/nova-cell0-356a-account-create-update-6mtbv" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.624138 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5984ffba-4d1d-4976-b97f-606e2f557b03-operator-scripts\") pod \"nova-cell0-356a-account-create-update-6mtbv\" (UID: \"5984ffba-4d1d-4976-b97f-606e2f557b03\") " pod="openstack/nova-cell0-356a-account-create-update-6mtbv" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.625419 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/412c5e46-e0e4-4b8a-af43-616019f9b619-operator-scripts\") pod \"nova-cell1-db-create-fb6zj\" (UID: \"412c5e46-e0e4-4b8a-af43-616019f9b619\") " pod="openstack/nova-cell1-db-create-fb6zj" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.680729 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.690908 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrh89\" (UniqueName: \"kubernetes.io/projected/412c5e46-e0e4-4b8a-af43-616019f9b619-kube-api-access-xrh89\") pod \"nova-cell1-db-create-fb6zj\" (UID: \"412c5e46-e0e4-4b8a-af43-616019f9b619\") " pod="openstack/nova-cell1-db-create-fb6zj" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.719761 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27c8j\" (UniqueName: \"kubernetes.io/projected/7494a97b-56a3-45e8-a189-48eea155d3b3-kube-api-access-27c8j\") pod \"nova-cell1-8841-account-create-update-qhld2\" (UID: \"7494a97b-56a3-45e8-a189-48eea155d3b3\") " pod="openstack/nova-cell1-8841-account-create-update-qhld2" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.719951 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7494a97b-56a3-45e8-a189-48eea155d3b3-operator-scripts\") pod \"nova-cell1-8841-account-create-update-qhld2\" (UID: \"7494a97b-56a3-45e8-a189-48eea155d3b3\") " pod="openstack/nova-cell1-8841-account-create-update-qhld2" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.720634 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7494a97b-56a3-45e8-a189-48eea155d3b3-operator-scripts\") pod \"nova-cell1-8841-account-create-update-qhld2\" (UID: \"7494a97b-56a3-45e8-a189-48eea155d3b3\") " pod="openstack/nova-cell1-8841-account-create-update-qhld2" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.691721 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5w7j\" (UniqueName: \"kubernetes.io/projected/5984ffba-4d1d-4976-b97f-606e2f557b03-kube-api-access-k5w7j\") pod \"nova-cell0-356a-account-create-update-6mtbv\" (UID: \"5984ffba-4d1d-4976-b97f-606e2f557b03\") " pod="openstack/nova-cell0-356a-account-create-update-6mtbv" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.751349 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27c8j\" (UniqueName: \"kubernetes.io/projected/7494a97b-56a3-45e8-a189-48eea155d3b3-kube-api-access-27c8j\") pod \"nova-cell1-8841-account-create-update-qhld2\" (UID: \"7494a97b-56a3-45e8-a189-48eea155d3b3\") " pod="openstack/nova-cell1-8841-account-create-update-qhld2" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.801294 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-4733-account-create-update-z8gxx" Dec 06 06:18:16 crc kubenswrapper[4809]: I1206 06:18:16.889191 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-fb6zj" Dec 06 06:18:17 crc kubenswrapper[4809]: I1206 06:18:17.000379 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-356a-account-create-update-6mtbv" Dec 06 06:18:17 crc kubenswrapper[4809]: I1206 06:18:17.014828 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-8841-account-create-update-qhld2" Dec 06 06:18:17 crc kubenswrapper[4809]: I1206 06:18:17.126462 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-bmh2p"] Dec 06 06:18:17 crc kubenswrapper[4809]: I1206 06:18:17.143684 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-sd9ds"] Dec 06 06:18:17 crc kubenswrapper[4809]: W1206 06:18:17.159875 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee554177_cc9d_4969_9085_e507976f0f28.slice/crio-f02ec91ea32b5d9b285dcbe3f6d0de2d62c18859d1b7e39b834360bfa1af6194 WatchSource:0}: Error finding container f02ec91ea32b5d9b285dcbe3f6d0de2d62c18859d1b7e39b834360bfa1af6194: Status 404 returned error can't find the container with id f02ec91ea32b5d9b285dcbe3f6d0de2d62c18859d1b7e39b834360bfa1af6194 Dec 06 06:18:17 crc kubenswrapper[4809]: I1206 06:18:17.174432 4809 generic.go:334] "Generic (PLEG): container finished" podID="85faa375-eb95-49b6-ba0b-967eeeebb2e8" containerID="c5a34db92db6dc08e7a02ec45c3ae41092245c47d6f77e2fde124fd9d8fbdc19" exitCode=1 Dec 06 06:18:17 crc kubenswrapper[4809]: I1206 06:18:17.174560 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-97f66c66d-jlz22" event={"ID":"85faa375-eb95-49b6-ba0b-967eeeebb2e8","Type":"ContainerDied","Data":"c5a34db92db6dc08e7a02ec45c3ae41092245c47d6f77e2fde124fd9d8fbdc19"} Dec 06 06:18:17 crc kubenswrapper[4809]: I1206 06:18:17.174601 4809 scope.go:117] "RemoveContainer" containerID="d6aae92b638fa40b2193bcc2adaf0beb360877b71d2ee4b2d3d12d01a286f321" Dec 06 06:18:17 crc kubenswrapper[4809]: I1206 06:18:17.175385 4809 scope.go:117] "RemoveContainer" containerID="c5a34db92db6dc08e7a02ec45c3ae41092245c47d6f77e2fde124fd9d8fbdc19" Dec 06 06:18:17 crc kubenswrapper[4809]: E1206 06:18:17.175713 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-97f66c66d-jlz22_openstack(85faa375-eb95-49b6-ba0b-967eeeebb2e8)\"" pod="openstack/heat-api-97f66c66d-jlz22" podUID="85faa375-eb95-49b6-ba0b-967eeeebb2e8" Dec 06 06:18:17 crc kubenswrapper[4809]: I1206 06:18:17.207327 4809 scope.go:117] "RemoveContainer" containerID="1d0263157b9458dfb5dfb10941e0c8b0e1d1278bb04e62b0c7baa14522ce3dd1" Dec 06 06:18:17 crc kubenswrapper[4809]: E1206 06:18:17.207537 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-c48964f4f-ktxp6_openstack(0a5c1cb9-6796-43cd-8c09-68ade099a586)\"" pod="openstack/heat-cfnapi-c48964f4f-ktxp6" podUID="0a5c1cb9-6796-43cd-8c09-68ade099a586" Dec 06 06:18:17 crc kubenswrapper[4809]: I1206 06:18:17.240083 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7756b9d78c-5drs7" Dec 06 06:18:17 crc kubenswrapper[4809]: I1206 06:18:17.356135 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-6dldb"] Dec 06 06:18:17 crc kubenswrapper[4809]: I1206 06:18:17.356336 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9776ccc5-6dldb" podUID="8757fd43-0a69-40c8-8832-53a4acd0a912" containerName="dnsmasq-dns" containerID="cri-o://f8f0dee995d02f998c3002eff13de0219a2a4ebafa905444e8aed7037afa2ce0" gracePeriod=10 Dec 06 06:18:17 crc kubenswrapper[4809]: I1206 06:18:17.533499 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3eed0435-bed9-4713-a316-acf245fd8872" path="/var/lib/kubelet/pods/3eed0435-bed9-4713-a316-acf245fd8872/volumes" Dec 06 06:18:17 crc kubenswrapper[4809]: I1206 06:18:17.534253 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="413fcc56-206e-402f-a6f9-a5f9a4cab527" path="/var/lib/kubelet/pods/413fcc56-206e-402f-a6f9-a5f9a4cab527/volumes" Dec 06 06:18:17 crc kubenswrapper[4809]: I1206 06:18:17.877615 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-4733-account-create-update-z8gxx"] Dec 06 06:18:17 crc kubenswrapper[4809]: W1206 06:18:17.919030 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda0866cc1_8f06_4514_9939_5a4cfadc72b3.slice/crio-70f02532c7aea47b6ab049074dcd5714ad04c59ef361d6e80dee6c286a441fd2 WatchSource:0}: Error finding container 70f02532c7aea47b6ab049074dcd5714ad04c59ef361d6e80dee6c286a441fd2: Status 404 returned error can't find the container with id 70f02532c7aea47b6ab049074dcd5714ad04c59ef361d6e80dee6c286a441fd2 Dec 06 06:18:18 crc kubenswrapper[4809]: I1206 06:18:18.226620 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-fb6zj"] Dec 06 06:18:18 crc kubenswrapper[4809]: I1206 06:18:18.343215 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"53615cf8-5ea5-4d28-a4e7-1491e2b33744","Type":"ContainerStarted","Data":"469ecb481d66fc349315f13547379057606da423f7767718d0b6694ad4cf95a3"} Dec 06 06:18:18 crc kubenswrapper[4809]: I1206 06:18:18.343407 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="53615cf8-5ea5-4d28-a4e7-1491e2b33744" containerName="ceilometer-central-agent" containerID="cri-o://9dc198feff550848f012dfdf8ff0459b0a049a7da9c085bb6ac08abdb3e3c283" gracePeriod=30 Dec 06 06:18:18 crc kubenswrapper[4809]: I1206 06:18:18.343640 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 06:18:18 crc kubenswrapper[4809]: I1206 06:18:18.343964 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="53615cf8-5ea5-4d28-a4e7-1491e2b33744" containerName="proxy-httpd" containerID="cri-o://469ecb481d66fc349315f13547379057606da423f7767718d0b6694ad4cf95a3" gracePeriod=30 Dec 06 06:18:18 crc kubenswrapper[4809]: I1206 06:18:18.344018 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="53615cf8-5ea5-4d28-a4e7-1491e2b33744" containerName="sg-core" containerID="cri-o://89b31f5e6af406027c0bbe907db9671a164bbfd100c01beee0a9a23b402a0893" gracePeriod=30 Dec 06 06:18:18 crc kubenswrapper[4809]: I1206 06:18:18.344057 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="53615cf8-5ea5-4d28-a4e7-1491e2b33744" containerName="ceilometer-notification-agent" containerID="cri-o://999d97cab57c383062b33ab0f30fd1a595134ddd3aaef5109e965c3b0cca4b63" gracePeriod=30 Dec 06 06:18:18 crc kubenswrapper[4809]: I1206 06:18:18.392903 4809 generic.go:334] "Generic (PLEG): container finished" podID="8757fd43-0a69-40c8-8832-53a4acd0a912" containerID="f8f0dee995d02f998c3002eff13de0219a2a4ebafa905444e8aed7037afa2ce0" exitCode=0 Dec 06 06:18:18 crc kubenswrapper[4809]: I1206 06:18:18.393047 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-6dldb" event={"ID":"8757fd43-0a69-40c8-8832-53a4acd0a912","Type":"ContainerDied","Data":"f8f0dee995d02f998c3002eff13de0219a2a4ebafa905444e8aed7037afa2ce0"} Dec 06 06:18:18 crc kubenswrapper[4809]: I1206 06:18:18.422299 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=4.577722918 podStartE2EDuration="11.422284905s" podCreationTimestamp="2025-12-06 06:18:07 +0000 UTC" firstStartedPulling="2025-12-06 06:18:10.224162197 +0000 UTC m=+1615.113145139" lastFinishedPulling="2025-12-06 06:18:17.068724184 +0000 UTC m=+1621.957707126" observedRunningTime="2025-12-06 06:18:18.393425807 +0000 UTC m=+1623.282408749" watchObservedRunningTime="2025-12-06 06:18:18.422284905 +0000 UTC m=+1623.311267847" Dec 06 06:18:18 crc kubenswrapper[4809]: I1206 06:18:18.457905 4809 scope.go:117] "RemoveContainer" containerID="c5a34db92db6dc08e7a02ec45c3ae41092245c47d6f77e2fde124fd9d8fbdc19" Dec 06 06:18:18 crc kubenswrapper[4809]: E1206 06:18:18.458472 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-97f66c66d-jlz22_openstack(85faa375-eb95-49b6-ba0b-967eeeebb2e8)\"" pod="openstack/heat-api-97f66c66d-jlz22" podUID="85faa375-eb95-49b6-ba0b-967eeeebb2e8" Dec 06 06:18:18 crc kubenswrapper[4809]: I1206 06:18:18.464469 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-bmh2p" event={"ID":"ee554177-cc9d-4969-9085-e507976f0f28","Type":"ContainerStarted","Data":"308715c3c15148bf3a30aa4c43e68ba387da20816fc7ecf94ae2b946ccd0e9ac"} Dec 06 06:18:18 crc kubenswrapper[4809]: I1206 06:18:18.464508 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-bmh2p" event={"ID":"ee554177-cc9d-4969-9085-e507976f0f28","Type":"ContainerStarted","Data":"f02ec91ea32b5d9b285dcbe3f6d0de2d62c18859d1b7e39b834360bfa1af6194"} Dec 06 06:18:18 crc kubenswrapper[4809]: I1206 06:18:18.468563 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-fb6zj" event={"ID":"412c5e46-e0e4-4b8a-af43-616019f9b619","Type":"ContainerStarted","Data":"87701fbc106da4b39116db9ecfe6b1722842de90a9984d0a9c462d7876159d2b"} Dec 06 06:18:18 crc kubenswrapper[4809]: I1206 06:18:18.471506 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-4733-account-create-update-z8gxx" event={"ID":"a0866cc1-8f06-4514-9939-5a4cfadc72b3","Type":"ContainerStarted","Data":"70f02532c7aea47b6ab049074dcd5714ad04c59ef361d6e80dee6c286a441fd2"} Dec 06 06:18:18 crc kubenswrapper[4809]: I1206 06:18:18.480684 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-sd9ds" event={"ID":"c6b76b1b-c52d-464c-9b94-1e67d459f3b3","Type":"ContainerStarted","Data":"e77b147ecc98479207fbf89e2b41c04292a66b4172aced62be7f5170feb90686"} Dec 06 06:18:18 crc kubenswrapper[4809]: I1206 06:18:18.480752 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-sd9ds" event={"ID":"c6b76b1b-c52d-464c-9b94-1e67d459f3b3","Type":"ContainerStarted","Data":"81675f90ca51ed353720a1d409e626e14d925c61e319c27fbbeef8681b71ccd8"} Dec 06 06:18:18 crc kubenswrapper[4809]: I1206 06:18:18.532736 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-356a-account-create-update-6mtbv"] Dec 06 06:18:18 crc kubenswrapper[4809]: I1206 06:18:18.619382 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-4733-account-create-update-z8gxx" podStartSLOduration=2.619363183 podStartE2EDuration="2.619363183s" podCreationTimestamp="2025-12-06 06:18:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:18:18.525767517 +0000 UTC m=+1623.414750459" watchObservedRunningTime="2025-12-06 06:18:18.619363183 +0000 UTC m=+1623.508346125" Dec 06 06:18:18 crc kubenswrapper[4809]: I1206 06:18:18.674338 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-6dldb" Dec 06 06:18:18 crc kubenswrapper[4809]: I1206 06:18:18.722700 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-8841-account-create-update-qhld2"] Dec 06 06:18:18 crc kubenswrapper[4809]: I1206 06:18:18.834270 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbv45\" (UniqueName: \"kubernetes.io/projected/8757fd43-0a69-40c8-8832-53a4acd0a912-kube-api-access-bbv45\") pod \"8757fd43-0a69-40c8-8832-53a4acd0a912\" (UID: \"8757fd43-0a69-40c8-8832-53a4acd0a912\") " Dec 06 06:18:18 crc kubenswrapper[4809]: I1206 06:18:18.838715 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8757fd43-0a69-40c8-8832-53a4acd0a912-config\") pod \"8757fd43-0a69-40c8-8832-53a4acd0a912\" (UID: \"8757fd43-0a69-40c8-8832-53a4acd0a912\") " Dec 06 06:18:18 crc kubenswrapper[4809]: I1206 06:18:18.852173 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8757fd43-0a69-40c8-8832-53a4acd0a912-ovsdbserver-nb\") pod \"8757fd43-0a69-40c8-8832-53a4acd0a912\" (UID: \"8757fd43-0a69-40c8-8832-53a4acd0a912\") " Dec 06 06:18:18 crc kubenswrapper[4809]: I1206 06:18:18.852359 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8757fd43-0a69-40c8-8832-53a4acd0a912-dns-swift-storage-0\") pod \"8757fd43-0a69-40c8-8832-53a4acd0a912\" (UID: \"8757fd43-0a69-40c8-8832-53a4acd0a912\") " Dec 06 06:18:18 crc kubenswrapper[4809]: I1206 06:18:18.852458 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8757fd43-0a69-40c8-8832-53a4acd0a912-ovsdbserver-sb\") pod \"8757fd43-0a69-40c8-8832-53a4acd0a912\" (UID: \"8757fd43-0a69-40c8-8832-53a4acd0a912\") " Dec 06 06:18:18 crc kubenswrapper[4809]: I1206 06:18:18.852622 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8757fd43-0a69-40c8-8832-53a4acd0a912-dns-svc\") pod \"8757fd43-0a69-40c8-8832-53a4acd0a912\" (UID: \"8757fd43-0a69-40c8-8832-53a4acd0a912\") " Dec 06 06:18:18 crc kubenswrapper[4809]: I1206 06:18:18.847147 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8757fd43-0a69-40c8-8832-53a4acd0a912-kube-api-access-bbv45" (OuterVolumeSpecName: "kube-api-access-bbv45") pod "8757fd43-0a69-40c8-8832-53a4acd0a912" (UID: "8757fd43-0a69-40c8-8832-53a4acd0a912"). InnerVolumeSpecName "kube-api-access-bbv45". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:18:18 crc kubenswrapper[4809]: I1206 06:18:18.864856 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbv45\" (UniqueName: \"kubernetes.io/projected/8757fd43-0a69-40c8-8832-53a4acd0a912-kube-api-access-bbv45\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:18 crc kubenswrapper[4809]: I1206 06:18:18.942879 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8757fd43-0a69-40c8-8832-53a4acd0a912-config" (OuterVolumeSpecName: "config") pod "8757fd43-0a69-40c8-8832-53a4acd0a912" (UID: "8757fd43-0a69-40c8-8832-53a4acd0a912"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:18:18 crc kubenswrapper[4809]: I1206 06:18:18.975362 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8757fd43-0a69-40c8-8832-53a4acd0a912-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:18 crc kubenswrapper[4809]: I1206 06:18:18.989564 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8757fd43-0a69-40c8-8832-53a4acd0a912-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8757fd43-0a69-40c8-8832-53a4acd0a912" (UID: "8757fd43-0a69-40c8-8832-53a4acd0a912"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:18:19 crc kubenswrapper[4809]: I1206 06:18:19.013147 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8757fd43-0a69-40c8-8832-53a4acd0a912-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8757fd43-0a69-40c8-8832-53a4acd0a912" (UID: "8757fd43-0a69-40c8-8832-53a4acd0a912"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:18:19 crc kubenswrapper[4809]: I1206 06:18:19.028319 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8757fd43-0a69-40c8-8832-53a4acd0a912-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8757fd43-0a69-40c8-8832-53a4acd0a912" (UID: "8757fd43-0a69-40c8-8832-53a4acd0a912"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:18:19 crc kubenswrapper[4809]: I1206 06:18:19.080624 4809 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8757fd43-0a69-40c8-8832-53a4acd0a912-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:19 crc kubenswrapper[4809]: I1206 06:18:19.085548 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8757fd43-0a69-40c8-8832-53a4acd0a912-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:19 crc kubenswrapper[4809]: I1206 06:18:19.085780 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8757fd43-0a69-40c8-8832-53a4acd0a912-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:19 crc kubenswrapper[4809]: I1206 06:18:19.084456 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8757fd43-0a69-40c8-8832-53a4acd0a912-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8757fd43-0a69-40c8-8832-53a4acd0a912" (UID: "8757fd43-0a69-40c8-8832-53a4acd0a912"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:18:19 crc kubenswrapper[4809]: I1206 06:18:19.188445 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8757fd43-0a69-40c8-8832-53a4acd0a912-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:19 crc kubenswrapper[4809]: I1206 06:18:19.495765 4809 generic.go:334] "Generic (PLEG): container finished" podID="412c5e46-e0e4-4b8a-af43-616019f9b619" containerID="f8ba6acd951144b6cbf9feccba3d35a19374d54fcead893ca299bd97363407f4" exitCode=0 Dec 06 06:18:19 crc kubenswrapper[4809]: I1206 06:18:19.496046 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-fb6zj" event={"ID":"412c5e46-e0e4-4b8a-af43-616019f9b619","Type":"ContainerDied","Data":"f8ba6acd951144b6cbf9feccba3d35a19374d54fcead893ca299bd97363407f4"} Dec 06 06:18:19 crc kubenswrapper[4809]: I1206 06:18:19.498088 4809 generic.go:334] "Generic (PLEG): container finished" podID="ee554177-cc9d-4969-9085-e507976f0f28" containerID="308715c3c15148bf3a30aa4c43e68ba387da20816fc7ecf94ae2b946ccd0e9ac" exitCode=0 Dec 06 06:18:19 crc kubenswrapper[4809]: I1206 06:18:19.498148 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-bmh2p" event={"ID":"ee554177-cc9d-4969-9085-e507976f0f28","Type":"ContainerDied","Data":"308715c3c15148bf3a30aa4c43e68ba387da20816fc7ecf94ae2b946ccd0e9ac"} Dec 06 06:18:19 crc kubenswrapper[4809]: I1206 06:18:19.501438 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-6dldb" event={"ID":"8757fd43-0a69-40c8-8832-53a4acd0a912","Type":"ContainerDied","Data":"1d95a975f54748220b7d00d032185c87a19502fb67e16773b25d9f5beee93b5e"} Dec 06 06:18:19 crc kubenswrapper[4809]: I1206 06:18:19.501486 4809 scope.go:117] "RemoveContainer" containerID="f8f0dee995d02f998c3002eff13de0219a2a4ebafa905444e8aed7037afa2ce0" Dec 06 06:18:19 crc kubenswrapper[4809]: I1206 06:18:19.501657 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-6dldb" Dec 06 06:18:19 crc kubenswrapper[4809]: I1206 06:18:19.513450 4809 generic.go:334] "Generic (PLEG): container finished" podID="5984ffba-4d1d-4976-b97f-606e2f557b03" containerID="b30146671b82ca2efc77dadcc395a46bc9f67d0a5d0f6683eecfa651518e4103" exitCode=0 Dec 06 06:18:19 crc kubenswrapper[4809]: I1206 06:18:19.513530 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-356a-account-create-update-6mtbv" event={"ID":"5984ffba-4d1d-4976-b97f-606e2f557b03","Type":"ContainerDied","Data":"b30146671b82ca2efc77dadcc395a46bc9f67d0a5d0f6683eecfa651518e4103"} Dec 06 06:18:19 crc kubenswrapper[4809]: I1206 06:18:19.513555 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-356a-account-create-update-6mtbv" event={"ID":"5984ffba-4d1d-4976-b97f-606e2f557b03","Type":"ContainerStarted","Data":"f0b7fcb089c3953351ae89931db079663fe4ea684f87212623a42e73c074a99a"} Dec 06 06:18:19 crc kubenswrapper[4809]: I1206 06:18:19.550412 4809 generic.go:334] "Generic (PLEG): container finished" podID="a0866cc1-8f06-4514-9939-5a4cfadc72b3" containerID="fd7f07a5ac3909a5a15b8e1fdc9686616ea3e3cc8fbe64e0651e6f5fd586ebfe" exitCode=0 Dec 06 06:18:19 crc kubenswrapper[4809]: I1206 06:18:19.550499 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-4733-account-create-update-z8gxx" event={"ID":"a0866cc1-8f06-4514-9939-5a4cfadc72b3","Type":"ContainerDied","Data":"fd7f07a5ac3909a5a15b8e1fdc9686616ea3e3cc8fbe64e0651e6f5fd586ebfe"} Dec 06 06:18:19 crc kubenswrapper[4809]: I1206 06:18:19.564278 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-6dldb"] Dec 06 06:18:19 crc kubenswrapper[4809]: I1206 06:18:19.576002 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-6dldb"] Dec 06 06:18:19 crc kubenswrapper[4809]: I1206 06:18:19.587181 4809 generic.go:334] "Generic (PLEG): container finished" podID="53615cf8-5ea5-4d28-a4e7-1491e2b33744" containerID="469ecb481d66fc349315f13547379057606da423f7767718d0b6694ad4cf95a3" exitCode=0 Dec 06 06:18:19 crc kubenswrapper[4809]: I1206 06:18:19.587239 4809 generic.go:334] "Generic (PLEG): container finished" podID="53615cf8-5ea5-4d28-a4e7-1491e2b33744" containerID="89b31f5e6af406027c0bbe907db9671a164bbfd100c01beee0a9a23b402a0893" exitCode=2 Dec 06 06:18:19 crc kubenswrapper[4809]: I1206 06:18:19.587251 4809 generic.go:334] "Generic (PLEG): container finished" podID="53615cf8-5ea5-4d28-a4e7-1491e2b33744" containerID="999d97cab57c383062b33ab0f30fd1a595134ddd3aaef5109e965c3b0cca4b63" exitCode=0 Dec 06 06:18:19 crc kubenswrapper[4809]: I1206 06:18:19.587343 4809 scope.go:117] "RemoveContainer" containerID="c7b2a190e462fabd5f9d06e8ba8b95313e9ade5bf7e92b8ed03a2ecab9950af2" Dec 06 06:18:19 crc kubenswrapper[4809]: I1206 06:18:19.587460 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"53615cf8-5ea5-4d28-a4e7-1491e2b33744","Type":"ContainerDied","Data":"469ecb481d66fc349315f13547379057606da423f7767718d0b6694ad4cf95a3"} Dec 06 06:18:19 crc kubenswrapper[4809]: I1206 06:18:19.587484 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"53615cf8-5ea5-4d28-a4e7-1491e2b33744","Type":"ContainerDied","Data":"89b31f5e6af406027c0bbe907db9671a164bbfd100c01beee0a9a23b402a0893"} Dec 06 06:18:19 crc kubenswrapper[4809]: I1206 06:18:19.587497 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"53615cf8-5ea5-4d28-a4e7-1491e2b33744","Type":"ContainerDied","Data":"999d97cab57c383062b33ab0f30fd1a595134ddd3aaef5109e965c3b0cca4b63"} Dec 06 06:18:19 crc kubenswrapper[4809]: I1206 06:18:19.618418 4809 generic.go:334] "Generic (PLEG): container finished" podID="c6b76b1b-c52d-464c-9b94-1e67d459f3b3" containerID="e77b147ecc98479207fbf89e2b41c04292a66b4172aced62be7f5170feb90686" exitCode=0 Dec 06 06:18:19 crc kubenswrapper[4809]: I1206 06:18:19.618630 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-sd9ds" event={"ID":"c6b76b1b-c52d-464c-9b94-1e67d459f3b3","Type":"ContainerDied","Data":"e77b147ecc98479207fbf89e2b41c04292a66b4172aced62be7f5170feb90686"} Dec 06 06:18:19 crc kubenswrapper[4809]: I1206 06:18:19.622271 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-8841-account-create-update-qhld2" event={"ID":"7494a97b-56a3-45e8-a189-48eea155d3b3","Type":"ContainerStarted","Data":"fa80f87dee5acf1e4481c0a5cb3a2046e006e209b4ecea87d99c9b30c9815b13"} Dec 06 06:18:19 crc kubenswrapper[4809]: I1206 06:18:19.622314 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-8841-account-create-update-qhld2" event={"ID":"7494a97b-56a3-45e8-a189-48eea155d3b3","Type":"ContainerStarted","Data":"a5cd21d78a21bd283178c970ea5f757ce6575ada08025ca820b128f9f7055768"} Dec 06 06:18:19 crc kubenswrapper[4809]: I1206 06:18:19.679155 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-8841-account-create-update-qhld2" podStartSLOduration=3.679130217 podStartE2EDuration="3.679130217s" podCreationTimestamp="2025-12-06 06:18:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:18:19.643344561 +0000 UTC m=+1624.532327503" watchObservedRunningTime="2025-12-06 06:18:19.679130217 +0000 UTC m=+1624.568113169" Dec 06 06:18:19 crc kubenswrapper[4809]: I1206 06:18:19.941060 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-api-97f66c66d-jlz22" Dec 06 06:18:19 crc kubenswrapper[4809]: I1206 06:18:19.942154 4809 scope.go:117] "RemoveContainer" containerID="c5a34db92db6dc08e7a02ec45c3ae41092245c47d6f77e2fde124fd9d8fbdc19" Dec 06 06:18:19 crc kubenswrapper[4809]: E1206 06:18:19.942509 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-97f66c66d-jlz22_openstack(85faa375-eb95-49b6-ba0b-967eeeebb2e8)\"" pod="openstack/heat-api-97f66c66d-jlz22" podUID="85faa375-eb95-49b6-ba0b-967eeeebb2e8" Dec 06 06:18:19 crc kubenswrapper[4809]: I1206 06:18:19.948622 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-97f66c66d-jlz22" Dec 06 06:18:20 crc kubenswrapper[4809]: I1206 06:18:20.419730 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-bmh2p" Dec 06 06:18:20 crc kubenswrapper[4809]: I1206 06:18:20.426667 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-sd9ds" Dec 06 06:18:20 crc kubenswrapper[4809]: I1206 06:18:20.550425 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hsgpz\" (UniqueName: \"kubernetes.io/projected/ee554177-cc9d-4969-9085-e507976f0f28-kube-api-access-hsgpz\") pod \"ee554177-cc9d-4969-9085-e507976f0f28\" (UID: \"ee554177-cc9d-4969-9085-e507976f0f28\") " Dec 06 06:18:20 crc kubenswrapper[4809]: I1206 06:18:20.550476 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee554177-cc9d-4969-9085-e507976f0f28-operator-scripts\") pod \"ee554177-cc9d-4969-9085-e507976f0f28\" (UID: \"ee554177-cc9d-4969-9085-e507976f0f28\") " Dec 06 06:18:20 crc kubenswrapper[4809]: I1206 06:18:20.550502 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l7z7g\" (UniqueName: \"kubernetes.io/projected/c6b76b1b-c52d-464c-9b94-1e67d459f3b3-kube-api-access-l7z7g\") pod \"c6b76b1b-c52d-464c-9b94-1e67d459f3b3\" (UID: \"c6b76b1b-c52d-464c-9b94-1e67d459f3b3\") " Dec 06 06:18:20 crc kubenswrapper[4809]: I1206 06:18:20.550760 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c6b76b1b-c52d-464c-9b94-1e67d459f3b3-operator-scripts\") pod \"c6b76b1b-c52d-464c-9b94-1e67d459f3b3\" (UID: \"c6b76b1b-c52d-464c-9b94-1e67d459f3b3\") " Dec 06 06:18:20 crc kubenswrapper[4809]: I1206 06:18:20.552536 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee554177-cc9d-4969-9085-e507976f0f28-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ee554177-cc9d-4969-9085-e507976f0f28" (UID: "ee554177-cc9d-4969-9085-e507976f0f28"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:18:20 crc kubenswrapper[4809]: I1206 06:18:20.552724 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c6b76b1b-c52d-464c-9b94-1e67d459f3b3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c6b76b1b-c52d-464c-9b94-1e67d459f3b3" (UID: "c6b76b1b-c52d-464c-9b94-1e67d459f3b3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:18:20 crc kubenswrapper[4809]: I1206 06:18:20.559803 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee554177-cc9d-4969-9085-e507976f0f28-kube-api-access-hsgpz" (OuterVolumeSpecName: "kube-api-access-hsgpz") pod "ee554177-cc9d-4969-9085-e507976f0f28" (UID: "ee554177-cc9d-4969-9085-e507976f0f28"). InnerVolumeSpecName "kube-api-access-hsgpz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:18:20 crc kubenswrapper[4809]: I1206 06:18:20.564736 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6b76b1b-c52d-464c-9b94-1e67d459f3b3-kube-api-access-l7z7g" (OuterVolumeSpecName: "kube-api-access-l7z7g") pod "c6b76b1b-c52d-464c-9b94-1e67d459f3b3" (UID: "c6b76b1b-c52d-464c-9b94-1e67d459f3b3"). InnerVolumeSpecName "kube-api-access-l7z7g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:18:20 crc kubenswrapper[4809]: I1206 06:18:20.634948 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-sd9ds" Dec 06 06:18:20 crc kubenswrapper[4809]: I1206 06:18:20.634963 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-sd9ds" event={"ID":"c6b76b1b-c52d-464c-9b94-1e67d459f3b3","Type":"ContainerDied","Data":"81675f90ca51ed353720a1d409e626e14d925c61e319c27fbbeef8681b71ccd8"} Dec 06 06:18:20 crc kubenswrapper[4809]: I1206 06:18:20.634997 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="81675f90ca51ed353720a1d409e626e14d925c61e319c27fbbeef8681b71ccd8" Dec 06 06:18:20 crc kubenswrapper[4809]: I1206 06:18:20.637222 4809 generic.go:334] "Generic (PLEG): container finished" podID="7494a97b-56a3-45e8-a189-48eea155d3b3" containerID="fa80f87dee5acf1e4481c0a5cb3a2046e006e209b4ecea87d99c9b30c9815b13" exitCode=0 Dec 06 06:18:20 crc kubenswrapper[4809]: I1206 06:18:20.637387 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-8841-account-create-update-qhld2" event={"ID":"7494a97b-56a3-45e8-a189-48eea155d3b3","Type":"ContainerDied","Data":"fa80f87dee5acf1e4481c0a5cb3a2046e006e209b4ecea87d99c9b30c9815b13"} Dec 06 06:18:20 crc kubenswrapper[4809]: I1206 06:18:20.640058 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-bmh2p" Dec 06 06:18:20 crc kubenswrapper[4809]: I1206 06:18:20.640051 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-bmh2p" event={"ID":"ee554177-cc9d-4969-9085-e507976f0f28","Type":"ContainerDied","Data":"f02ec91ea32b5d9b285dcbe3f6d0de2d62c18859d1b7e39b834360bfa1af6194"} Dec 06 06:18:20 crc kubenswrapper[4809]: I1206 06:18:20.640208 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f02ec91ea32b5d9b285dcbe3f6d0de2d62c18859d1b7e39b834360bfa1af6194" Dec 06 06:18:20 crc kubenswrapper[4809]: I1206 06:18:20.642531 4809 scope.go:117] "RemoveContainer" containerID="c5a34db92db6dc08e7a02ec45c3ae41092245c47d6f77e2fde124fd9d8fbdc19" Dec 06 06:18:20 crc kubenswrapper[4809]: E1206 06:18:20.643135 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-97f66c66d-jlz22_openstack(85faa375-eb95-49b6-ba0b-967eeeebb2e8)\"" pod="openstack/heat-api-97f66c66d-jlz22" podUID="85faa375-eb95-49b6-ba0b-967eeeebb2e8" Dec 06 06:18:20 crc kubenswrapper[4809]: I1206 06:18:20.653472 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hsgpz\" (UniqueName: \"kubernetes.io/projected/ee554177-cc9d-4969-9085-e507976f0f28-kube-api-access-hsgpz\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:20 crc kubenswrapper[4809]: I1206 06:18:20.653527 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee554177-cc9d-4969-9085-e507976f0f28-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:20 crc kubenswrapper[4809]: I1206 06:18:20.653537 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l7z7g\" (UniqueName: \"kubernetes.io/projected/c6b76b1b-c52d-464c-9b94-1e67d459f3b3-kube-api-access-l7z7g\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:20 crc kubenswrapper[4809]: I1206 06:18:20.653548 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c6b76b1b-c52d-464c-9b94-1e67d459f3b3-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:20 crc kubenswrapper[4809]: I1206 06:18:20.707127 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-bdfbd85-ntgxq" Dec 06 06:18:20 crc kubenswrapper[4809]: I1206 06:18:20.817350 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-97f66c66d-jlz22"] Dec 06 06:18:21 crc kubenswrapper[4809]: I1206 06:18:21.398073 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-356a-account-create-update-6mtbv" Dec 06 06:18:21 crc kubenswrapper[4809]: I1206 06:18:21.414441 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8757fd43-0a69-40c8-8832-53a4acd0a912" path="/var/lib/kubelet/pods/8757fd43-0a69-40c8-8832-53a4acd0a912/volumes" Dec 06 06:18:21 crc kubenswrapper[4809]: I1206 06:18:21.476824 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k5w7j\" (UniqueName: \"kubernetes.io/projected/5984ffba-4d1d-4976-b97f-606e2f557b03-kube-api-access-k5w7j\") pod \"5984ffba-4d1d-4976-b97f-606e2f557b03\" (UID: \"5984ffba-4d1d-4976-b97f-606e2f557b03\") " Dec 06 06:18:21 crc kubenswrapper[4809]: I1206 06:18:21.476993 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5984ffba-4d1d-4976-b97f-606e2f557b03-operator-scripts\") pod \"5984ffba-4d1d-4976-b97f-606e2f557b03\" (UID: \"5984ffba-4d1d-4976-b97f-606e2f557b03\") " Dec 06 06:18:21 crc kubenswrapper[4809]: I1206 06:18:21.478379 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5984ffba-4d1d-4976-b97f-606e2f557b03-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5984ffba-4d1d-4976-b97f-606e2f557b03" (UID: "5984ffba-4d1d-4976-b97f-606e2f557b03"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:18:21 crc kubenswrapper[4809]: I1206 06:18:21.533706 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5984ffba-4d1d-4976-b97f-606e2f557b03-kube-api-access-k5w7j" (OuterVolumeSpecName: "kube-api-access-k5w7j") pod "5984ffba-4d1d-4976-b97f-606e2f557b03" (UID: "5984ffba-4d1d-4976-b97f-606e2f557b03"). InnerVolumeSpecName "kube-api-access-k5w7j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:18:21 crc kubenswrapper[4809]: I1206 06:18:21.584419 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k5w7j\" (UniqueName: \"kubernetes.io/projected/5984ffba-4d1d-4976-b97f-606e2f557b03-kube-api-access-k5w7j\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:21 crc kubenswrapper[4809]: I1206 06:18:21.584464 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5984ffba-4d1d-4976-b97f-606e2f557b03-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:21 crc kubenswrapper[4809]: I1206 06:18:21.607053 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-4733-account-create-update-z8gxx" Dec 06 06:18:21 crc kubenswrapper[4809]: I1206 06:18:21.621966 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-fb6zj" Dec 06 06:18:21 crc kubenswrapper[4809]: I1206 06:18:21.666457 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"6c6578f3-bf7b-4892-abc9-5f4f145c8d36","Type":"ContainerStarted","Data":"bf99d69a932051f62ab7f6b30fbdad0e84e4dff2e95a115a9704547c15ae138b"} Dec 06 06:18:21 crc kubenswrapper[4809]: I1206 06:18:21.671741 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-356a-account-create-update-6mtbv" event={"ID":"5984ffba-4d1d-4976-b97f-606e2f557b03","Type":"ContainerDied","Data":"f0b7fcb089c3953351ae89931db079663fe4ea684f87212623a42e73c074a99a"} Dec 06 06:18:21 crc kubenswrapper[4809]: I1206 06:18:21.671777 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f0b7fcb089c3953351ae89931db079663fe4ea684f87212623a42e73c074a99a" Dec 06 06:18:21 crc kubenswrapper[4809]: I1206 06:18:21.671828 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-356a-account-create-update-6mtbv" Dec 06 06:18:21 crc kubenswrapper[4809]: I1206 06:18:21.680687 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-4733-account-create-update-z8gxx" event={"ID":"a0866cc1-8f06-4514-9939-5a4cfadc72b3","Type":"ContainerDied","Data":"70f02532c7aea47b6ab049074dcd5714ad04c59ef361d6e80dee6c286a441fd2"} Dec 06 06:18:21 crc kubenswrapper[4809]: I1206 06:18:21.680731 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="70f02532c7aea47b6ab049074dcd5714ad04c59ef361d6e80dee6c286a441fd2" Dec 06 06:18:21 crc kubenswrapper[4809]: I1206 06:18:21.680787 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-4733-account-create-update-z8gxx" Dec 06 06:18:21 crc kubenswrapper[4809]: I1206 06:18:21.685264 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a0866cc1-8f06-4514-9939-5a4cfadc72b3-operator-scripts\") pod \"a0866cc1-8f06-4514-9939-5a4cfadc72b3\" (UID: \"a0866cc1-8f06-4514-9939-5a4cfadc72b3\") " Dec 06 06:18:21 crc kubenswrapper[4809]: I1206 06:18:21.685351 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m84mp\" (UniqueName: \"kubernetes.io/projected/a0866cc1-8f06-4514-9939-5a4cfadc72b3-kube-api-access-m84mp\") pod \"a0866cc1-8f06-4514-9939-5a4cfadc72b3\" (UID: \"a0866cc1-8f06-4514-9939-5a4cfadc72b3\") " Dec 06 06:18:21 crc kubenswrapper[4809]: I1206 06:18:21.685438 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/412c5e46-e0e4-4b8a-af43-616019f9b619-operator-scripts\") pod \"412c5e46-e0e4-4b8a-af43-616019f9b619\" (UID: \"412c5e46-e0e4-4b8a-af43-616019f9b619\") " Dec 06 06:18:21 crc kubenswrapper[4809]: I1206 06:18:21.685624 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrh89\" (UniqueName: \"kubernetes.io/projected/412c5e46-e0e4-4b8a-af43-616019f9b619-kube-api-access-xrh89\") pod \"412c5e46-e0e4-4b8a-af43-616019f9b619\" (UID: \"412c5e46-e0e4-4b8a-af43-616019f9b619\") " Dec 06 06:18:21 crc kubenswrapper[4809]: I1206 06:18:21.686482 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/412c5e46-e0e4-4b8a-af43-616019f9b619-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "412c5e46-e0e4-4b8a-af43-616019f9b619" (UID: "412c5e46-e0e4-4b8a-af43-616019f9b619"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:18:21 crc kubenswrapper[4809]: I1206 06:18:21.687245 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/412c5e46-e0e4-4b8a-af43-616019f9b619-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:21 crc kubenswrapper[4809]: I1206 06:18:21.687670 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-fb6zj" Dec 06 06:18:21 crc kubenswrapper[4809]: I1206 06:18:21.688074 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-fb6zj" event={"ID":"412c5e46-e0e4-4b8a-af43-616019f9b619","Type":"ContainerDied","Data":"87701fbc106da4b39116db9ecfe6b1722842de90a9984d0a9c462d7876159d2b"} Dec 06 06:18:21 crc kubenswrapper[4809]: I1206 06:18:21.688100 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="87701fbc106da4b39116db9ecfe6b1722842de90a9984d0a9c462d7876159d2b" Dec 06 06:18:21 crc kubenswrapper[4809]: I1206 06:18:21.688524 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a0866cc1-8f06-4514-9939-5a4cfadc72b3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a0866cc1-8f06-4514-9939-5a4cfadc72b3" (UID: "a0866cc1-8f06-4514-9939-5a4cfadc72b3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:18:21 crc kubenswrapper[4809]: I1206 06:18:21.693331 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/412c5e46-e0e4-4b8a-af43-616019f9b619-kube-api-access-xrh89" (OuterVolumeSpecName: "kube-api-access-xrh89") pod "412c5e46-e0e4-4b8a-af43-616019f9b619" (UID: "412c5e46-e0e4-4b8a-af43-616019f9b619"). InnerVolumeSpecName "kube-api-access-xrh89". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:18:21 crc kubenswrapper[4809]: I1206 06:18:21.712024 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.977562791 podStartE2EDuration="37.712001083s" podCreationTimestamp="2025-12-06 06:17:44 +0000 UTC" firstStartedPulling="2025-12-06 06:17:45.166227511 +0000 UTC m=+1590.055210453" lastFinishedPulling="2025-12-06 06:18:20.900665803 +0000 UTC m=+1625.789648745" observedRunningTime="2025-12-06 06:18:21.688737455 +0000 UTC m=+1626.577720387" watchObservedRunningTime="2025-12-06 06:18:21.712001083 +0000 UTC m=+1626.600984025" Dec 06 06:18:21 crc kubenswrapper[4809]: I1206 06:18:21.718230 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0866cc1-8f06-4514-9939-5a4cfadc72b3-kube-api-access-m84mp" (OuterVolumeSpecName: "kube-api-access-m84mp") pod "a0866cc1-8f06-4514-9939-5a4cfadc72b3" (UID: "a0866cc1-8f06-4514-9939-5a4cfadc72b3"). InnerVolumeSpecName "kube-api-access-m84mp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:18:21 crc kubenswrapper[4809]: I1206 06:18:21.789520 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a0866cc1-8f06-4514-9939-5a4cfadc72b3-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:21 crc kubenswrapper[4809]: I1206 06:18:21.789553 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m84mp\" (UniqueName: \"kubernetes.io/projected/a0866cc1-8f06-4514-9939-5a4cfadc72b3-kube-api-access-m84mp\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:21 crc kubenswrapper[4809]: I1206 06:18:21.789567 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrh89\" (UniqueName: \"kubernetes.io/projected/412c5e46-e0e4-4b8a-af43-616019f9b619-kube-api-access-xrh89\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:22 crc kubenswrapper[4809]: I1206 06:18:22.080537 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-7c94c94748-57b42" Dec 06 06:18:22 crc kubenswrapper[4809]: I1206 06:18:22.339011 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-8841-account-create-update-qhld2" Dec 06 06:18:22 crc kubenswrapper[4809]: I1206 06:18:22.345758 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-97f66c66d-jlz22" Dec 06 06:18:22 crc kubenswrapper[4809]: I1206 06:18:22.405541 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85faa375-eb95-49b6-ba0b-967eeeebb2e8-combined-ca-bundle\") pod \"85faa375-eb95-49b6-ba0b-967eeeebb2e8\" (UID: \"85faa375-eb95-49b6-ba0b-967eeeebb2e8\") " Dec 06 06:18:22 crc kubenswrapper[4809]: I1206 06:18:22.405719 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8d6v8\" (UniqueName: \"kubernetes.io/projected/85faa375-eb95-49b6-ba0b-967eeeebb2e8-kube-api-access-8d6v8\") pod \"85faa375-eb95-49b6-ba0b-967eeeebb2e8\" (UID: \"85faa375-eb95-49b6-ba0b-967eeeebb2e8\") " Dec 06 06:18:22 crc kubenswrapper[4809]: I1206 06:18:22.405797 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85faa375-eb95-49b6-ba0b-967eeeebb2e8-config-data\") pod \"85faa375-eb95-49b6-ba0b-967eeeebb2e8\" (UID: \"85faa375-eb95-49b6-ba0b-967eeeebb2e8\") " Dec 06 06:18:22 crc kubenswrapper[4809]: I1206 06:18:22.405851 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/85faa375-eb95-49b6-ba0b-967eeeebb2e8-config-data-custom\") pod \"85faa375-eb95-49b6-ba0b-967eeeebb2e8\" (UID: \"85faa375-eb95-49b6-ba0b-967eeeebb2e8\") " Dec 06 06:18:22 crc kubenswrapper[4809]: I1206 06:18:22.406037 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7494a97b-56a3-45e8-a189-48eea155d3b3-operator-scripts\") pod \"7494a97b-56a3-45e8-a189-48eea155d3b3\" (UID: \"7494a97b-56a3-45e8-a189-48eea155d3b3\") " Dec 06 06:18:22 crc kubenswrapper[4809]: I1206 06:18:22.406068 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27c8j\" (UniqueName: \"kubernetes.io/projected/7494a97b-56a3-45e8-a189-48eea155d3b3-kube-api-access-27c8j\") pod \"7494a97b-56a3-45e8-a189-48eea155d3b3\" (UID: \"7494a97b-56a3-45e8-a189-48eea155d3b3\") " Dec 06 06:18:22 crc kubenswrapper[4809]: I1206 06:18:22.409349 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7494a97b-56a3-45e8-a189-48eea155d3b3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7494a97b-56a3-45e8-a189-48eea155d3b3" (UID: "7494a97b-56a3-45e8-a189-48eea155d3b3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:18:22 crc kubenswrapper[4809]: I1206 06:18:22.413259 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85faa375-eb95-49b6-ba0b-967eeeebb2e8-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "85faa375-eb95-49b6-ba0b-967eeeebb2e8" (UID: "85faa375-eb95-49b6-ba0b-967eeeebb2e8"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:18:22 crc kubenswrapper[4809]: I1206 06:18:22.420801 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85faa375-eb95-49b6-ba0b-967eeeebb2e8-kube-api-access-8d6v8" (OuterVolumeSpecName: "kube-api-access-8d6v8") pod "85faa375-eb95-49b6-ba0b-967eeeebb2e8" (UID: "85faa375-eb95-49b6-ba0b-967eeeebb2e8"). InnerVolumeSpecName "kube-api-access-8d6v8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:18:22 crc kubenswrapper[4809]: I1206 06:18:22.425076 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7494a97b-56a3-45e8-a189-48eea155d3b3-kube-api-access-27c8j" (OuterVolumeSpecName: "kube-api-access-27c8j") pod "7494a97b-56a3-45e8-a189-48eea155d3b3" (UID: "7494a97b-56a3-45e8-a189-48eea155d3b3"). InnerVolumeSpecName "kube-api-access-27c8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:18:22 crc kubenswrapper[4809]: I1206 06:18:22.487135 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85faa375-eb95-49b6-ba0b-967eeeebb2e8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "85faa375-eb95-49b6-ba0b-967eeeebb2e8" (UID: "85faa375-eb95-49b6-ba0b-967eeeebb2e8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:18:22 crc kubenswrapper[4809]: I1206 06:18:22.509242 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85faa375-eb95-49b6-ba0b-967eeeebb2e8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:22 crc kubenswrapper[4809]: I1206 06:18:22.509277 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8d6v8\" (UniqueName: \"kubernetes.io/projected/85faa375-eb95-49b6-ba0b-967eeeebb2e8-kube-api-access-8d6v8\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:22 crc kubenswrapper[4809]: I1206 06:18:22.509290 4809 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/85faa375-eb95-49b6-ba0b-967eeeebb2e8-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:22 crc kubenswrapper[4809]: I1206 06:18:22.509299 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7494a97b-56a3-45e8-a189-48eea155d3b3-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:22 crc kubenswrapper[4809]: I1206 06:18:22.509308 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27c8j\" (UniqueName: \"kubernetes.io/projected/7494a97b-56a3-45e8-a189-48eea155d3b3-kube-api-access-27c8j\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:22 crc kubenswrapper[4809]: I1206 06:18:22.530006 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85faa375-eb95-49b6-ba0b-967eeeebb2e8-config-data" (OuterVolumeSpecName: "config-data") pod "85faa375-eb95-49b6-ba0b-967eeeebb2e8" (UID: "85faa375-eb95-49b6-ba0b-967eeeebb2e8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:18:22 crc kubenswrapper[4809]: I1206 06:18:22.611297 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85faa375-eb95-49b6-ba0b-967eeeebb2e8-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:22 crc kubenswrapper[4809]: I1206 06:18:22.701133 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-97f66c66d-jlz22" event={"ID":"85faa375-eb95-49b6-ba0b-967eeeebb2e8","Type":"ContainerDied","Data":"5196b40cec8315a33e9d4b30a8977972d17c2168ecc1c198817eb75438793ac6"} Dec 06 06:18:22 crc kubenswrapper[4809]: I1206 06:18:22.701149 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-97f66c66d-jlz22" Dec 06 06:18:22 crc kubenswrapper[4809]: I1206 06:18:22.701207 4809 scope.go:117] "RemoveContainer" containerID="c5a34db92db6dc08e7a02ec45c3ae41092245c47d6f77e2fde124fd9d8fbdc19" Dec 06 06:18:22 crc kubenswrapper[4809]: I1206 06:18:22.702652 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-8841-account-create-update-qhld2" event={"ID":"7494a97b-56a3-45e8-a189-48eea155d3b3","Type":"ContainerDied","Data":"a5cd21d78a21bd283178c970ea5f757ce6575ada08025ca820b128f9f7055768"} Dec 06 06:18:22 crc kubenswrapper[4809]: I1206 06:18:22.702685 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a5cd21d78a21bd283178c970ea5f757ce6575ada08025ca820b128f9f7055768" Dec 06 06:18:22 crc kubenswrapper[4809]: I1206 06:18:22.702707 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-8841-account-create-update-qhld2" Dec 06 06:18:22 crc kubenswrapper[4809]: I1206 06:18:22.745586 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-97f66c66d-jlz22"] Dec 06 06:18:22 crc kubenswrapper[4809]: I1206 06:18:22.766383 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-97f66c66d-jlz22"] Dec 06 06:18:23 crc kubenswrapper[4809]: I1206 06:18:23.422822 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85faa375-eb95-49b6-ba0b-967eeeebb2e8" path="/var/lib/kubelet/pods/85faa375-eb95-49b6-ba0b-967eeeebb2e8/volumes" Dec 06 06:18:23 crc kubenswrapper[4809]: I1206 06:18:23.652232 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-9f9dd4fbb-4pcn7" Dec 06 06:18:23 crc kubenswrapper[4809]: I1206 06:18:23.734965 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-c48964f4f-ktxp6"] Dec 06 06:18:24 crc kubenswrapper[4809]: I1206 06:18:24.344917 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-c48964f4f-ktxp6" Dec 06 06:18:24 crc kubenswrapper[4809]: I1206 06:18:24.459551 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a5c1cb9-6796-43cd-8c09-68ade099a586-config-data\") pod \"0a5c1cb9-6796-43cd-8c09-68ade099a586\" (UID: \"0a5c1cb9-6796-43cd-8c09-68ade099a586\") " Dec 06 06:18:24 crc kubenswrapper[4809]: I1206 06:18:24.459800 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a5c1cb9-6796-43cd-8c09-68ade099a586-combined-ca-bundle\") pod \"0a5c1cb9-6796-43cd-8c09-68ade099a586\" (UID: \"0a5c1cb9-6796-43cd-8c09-68ade099a586\") " Dec 06 06:18:24 crc kubenswrapper[4809]: I1206 06:18:24.459846 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0a5c1cb9-6796-43cd-8c09-68ade099a586-config-data-custom\") pod \"0a5c1cb9-6796-43cd-8c09-68ade099a586\" (UID: \"0a5c1cb9-6796-43cd-8c09-68ade099a586\") " Dec 06 06:18:24 crc kubenswrapper[4809]: I1206 06:18:24.459895 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgj5f\" (UniqueName: \"kubernetes.io/projected/0a5c1cb9-6796-43cd-8c09-68ade099a586-kube-api-access-bgj5f\") pod \"0a5c1cb9-6796-43cd-8c09-68ade099a586\" (UID: \"0a5c1cb9-6796-43cd-8c09-68ade099a586\") " Dec 06 06:18:24 crc kubenswrapper[4809]: I1206 06:18:24.476222 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a5c1cb9-6796-43cd-8c09-68ade099a586-kube-api-access-bgj5f" (OuterVolumeSpecName: "kube-api-access-bgj5f") pod "0a5c1cb9-6796-43cd-8c09-68ade099a586" (UID: "0a5c1cb9-6796-43cd-8c09-68ade099a586"). InnerVolumeSpecName "kube-api-access-bgj5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:18:24 crc kubenswrapper[4809]: I1206 06:18:24.491089 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a5c1cb9-6796-43cd-8c09-68ade099a586-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "0a5c1cb9-6796-43cd-8c09-68ade099a586" (UID: "0a5c1cb9-6796-43cd-8c09-68ade099a586"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:18:24 crc kubenswrapper[4809]: I1206 06:18:24.508013 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a5c1cb9-6796-43cd-8c09-68ade099a586-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0a5c1cb9-6796-43cd-8c09-68ade099a586" (UID: "0a5c1cb9-6796-43cd-8c09-68ade099a586"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:18:24 crc kubenswrapper[4809]: I1206 06:18:24.529891 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a5c1cb9-6796-43cd-8c09-68ade099a586-config-data" (OuterVolumeSpecName: "config-data") pod "0a5c1cb9-6796-43cd-8c09-68ade099a586" (UID: "0a5c1cb9-6796-43cd-8c09-68ade099a586"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:18:24 crc kubenswrapper[4809]: I1206 06:18:24.562342 4809 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0a5c1cb9-6796-43cd-8c09-68ade099a586-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:24 crc kubenswrapper[4809]: I1206 06:18:24.562378 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgj5f\" (UniqueName: \"kubernetes.io/projected/0a5c1cb9-6796-43cd-8c09-68ade099a586-kube-api-access-bgj5f\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:24 crc kubenswrapper[4809]: I1206 06:18:24.562393 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a5c1cb9-6796-43cd-8c09-68ade099a586-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:24 crc kubenswrapper[4809]: I1206 06:18:24.562406 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a5c1cb9-6796-43cd-8c09-68ade099a586-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:24 crc kubenswrapper[4809]: I1206 06:18:24.784181 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-c48964f4f-ktxp6" event={"ID":"0a5c1cb9-6796-43cd-8c09-68ade099a586","Type":"ContainerDied","Data":"88ee53f5e833ba9abf430b5608360b12bdf739e3bfb70da2ac427465fc1e150f"} Dec 06 06:18:24 crc kubenswrapper[4809]: I1206 06:18:24.784569 4809 scope.go:117] "RemoveContainer" containerID="1d0263157b9458dfb5dfb10941e0c8b0e1d1278bb04e62b0c7baa14522ce3dd1" Dec 06 06:18:24 crc kubenswrapper[4809]: I1206 06:18:24.784241 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-c48964f4f-ktxp6" Dec 06 06:18:24 crc kubenswrapper[4809]: I1206 06:18:24.826043 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-c48964f4f-ktxp6"] Dec 06 06:18:24 crc kubenswrapper[4809]: I1206 06:18:24.838085 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-c48964f4f-ktxp6"] Dec 06 06:18:25 crc kubenswrapper[4809]: I1206 06:18:25.412418 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a5c1cb9-6796-43cd-8c09-68ade099a586" path="/var/lib/kubelet/pods/0a5c1cb9-6796-43cd-8c09-68ade099a586/volumes" Dec 06 06:18:26 crc kubenswrapper[4809]: I1206 06:18:26.620521 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-vt99m"] Dec 06 06:18:26 crc kubenswrapper[4809]: E1206 06:18:26.621466 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85faa375-eb95-49b6-ba0b-967eeeebb2e8" containerName="heat-api" Dec 06 06:18:26 crc kubenswrapper[4809]: I1206 06:18:26.621489 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="85faa375-eb95-49b6-ba0b-967eeeebb2e8" containerName="heat-api" Dec 06 06:18:26 crc kubenswrapper[4809]: E1206 06:18:26.621508 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="412c5e46-e0e4-4b8a-af43-616019f9b619" containerName="mariadb-database-create" Dec 06 06:18:26 crc kubenswrapper[4809]: I1206 06:18:26.621516 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="412c5e46-e0e4-4b8a-af43-616019f9b619" containerName="mariadb-database-create" Dec 06 06:18:26 crc kubenswrapper[4809]: E1206 06:18:26.621533 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5984ffba-4d1d-4976-b97f-606e2f557b03" containerName="mariadb-account-create-update" Dec 06 06:18:26 crc kubenswrapper[4809]: I1206 06:18:26.621543 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="5984ffba-4d1d-4976-b97f-606e2f557b03" containerName="mariadb-account-create-update" Dec 06 06:18:26 crc kubenswrapper[4809]: E1206 06:18:26.621564 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0866cc1-8f06-4514-9939-5a4cfadc72b3" containerName="mariadb-account-create-update" Dec 06 06:18:26 crc kubenswrapper[4809]: I1206 06:18:26.621574 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0866cc1-8f06-4514-9939-5a4cfadc72b3" containerName="mariadb-account-create-update" Dec 06 06:18:26 crc kubenswrapper[4809]: E1206 06:18:26.621593 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8757fd43-0a69-40c8-8832-53a4acd0a912" containerName="dnsmasq-dns" Dec 06 06:18:26 crc kubenswrapper[4809]: I1206 06:18:26.621601 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8757fd43-0a69-40c8-8832-53a4acd0a912" containerName="dnsmasq-dns" Dec 06 06:18:26 crc kubenswrapper[4809]: E1206 06:18:26.621618 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6b76b1b-c52d-464c-9b94-1e67d459f3b3" containerName="mariadb-database-create" Dec 06 06:18:26 crc kubenswrapper[4809]: I1206 06:18:26.621626 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6b76b1b-c52d-464c-9b94-1e67d459f3b3" containerName="mariadb-database-create" Dec 06 06:18:26 crc kubenswrapper[4809]: E1206 06:18:26.621637 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee554177-cc9d-4969-9085-e507976f0f28" containerName="mariadb-database-create" Dec 06 06:18:26 crc kubenswrapper[4809]: I1206 06:18:26.621645 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee554177-cc9d-4969-9085-e507976f0f28" containerName="mariadb-database-create" Dec 06 06:18:26 crc kubenswrapper[4809]: E1206 06:18:26.621679 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7494a97b-56a3-45e8-a189-48eea155d3b3" containerName="mariadb-account-create-update" Dec 06 06:18:26 crc kubenswrapper[4809]: I1206 06:18:26.621689 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7494a97b-56a3-45e8-a189-48eea155d3b3" containerName="mariadb-account-create-update" Dec 06 06:18:26 crc kubenswrapper[4809]: E1206 06:18:26.621700 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a5c1cb9-6796-43cd-8c09-68ade099a586" containerName="heat-cfnapi" Dec 06 06:18:26 crc kubenswrapper[4809]: I1206 06:18:26.621708 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a5c1cb9-6796-43cd-8c09-68ade099a586" containerName="heat-cfnapi" Dec 06 06:18:26 crc kubenswrapper[4809]: E1206 06:18:26.621739 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a5c1cb9-6796-43cd-8c09-68ade099a586" containerName="heat-cfnapi" Dec 06 06:18:26 crc kubenswrapper[4809]: I1206 06:18:26.621747 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a5c1cb9-6796-43cd-8c09-68ade099a586" containerName="heat-cfnapi" Dec 06 06:18:26 crc kubenswrapper[4809]: E1206 06:18:26.621764 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8757fd43-0a69-40c8-8832-53a4acd0a912" containerName="init" Dec 06 06:18:26 crc kubenswrapper[4809]: I1206 06:18:26.621771 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8757fd43-0a69-40c8-8832-53a4acd0a912" containerName="init" Dec 06 06:18:26 crc kubenswrapper[4809]: I1206 06:18:26.622119 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="85faa375-eb95-49b6-ba0b-967eeeebb2e8" containerName="heat-api" Dec 06 06:18:26 crc kubenswrapper[4809]: I1206 06:18:26.622142 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6b76b1b-c52d-464c-9b94-1e67d459f3b3" containerName="mariadb-database-create" Dec 06 06:18:26 crc kubenswrapper[4809]: I1206 06:18:26.622159 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="5984ffba-4d1d-4976-b97f-606e2f557b03" containerName="mariadb-account-create-update" Dec 06 06:18:26 crc kubenswrapper[4809]: I1206 06:18:26.622176 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="412c5e46-e0e4-4b8a-af43-616019f9b619" containerName="mariadb-database-create" Dec 06 06:18:26 crc kubenswrapper[4809]: I1206 06:18:26.622193 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a5c1cb9-6796-43cd-8c09-68ade099a586" containerName="heat-cfnapi" Dec 06 06:18:26 crc kubenswrapper[4809]: I1206 06:18:26.622202 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a5c1cb9-6796-43cd-8c09-68ade099a586" containerName="heat-cfnapi" Dec 06 06:18:26 crc kubenswrapper[4809]: I1206 06:18:26.622211 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee554177-cc9d-4969-9085-e507976f0f28" containerName="mariadb-database-create" Dec 06 06:18:26 crc kubenswrapper[4809]: I1206 06:18:26.622227 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="8757fd43-0a69-40c8-8832-53a4acd0a912" containerName="dnsmasq-dns" Dec 06 06:18:26 crc kubenswrapper[4809]: I1206 06:18:26.622243 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="7494a97b-56a3-45e8-a189-48eea155d3b3" containerName="mariadb-account-create-update" Dec 06 06:18:26 crc kubenswrapper[4809]: I1206 06:18:26.622256 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0866cc1-8f06-4514-9939-5a4cfadc72b3" containerName="mariadb-account-create-update" Dec 06 06:18:26 crc kubenswrapper[4809]: I1206 06:18:26.623285 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-vt99m" Dec 06 06:18:26 crc kubenswrapper[4809]: I1206 06:18:26.624749 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 06 06:18:26 crc kubenswrapper[4809]: I1206 06:18:26.625602 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 06 06:18:26 crc kubenswrapper[4809]: I1206 06:18:26.627278 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-tjnnd" Dec 06 06:18:26 crc kubenswrapper[4809]: I1206 06:18:26.650084 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-vt99m"] Dec 06 06:18:26 crc kubenswrapper[4809]: I1206 06:18:26.711547 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13e0242d-c42e-4aad-95a2-62de95d87e42-scripts\") pod \"nova-cell0-conductor-db-sync-vt99m\" (UID: \"13e0242d-c42e-4aad-95a2-62de95d87e42\") " pod="openstack/nova-cell0-conductor-db-sync-vt99m" Dec 06 06:18:26 crc kubenswrapper[4809]: I1206 06:18:26.711974 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13e0242d-c42e-4aad-95a2-62de95d87e42-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-vt99m\" (UID: \"13e0242d-c42e-4aad-95a2-62de95d87e42\") " pod="openstack/nova-cell0-conductor-db-sync-vt99m" Dec 06 06:18:26 crc kubenswrapper[4809]: I1206 06:18:26.712112 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bxt9\" (UniqueName: \"kubernetes.io/projected/13e0242d-c42e-4aad-95a2-62de95d87e42-kube-api-access-4bxt9\") pod \"nova-cell0-conductor-db-sync-vt99m\" (UID: \"13e0242d-c42e-4aad-95a2-62de95d87e42\") " pod="openstack/nova-cell0-conductor-db-sync-vt99m" Dec 06 06:18:26 crc kubenswrapper[4809]: I1206 06:18:26.712445 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13e0242d-c42e-4aad-95a2-62de95d87e42-config-data\") pod \"nova-cell0-conductor-db-sync-vt99m\" (UID: \"13e0242d-c42e-4aad-95a2-62de95d87e42\") " pod="openstack/nova-cell0-conductor-db-sync-vt99m" Dec 06 06:18:26 crc kubenswrapper[4809]: I1206 06:18:26.815078 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13e0242d-c42e-4aad-95a2-62de95d87e42-config-data\") pod \"nova-cell0-conductor-db-sync-vt99m\" (UID: \"13e0242d-c42e-4aad-95a2-62de95d87e42\") " pod="openstack/nova-cell0-conductor-db-sync-vt99m" Dec 06 06:18:26 crc kubenswrapper[4809]: I1206 06:18:26.815405 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13e0242d-c42e-4aad-95a2-62de95d87e42-scripts\") pod \"nova-cell0-conductor-db-sync-vt99m\" (UID: \"13e0242d-c42e-4aad-95a2-62de95d87e42\") " pod="openstack/nova-cell0-conductor-db-sync-vt99m" Dec 06 06:18:26 crc kubenswrapper[4809]: I1206 06:18:26.815597 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13e0242d-c42e-4aad-95a2-62de95d87e42-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-vt99m\" (UID: \"13e0242d-c42e-4aad-95a2-62de95d87e42\") " pod="openstack/nova-cell0-conductor-db-sync-vt99m" Dec 06 06:18:26 crc kubenswrapper[4809]: I1206 06:18:26.815738 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bxt9\" (UniqueName: \"kubernetes.io/projected/13e0242d-c42e-4aad-95a2-62de95d87e42-kube-api-access-4bxt9\") pod \"nova-cell0-conductor-db-sync-vt99m\" (UID: \"13e0242d-c42e-4aad-95a2-62de95d87e42\") " pod="openstack/nova-cell0-conductor-db-sync-vt99m" Dec 06 06:18:26 crc kubenswrapper[4809]: I1206 06:18:26.821489 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13e0242d-c42e-4aad-95a2-62de95d87e42-scripts\") pod \"nova-cell0-conductor-db-sync-vt99m\" (UID: \"13e0242d-c42e-4aad-95a2-62de95d87e42\") " pod="openstack/nova-cell0-conductor-db-sync-vt99m" Dec 06 06:18:26 crc kubenswrapper[4809]: I1206 06:18:26.821595 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13e0242d-c42e-4aad-95a2-62de95d87e42-config-data\") pod \"nova-cell0-conductor-db-sync-vt99m\" (UID: \"13e0242d-c42e-4aad-95a2-62de95d87e42\") " pod="openstack/nova-cell0-conductor-db-sync-vt99m" Dec 06 06:18:26 crc kubenswrapper[4809]: I1206 06:18:26.827597 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13e0242d-c42e-4aad-95a2-62de95d87e42-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-vt99m\" (UID: \"13e0242d-c42e-4aad-95a2-62de95d87e42\") " pod="openstack/nova-cell0-conductor-db-sync-vt99m" Dec 06 06:18:26 crc kubenswrapper[4809]: I1206 06:18:26.838242 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bxt9\" (UniqueName: \"kubernetes.io/projected/13e0242d-c42e-4aad-95a2-62de95d87e42-kube-api-access-4bxt9\") pod \"nova-cell0-conductor-db-sync-vt99m\" (UID: \"13e0242d-c42e-4aad-95a2-62de95d87e42\") " pod="openstack/nova-cell0-conductor-db-sync-vt99m" Dec 06 06:18:26 crc kubenswrapper[4809]: I1206 06:18:26.953724 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-vt99m" Dec 06 06:18:27 crc kubenswrapper[4809]: I1206 06:18:27.494278 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-vt99m"] Dec 06 06:18:27 crc kubenswrapper[4809]: W1206 06:18:27.496264 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod13e0242d_c42e_4aad_95a2_62de95d87e42.slice/crio-5feeb302e32bc6c326fbdd2d33d8d85d6cabddb04bc953da8551a17a3e501fb8 WatchSource:0}: Error finding container 5feeb302e32bc6c326fbdd2d33d8d85d6cabddb04bc953da8551a17a3e501fb8: Status 404 returned error can't find the container with id 5feeb302e32bc6c326fbdd2d33d8d85d6cabddb04bc953da8551a17a3e501fb8 Dec 06 06:18:27 crc kubenswrapper[4809]: I1206 06:18:27.819692 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-vt99m" event={"ID":"13e0242d-c42e-4aad-95a2-62de95d87e42","Type":"ContainerStarted","Data":"5feeb302e32bc6c326fbdd2d33d8d85d6cabddb04bc953da8551a17a3e501fb8"} Dec 06 06:18:27 crc kubenswrapper[4809]: I1206 06:18:27.952213 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-v6pcg"] Dec 06 06:18:27 crc kubenswrapper[4809]: E1206 06:18:27.952948 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85faa375-eb95-49b6-ba0b-967eeeebb2e8" containerName="heat-api" Dec 06 06:18:27 crc kubenswrapper[4809]: I1206 06:18:27.952964 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="85faa375-eb95-49b6-ba0b-967eeeebb2e8" containerName="heat-api" Dec 06 06:18:27 crc kubenswrapper[4809]: I1206 06:18:27.953201 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="85faa375-eb95-49b6-ba0b-967eeeebb2e8" containerName="heat-api" Dec 06 06:18:27 crc kubenswrapper[4809]: I1206 06:18:27.955355 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v6pcg" Dec 06 06:18:27 crc kubenswrapper[4809]: I1206 06:18:27.995304 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v6pcg"] Dec 06 06:18:28 crc kubenswrapper[4809]: I1206 06:18:28.052248 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d97f700e-65e3-41e3-a053-a9e97bd9857f-utilities\") pod \"community-operators-v6pcg\" (UID: \"d97f700e-65e3-41e3-a053-a9e97bd9857f\") " pod="openshift-marketplace/community-operators-v6pcg" Dec 06 06:18:28 crc kubenswrapper[4809]: I1206 06:18:28.052318 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwr5s\" (UniqueName: \"kubernetes.io/projected/d97f700e-65e3-41e3-a053-a9e97bd9857f-kube-api-access-bwr5s\") pod \"community-operators-v6pcg\" (UID: \"d97f700e-65e3-41e3-a053-a9e97bd9857f\") " pod="openshift-marketplace/community-operators-v6pcg" Dec 06 06:18:28 crc kubenswrapper[4809]: I1206 06:18:28.052343 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d97f700e-65e3-41e3-a053-a9e97bd9857f-catalog-content\") pod \"community-operators-v6pcg\" (UID: \"d97f700e-65e3-41e3-a053-a9e97bd9857f\") " pod="openshift-marketplace/community-operators-v6pcg" Dec 06 06:18:28 crc kubenswrapper[4809]: I1206 06:18:28.154376 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d97f700e-65e3-41e3-a053-a9e97bd9857f-utilities\") pod \"community-operators-v6pcg\" (UID: \"d97f700e-65e3-41e3-a053-a9e97bd9857f\") " pod="openshift-marketplace/community-operators-v6pcg" Dec 06 06:18:28 crc kubenswrapper[4809]: I1206 06:18:28.154695 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwr5s\" (UniqueName: \"kubernetes.io/projected/d97f700e-65e3-41e3-a053-a9e97bd9857f-kube-api-access-bwr5s\") pod \"community-operators-v6pcg\" (UID: \"d97f700e-65e3-41e3-a053-a9e97bd9857f\") " pod="openshift-marketplace/community-operators-v6pcg" Dec 06 06:18:28 crc kubenswrapper[4809]: I1206 06:18:28.154800 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d97f700e-65e3-41e3-a053-a9e97bd9857f-catalog-content\") pod \"community-operators-v6pcg\" (UID: \"d97f700e-65e3-41e3-a053-a9e97bd9857f\") " pod="openshift-marketplace/community-operators-v6pcg" Dec 06 06:18:28 crc kubenswrapper[4809]: I1206 06:18:28.155011 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d97f700e-65e3-41e3-a053-a9e97bd9857f-utilities\") pod \"community-operators-v6pcg\" (UID: \"d97f700e-65e3-41e3-a053-a9e97bd9857f\") " pod="openshift-marketplace/community-operators-v6pcg" Dec 06 06:18:28 crc kubenswrapper[4809]: I1206 06:18:28.155472 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d97f700e-65e3-41e3-a053-a9e97bd9857f-catalog-content\") pod \"community-operators-v6pcg\" (UID: \"d97f700e-65e3-41e3-a053-a9e97bd9857f\") " pod="openshift-marketplace/community-operators-v6pcg" Dec 06 06:18:28 crc kubenswrapper[4809]: I1206 06:18:28.177564 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwr5s\" (UniqueName: \"kubernetes.io/projected/d97f700e-65e3-41e3-a053-a9e97bd9857f-kube-api-access-bwr5s\") pod \"community-operators-v6pcg\" (UID: \"d97f700e-65e3-41e3-a053-a9e97bd9857f\") " pod="openshift-marketplace/community-operators-v6pcg" Dec 06 06:18:28 crc kubenswrapper[4809]: I1206 06:18:28.288095 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v6pcg" Dec 06 06:18:28 crc kubenswrapper[4809]: W1206 06:18:28.805215 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd97f700e_65e3_41e3_a053_a9e97bd9857f.slice/crio-a9dca06c39aa1bc14016f04104c55bdb8599488fe076e01ed59cdb8a2e4f6184 WatchSource:0}: Error finding container a9dca06c39aa1bc14016f04104c55bdb8599488fe076e01ed59cdb8a2e4f6184: Status 404 returned error can't find the container with id a9dca06c39aa1bc14016f04104c55bdb8599488fe076e01ed59cdb8a2e4f6184 Dec 06 06:18:28 crc kubenswrapper[4809]: I1206 06:18:28.813069 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v6pcg"] Dec 06 06:18:28 crc kubenswrapper[4809]: I1206 06:18:28.856359 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6pcg" event={"ID":"d97f700e-65e3-41e3-a053-a9e97bd9857f","Type":"ContainerStarted","Data":"a9dca06c39aa1bc14016f04104c55bdb8599488fe076e01ed59cdb8a2e4f6184"} Dec 06 06:18:29 crc kubenswrapper[4809]: I1206 06:18:29.869521 4809 generic.go:334] "Generic (PLEG): container finished" podID="d97f700e-65e3-41e3-a053-a9e97bd9857f" containerID="51eff57acd4a5b48273d67a5859417435e2ab837f35dcb56b80ea46a0e230ef4" exitCode=0 Dec 06 06:18:29 crc kubenswrapper[4809]: I1206 06:18:29.869578 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6pcg" event={"ID":"d97f700e-65e3-41e3-a053-a9e97bd9857f","Type":"ContainerDied","Data":"51eff57acd4a5b48273d67a5859417435e2ab837f35dcb56b80ea46a0e230ef4"} Dec 06 06:18:29 crc kubenswrapper[4809]: I1206 06:18:29.934717 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-6d8ccd7c9f-nbxnr" Dec 06 06:18:29 crc kubenswrapper[4809]: I1206 06:18:29.987438 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-7c94c94748-57b42"] Dec 06 06:18:29 crc kubenswrapper[4809]: I1206 06:18:29.987686 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-7c94c94748-57b42" podUID="515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a" containerName="heat-engine" containerID="cri-o://1b4497e7e985b3dcecad685e6cbc0fe4e3ac63269895ab1f86489595826907ef" gracePeriod=60 Dec 06 06:18:31 crc kubenswrapper[4809]: I1206 06:18:31.900430 4809 generic.go:334] "Generic (PLEG): container finished" podID="53615cf8-5ea5-4d28-a4e7-1491e2b33744" containerID="9dc198feff550848f012dfdf8ff0459b0a049a7da9c085bb6ac08abdb3e3c283" exitCode=0 Dec 06 06:18:31 crc kubenswrapper[4809]: I1206 06:18:31.900655 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"53615cf8-5ea5-4d28-a4e7-1491e2b33744","Type":"ContainerDied","Data":"9dc198feff550848f012dfdf8ff0459b0a049a7da9c085bb6ac08abdb3e3c283"} Dec 06 06:18:31 crc kubenswrapper[4809]: E1206 06:18:31.983042 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1b4497e7e985b3dcecad685e6cbc0fe4e3ac63269895ab1f86489595826907ef" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 06 06:18:31 crc kubenswrapper[4809]: E1206 06:18:31.985034 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1b4497e7e985b3dcecad685e6cbc0fe4e3ac63269895ab1f86489595826907ef" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 06 06:18:31 crc kubenswrapper[4809]: E1206 06:18:31.986414 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1b4497e7e985b3dcecad685e6cbc0fe4e3ac63269895ab1f86489595826907ef" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 06 06:18:31 crc kubenswrapper[4809]: E1206 06:18:31.986470 4809 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-7c94c94748-57b42" podUID="515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a" containerName="heat-engine" Dec 06 06:18:32 crc kubenswrapper[4809]: I1206 06:18:32.326995 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zx9rg"] Dec 06 06:18:32 crc kubenswrapper[4809]: I1206 06:18:32.331168 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zx9rg" Dec 06 06:18:32 crc kubenswrapper[4809]: I1206 06:18:32.352896 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zx9rg"] Dec 06 06:18:32 crc kubenswrapper[4809]: I1206 06:18:32.481915 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4adfb97-c4f0-4093-80c5-e0229644834c-catalog-content\") pod \"redhat-marketplace-zx9rg\" (UID: \"d4adfb97-c4f0-4093-80c5-e0229644834c\") " pod="openshift-marketplace/redhat-marketplace-zx9rg" Dec 06 06:18:32 crc kubenswrapper[4809]: I1206 06:18:32.482047 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4adfb97-c4f0-4093-80c5-e0229644834c-utilities\") pod \"redhat-marketplace-zx9rg\" (UID: \"d4adfb97-c4f0-4093-80c5-e0229644834c\") " pod="openshift-marketplace/redhat-marketplace-zx9rg" Dec 06 06:18:32 crc kubenswrapper[4809]: I1206 06:18:32.482130 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7s27g\" (UniqueName: \"kubernetes.io/projected/d4adfb97-c4f0-4093-80c5-e0229644834c-kube-api-access-7s27g\") pod \"redhat-marketplace-zx9rg\" (UID: \"d4adfb97-c4f0-4093-80c5-e0229644834c\") " pod="openshift-marketplace/redhat-marketplace-zx9rg" Dec 06 06:18:32 crc kubenswrapper[4809]: I1206 06:18:32.583756 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4adfb97-c4f0-4093-80c5-e0229644834c-catalog-content\") pod \"redhat-marketplace-zx9rg\" (UID: \"d4adfb97-c4f0-4093-80c5-e0229644834c\") " pod="openshift-marketplace/redhat-marketplace-zx9rg" Dec 06 06:18:32 crc kubenswrapper[4809]: I1206 06:18:32.583877 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4adfb97-c4f0-4093-80c5-e0229644834c-utilities\") pod \"redhat-marketplace-zx9rg\" (UID: \"d4adfb97-c4f0-4093-80c5-e0229644834c\") " pod="openshift-marketplace/redhat-marketplace-zx9rg" Dec 06 06:18:32 crc kubenswrapper[4809]: I1206 06:18:32.583978 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7s27g\" (UniqueName: \"kubernetes.io/projected/d4adfb97-c4f0-4093-80c5-e0229644834c-kube-api-access-7s27g\") pod \"redhat-marketplace-zx9rg\" (UID: \"d4adfb97-c4f0-4093-80c5-e0229644834c\") " pod="openshift-marketplace/redhat-marketplace-zx9rg" Dec 06 06:18:32 crc kubenswrapper[4809]: I1206 06:18:32.586892 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4adfb97-c4f0-4093-80c5-e0229644834c-catalog-content\") pod \"redhat-marketplace-zx9rg\" (UID: \"d4adfb97-c4f0-4093-80c5-e0229644834c\") " pod="openshift-marketplace/redhat-marketplace-zx9rg" Dec 06 06:18:32 crc kubenswrapper[4809]: I1206 06:18:32.587136 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4adfb97-c4f0-4093-80c5-e0229644834c-utilities\") pod \"redhat-marketplace-zx9rg\" (UID: \"d4adfb97-c4f0-4093-80c5-e0229644834c\") " pod="openshift-marketplace/redhat-marketplace-zx9rg" Dec 06 06:18:32 crc kubenswrapper[4809]: I1206 06:18:32.625025 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7s27g\" (UniqueName: \"kubernetes.io/projected/d4adfb97-c4f0-4093-80c5-e0229644834c-kube-api-access-7s27g\") pod \"redhat-marketplace-zx9rg\" (UID: \"d4adfb97-c4f0-4093-80c5-e0229644834c\") " pod="openshift-marketplace/redhat-marketplace-zx9rg" Dec 06 06:18:32 crc kubenswrapper[4809]: I1206 06:18:32.663776 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zx9rg" Dec 06 06:18:34 crc kubenswrapper[4809]: I1206 06:18:34.496213 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:18:34 crc kubenswrapper[4809]: I1206 06:18:34.496839 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:18:34 crc kubenswrapper[4809]: I1206 06:18:34.496894 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-npms2" Dec 06 06:18:34 crc kubenswrapper[4809]: I1206 06:18:34.497857 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9bc2dc88b3b9f309ce0169aa510326dd77022dece4c69cab3141df00fe5c0106"} pod="openshift-machine-config-operator/machine-config-daemon-npms2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 06:18:34 crc kubenswrapper[4809]: I1206 06:18:34.497925 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" containerID="cri-o://9bc2dc88b3b9f309ce0169aa510326dd77022dece4c69cab3141df00fe5c0106" gracePeriod=600 Dec 06 06:18:34 crc kubenswrapper[4809]: I1206 06:18:34.942096 4809 generic.go:334] "Generic (PLEG): container finished" podID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerID="9bc2dc88b3b9f309ce0169aa510326dd77022dece4c69cab3141df00fe5c0106" exitCode=0 Dec 06 06:18:34 crc kubenswrapper[4809]: I1206 06:18:34.942147 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" event={"ID":"cbf4cf62-024e-4703-a8b8-9aecda9cd26a","Type":"ContainerDied","Data":"9bc2dc88b3b9f309ce0169aa510326dd77022dece4c69cab3141df00fe5c0106"} Dec 06 06:18:34 crc kubenswrapper[4809]: I1206 06:18:34.942190 4809 scope.go:117] "RemoveContainer" containerID="fac0817a1ba54b752b764ab02ba5dbd6db50ecfcf2de9f18d391e8a61207257c" Dec 06 06:18:39 crc kubenswrapper[4809]: E1206 06:18:39.733499 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:18:40 crc kubenswrapper[4809]: I1206 06:18:40.030503 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:18:40 crc kubenswrapper[4809]: I1206 06:18:40.034370 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"53615cf8-5ea5-4d28-a4e7-1491e2b33744","Type":"ContainerDied","Data":"d3427867fc1f8bf40890103584f0e3c254bffa7e0aae8bc3ccceb041647e1727"} Dec 06 06:18:40 crc kubenswrapper[4809]: I1206 06:18:40.034455 4809 scope.go:117] "RemoveContainer" containerID="469ecb481d66fc349315f13547379057606da423f7767718d0b6694ad4cf95a3" Dec 06 06:18:40 crc kubenswrapper[4809]: I1206 06:18:40.034866 4809 scope.go:117] "RemoveContainer" containerID="9bc2dc88b3b9f309ce0169aa510326dd77022dece4c69cab3141df00fe5c0106" Dec 06 06:18:40 crc kubenswrapper[4809]: E1206 06:18:40.035136 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:18:40 crc kubenswrapper[4809]: I1206 06:18:40.077775 4809 scope.go:117] "RemoveContainer" containerID="89b31f5e6af406027c0bbe907db9671a164bbfd100c01beee0a9a23b402a0893" Dec 06 06:18:40 crc kubenswrapper[4809]: I1206 06:18:40.149153 4809 scope.go:117] "RemoveContainer" containerID="999d97cab57c383062b33ab0f30fd1a595134ddd3aaef5109e965c3b0cca4b63" Dec 06 06:18:40 crc kubenswrapper[4809]: I1206 06:18:40.159691 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53615cf8-5ea5-4d28-a4e7-1491e2b33744-scripts\") pod \"53615cf8-5ea5-4d28-a4e7-1491e2b33744\" (UID: \"53615cf8-5ea5-4d28-a4e7-1491e2b33744\") " Dec 06 06:18:40 crc kubenswrapper[4809]: I1206 06:18:40.159790 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53615cf8-5ea5-4d28-a4e7-1491e2b33744-run-httpd\") pod \"53615cf8-5ea5-4d28-a4e7-1491e2b33744\" (UID: \"53615cf8-5ea5-4d28-a4e7-1491e2b33744\") " Dec 06 06:18:40 crc kubenswrapper[4809]: I1206 06:18:40.159826 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/53615cf8-5ea5-4d28-a4e7-1491e2b33744-sg-core-conf-yaml\") pod \"53615cf8-5ea5-4d28-a4e7-1491e2b33744\" (UID: \"53615cf8-5ea5-4d28-a4e7-1491e2b33744\") " Dec 06 06:18:40 crc kubenswrapper[4809]: I1206 06:18:40.160015 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53615cf8-5ea5-4d28-a4e7-1491e2b33744-combined-ca-bundle\") pod \"53615cf8-5ea5-4d28-a4e7-1491e2b33744\" (UID: \"53615cf8-5ea5-4d28-a4e7-1491e2b33744\") " Dec 06 06:18:40 crc kubenswrapper[4809]: I1206 06:18:40.160115 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53615cf8-5ea5-4d28-a4e7-1491e2b33744-config-data\") pod \"53615cf8-5ea5-4d28-a4e7-1491e2b33744\" (UID: \"53615cf8-5ea5-4d28-a4e7-1491e2b33744\") " Dec 06 06:18:40 crc kubenswrapper[4809]: I1206 06:18:40.160349 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53615cf8-5ea5-4d28-a4e7-1491e2b33744-log-httpd\") pod \"53615cf8-5ea5-4d28-a4e7-1491e2b33744\" (UID: \"53615cf8-5ea5-4d28-a4e7-1491e2b33744\") " Dec 06 06:18:40 crc kubenswrapper[4809]: I1206 06:18:40.160404 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pskh5\" (UniqueName: \"kubernetes.io/projected/53615cf8-5ea5-4d28-a4e7-1491e2b33744-kube-api-access-pskh5\") pod \"53615cf8-5ea5-4d28-a4e7-1491e2b33744\" (UID: \"53615cf8-5ea5-4d28-a4e7-1491e2b33744\") " Dec 06 06:18:40 crc kubenswrapper[4809]: I1206 06:18:40.162489 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53615cf8-5ea5-4d28-a4e7-1491e2b33744-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "53615cf8-5ea5-4d28-a4e7-1491e2b33744" (UID: "53615cf8-5ea5-4d28-a4e7-1491e2b33744"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:18:40 crc kubenswrapper[4809]: I1206 06:18:40.162643 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53615cf8-5ea5-4d28-a4e7-1491e2b33744-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "53615cf8-5ea5-4d28-a4e7-1491e2b33744" (UID: "53615cf8-5ea5-4d28-a4e7-1491e2b33744"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:18:40 crc kubenswrapper[4809]: I1206 06:18:40.169687 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53615cf8-5ea5-4d28-a4e7-1491e2b33744-scripts" (OuterVolumeSpecName: "scripts") pod "53615cf8-5ea5-4d28-a4e7-1491e2b33744" (UID: "53615cf8-5ea5-4d28-a4e7-1491e2b33744"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:18:40 crc kubenswrapper[4809]: I1206 06:18:40.190217 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53615cf8-5ea5-4d28-a4e7-1491e2b33744-kube-api-access-pskh5" (OuterVolumeSpecName: "kube-api-access-pskh5") pod "53615cf8-5ea5-4d28-a4e7-1491e2b33744" (UID: "53615cf8-5ea5-4d28-a4e7-1491e2b33744"). InnerVolumeSpecName "kube-api-access-pskh5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:18:40 crc kubenswrapper[4809]: I1206 06:18:40.227067 4809 scope.go:117] "RemoveContainer" containerID="9dc198feff550848f012dfdf8ff0459b0a049a7da9c085bb6ac08abdb3e3c283" Dec 06 06:18:40 crc kubenswrapper[4809]: I1206 06:18:40.228963 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53615cf8-5ea5-4d28-a4e7-1491e2b33744-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "53615cf8-5ea5-4d28-a4e7-1491e2b33744" (UID: "53615cf8-5ea5-4d28-a4e7-1491e2b33744"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:18:40 crc kubenswrapper[4809]: I1206 06:18:40.267924 4809 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53615cf8-5ea5-4d28-a4e7-1491e2b33744-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:40 crc kubenswrapper[4809]: I1206 06:18:40.267974 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pskh5\" (UniqueName: \"kubernetes.io/projected/53615cf8-5ea5-4d28-a4e7-1491e2b33744-kube-api-access-pskh5\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:40 crc kubenswrapper[4809]: I1206 06:18:40.267990 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53615cf8-5ea5-4d28-a4e7-1491e2b33744-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:40 crc kubenswrapper[4809]: I1206 06:18:40.268000 4809 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53615cf8-5ea5-4d28-a4e7-1491e2b33744-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:40 crc kubenswrapper[4809]: I1206 06:18:40.268010 4809 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/53615cf8-5ea5-4d28-a4e7-1491e2b33744-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:40 crc kubenswrapper[4809]: I1206 06:18:40.326073 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zx9rg"] Dec 06 06:18:40 crc kubenswrapper[4809]: I1206 06:18:40.380699 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53615cf8-5ea5-4d28-a4e7-1491e2b33744-config-data" (OuterVolumeSpecName: "config-data") pod "53615cf8-5ea5-4d28-a4e7-1491e2b33744" (UID: "53615cf8-5ea5-4d28-a4e7-1491e2b33744"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:18:40 crc kubenswrapper[4809]: I1206 06:18:40.383667 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53615cf8-5ea5-4d28-a4e7-1491e2b33744-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "53615cf8-5ea5-4d28-a4e7-1491e2b33744" (UID: "53615cf8-5ea5-4d28-a4e7-1491e2b33744"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:18:40 crc kubenswrapper[4809]: I1206 06:18:40.471598 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53615cf8-5ea5-4d28-a4e7-1491e2b33744-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:40 crc kubenswrapper[4809]: I1206 06:18:40.471730 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53615cf8-5ea5-4d28-a4e7-1491e2b33744-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:40 crc kubenswrapper[4809]: I1206 06:18:40.811920 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 06:18:40 crc kubenswrapper[4809]: I1206 06:18:40.812503 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="dc087d56-1746-4ae2-a3d1-7a44f3084390" containerName="glance-log" containerID="cri-o://eceb1dd6a956119561296e96f6dd57c1a36fc552142dab032ead78002d98f409" gracePeriod=30 Dec 06 06:18:40 crc kubenswrapper[4809]: I1206 06:18:40.812656 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="dc087d56-1746-4ae2-a3d1-7a44f3084390" containerName="glance-httpd" containerID="cri-o://76b82b56f3e1173ce4aabac3674d4645c6199e126e35b484ef68d60197933ced" gracePeriod=30 Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.047557 4809 generic.go:334] "Generic (PLEG): container finished" podID="dc087d56-1746-4ae2-a3d1-7a44f3084390" containerID="eceb1dd6a956119561296e96f6dd57c1a36fc552142dab032ead78002d98f409" exitCode=143 Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.047650 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dc087d56-1746-4ae2-a3d1-7a44f3084390","Type":"ContainerDied","Data":"eceb1dd6a956119561296e96f6dd57c1a36fc552142dab032ead78002d98f409"} Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.049964 4809 generic.go:334] "Generic (PLEG): container finished" podID="d4adfb97-c4f0-4093-80c5-e0229644834c" containerID="7c94743d940eabfd47a0a92465458b1b3d9348c604c3a433c1da4fe15ed1eb61" exitCode=0 Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.050012 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zx9rg" event={"ID":"d4adfb97-c4f0-4093-80c5-e0229644834c","Type":"ContainerDied","Data":"7c94743d940eabfd47a0a92465458b1b3d9348c604c3a433c1da4fe15ed1eb61"} Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.050089 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zx9rg" event={"ID":"d4adfb97-c4f0-4093-80c5-e0229644834c","Type":"ContainerStarted","Data":"f101c865677027413942eb36929ac28a4f61696de2f786ee882cb7c0ce2f4bc5"} Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.053613 4809 generic.go:334] "Generic (PLEG): container finished" podID="d97f700e-65e3-41e3-a053-a9e97bd9857f" containerID="71495a991e495292715145b998352a921331f8b2e67b9803873b6ba262ecd42d" exitCode=0 Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.054641 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6pcg" event={"ID":"d97f700e-65e3-41e3-a053-a9e97bd9857f","Type":"ContainerDied","Data":"71495a991e495292715145b998352a921331f8b2e67b9803873b6ba262ecd42d"} Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.060528 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-vt99m" event={"ID":"13e0242d-c42e-4aad-95a2-62de95d87e42","Type":"ContainerStarted","Data":"2a58fd22495f3af98ef23bc24a54ae7cfc770faff52f1f4baae10ae5065326b0"} Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.062912 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.100427 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-vt99m" podStartSLOduration=2.75283116 podStartE2EDuration="15.100410492s" podCreationTimestamp="2025-12-06 06:18:26 +0000 UTC" firstStartedPulling="2025-12-06 06:18:27.499260639 +0000 UTC m=+1632.388243591" lastFinishedPulling="2025-12-06 06:18:39.846839981 +0000 UTC m=+1644.735822923" observedRunningTime="2025-12-06 06:18:41.098464619 +0000 UTC m=+1645.987447551" watchObservedRunningTime="2025-12-06 06:18:41.100410492 +0000 UTC m=+1645.989393434" Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.157216 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.224273 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.234599 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:18:41 crc kubenswrapper[4809]: E1206 06:18:41.235580 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53615cf8-5ea5-4d28-a4e7-1491e2b33744" containerName="ceilometer-central-agent" Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.235675 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="53615cf8-5ea5-4d28-a4e7-1491e2b33744" containerName="ceilometer-central-agent" Dec 06 06:18:41 crc kubenswrapper[4809]: E1206 06:18:41.235827 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53615cf8-5ea5-4d28-a4e7-1491e2b33744" containerName="ceilometer-notification-agent" Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.235902 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="53615cf8-5ea5-4d28-a4e7-1491e2b33744" containerName="ceilometer-notification-agent" Dec 06 06:18:41 crc kubenswrapper[4809]: E1206 06:18:41.235991 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53615cf8-5ea5-4d28-a4e7-1491e2b33744" containerName="sg-core" Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.236058 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="53615cf8-5ea5-4d28-a4e7-1491e2b33744" containerName="sg-core" Dec 06 06:18:41 crc kubenswrapper[4809]: E1206 06:18:41.236148 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53615cf8-5ea5-4d28-a4e7-1491e2b33744" containerName="proxy-httpd" Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.236218 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="53615cf8-5ea5-4d28-a4e7-1491e2b33744" containerName="proxy-httpd" Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.236529 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="53615cf8-5ea5-4d28-a4e7-1491e2b33744" containerName="sg-core" Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.236614 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="53615cf8-5ea5-4d28-a4e7-1491e2b33744" containerName="ceilometer-notification-agent" Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.236668 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="53615cf8-5ea5-4d28-a4e7-1491e2b33744" containerName="proxy-httpd" Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.236739 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="53615cf8-5ea5-4d28-a4e7-1491e2b33744" containerName="ceilometer-central-agent" Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.239262 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.243168 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.243827 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.243861 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.392332 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/395a8d1c-6b2e-4499-9aec-dfb78bf272d8-run-httpd\") pod \"ceilometer-0\" (UID: \"395a8d1c-6b2e-4499-9aec-dfb78bf272d8\") " pod="openstack/ceilometer-0" Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.392437 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/395a8d1c-6b2e-4499-9aec-dfb78bf272d8-config-data\") pod \"ceilometer-0\" (UID: \"395a8d1c-6b2e-4499-9aec-dfb78bf272d8\") " pod="openstack/ceilometer-0" Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.392882 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/395a8d1c-6b2e-4499-9aec-dfb78bf272d8-log-httpd\") pod \"ceilometer-0\" (UID: \"395a8d1c-6b2e-4499-9aec-dfb78bf272d8\") " pod="openstack/ceilometer-0" Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.393400 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/395a8d1c-6b2e-4499-9aec-dfb78bf272d8-scripts\") pod \"ceilometer-0\" (UID: \"395a8d1c-6b2e-4499-9aec-dfb78bf272d8\") " pod="openstack/ceilometer-0" Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.393491 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chtzh\" (UniqueName: \"kubernetes.io/projected/395a8d1c-6b2e-4499-9aec-dfb78bf272d8-kube-api-access-chtzh\") pod \"ceilometer-0\" (UID: \"395a8d1c-6b2e-4499-9aec-dfb78bf272d8\") " pod="openstack/ceilometer-0" Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.393541 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/395a8d1c-6b2e-4499-9aec-dfb78bf272d8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"395a8d1c-6b2e-4499-9aec-dfb78bf272d8\") " pod="openstack/ceilometer-0" Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.393578 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/395a8d1c-6b2e-4499-9aec-dfb78bf272d8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"395a8d1c-6b2e-4499-9aec-dfb78bf272d8\") " pod="openstack/ceilometer-0" Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.403740 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53615cf8-5ea5-4d28-a4e7-1491e2b33744" path="/var/lib/kubelet/pods/53615cf8-5ea5-4d28-a4e7-1491e2b33744/volumes" Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.495333 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/395a8d1c-6b2e-4499-9aec-dfb78bf272d8-run-httpd\") pod \"ceilometer-0\" (UID: \"395a8d1c-6b2e-4499-9aec-dfb78bf272d8\") " pod="openstack/ceilometer-0" Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.495725 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/395a8d1c-6b2e-4499-9aec-dfb78bf272d8-config-data\") pod \"ceilometer-0\" (UID: \"395a8d1c-6b2e-4499-9aec-dfb78bf272d8\") " pod="openstack/ceilometer-0" Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.495828 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/395a8d1c-6b2e-4499-9aec-dfb78bf272d8-log-httpd\") pod \"ceilometer-0\" (UID: \"395a8d1c-6b2e-4499-9aec-dfb78bf272d8\") " pod="openstack/ceilometer-0" Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.495993 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/395a8d1c-6b2e-4499-9aec-dfb78bf272d8-scripts\") pod \"ceilometer-0\" (UID: \"395a8d1c-6b2e-4499-9aec-dfb78bf272d8\") " pod="openstack/ceilometer-0" Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.496034 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/395a8d1c-6b2e-4499-9aec-dfb78bf272d8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"395a8d1c-6b2e-4499-9aec-dfb78bf272d8\") " pod="openstack/ceilometer-0" Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.496060 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chtzh\" (UniqueName: \"kubernetes.io/projected/395a8d1c-6b2e-4499-9aec-dfb78bf272d8-kube-api-access-chtzh\") pod \"ceilometer-0\" (UID: \"395a8d1c-6b2e-4499-9aec-dfb78bf272d8\") " pod="openstack/ceilometer-0" Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.496085 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/395a8d1c-6b2e-4499-9aec-dfb78bf272d8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"395a8d1c-6b2e-4499-9aec-dfb78bf272d8\") " pod="openstack/ceilometer-0" Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.496292 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/395a8d1c-6b2e-4499-9aec-dfb78bf272d8-run-httpd\") pod \"ceilometer-0\" (UID: \"395a8d1c-6b2e-4499-9aec-dfb78bf272d8\") " pod="openstack/ceilometer-0" Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.499438 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/395a8d1c-6b2e-4499-9aec-dfb78bf272d8-log-httpd\") pod \"ceilometer-0\" (UID: \"395a8d1c-6b2e-4499-9aec-dfb78bf272d8\") " pod="openstack/ceilometer-0" Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.501317 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/395a8d1c-6b2e-4499-9aec-dfb78bf272d8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"395a8d1c-6b2e-4499-9aec-dfb78bf272d8\") " pod="openstack/ceilometer-0" Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.501909 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/395a8d1c-6b2e-4499-9aec-dfb78bf272d8-config-data\") pod \"ceilometer-0\" (UID: \"395a8d1c-6b2e-4499-9aec-dfb78bf272d8\") " pod="openstack/ceilometer-0" Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.502976 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/395a8d1c-6b2e-4499-9aec-dfb78bf272d8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"395a8d1c-6b2e-4499-9aec-dfb78bf272d8\") " pod="openstack/ceilometer-0" Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.505496 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/395a8d1c-6b2e-4499-9aec-dfb78bf272d8-scripts\") pod \"ceilometer-0\" (UID: \"395a8d1c-6b2e-4499-9aec-dfb78bf272d8\") " pod="openstack/ceilometer-0" Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.517169 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chtzh\" (UniqueName: \"kubernetes.io/projected/395a8d1c-6b2e-4499-9aec-dfb78bf272d8-kube-api-access-chtzh\") pod \"ceilometer-0\" (UID: \"395a8d1c-6b2e-4499-9aec-dfb78bf272d8\") " pod="openstack/ceilometer-0" Dec 06 06:18:41 crc kubenswrapper[4809]: I1206 06:18:41.555410 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:18:41 crc kubenswrapper[4809]: E1206 06:18:41.980618 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1b4497e7e985b3dcecad685e6cbc0fe4e3ac63269895ab1f86489595826907ef" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 06 06:18:41 crc kubenswrapper[4809]: E1206 06:18:41.983197 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1b4497e7e985b3dcecad685e6cbc0fe4e3ac63269895ab1f86489595826907ef" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 06 06:18:41 crc kubenswrapper[4809]: E1206 06:18:41.986823 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1b4497e7e985b3dcecad685e6cbc0fe4e3ac63269895ab1f86489595826907ef" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 06 06:18:41 crc kubenswrapper[4809]: E1206 06:18:41.986900 4809 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-7c94c94748-57b42" podUID="515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a" containerName="heat-engine" Dec 06 06:18:42 crc kubenswrapper[4809]: I1206 06:18:42.124978 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:18:43 crc kubenswrapper[4809]: I1206 06:18:43.004730 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 06:18:43 crc kubenswrapper[4809]: I1206 06:18:43.005226 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="c40b4907-61d2-4ec7-b611-cb01d399c8d0" containerName="glance-log" containerID="cri-o://e66f47a441764738daccb70723b5c733763a3cad98cdd4e042a369cc6b7bce32" gracePeriod=30 Dec 06 06:18:43 crc kubenswrapper[4809]: I1206 06:18:43.005764 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="c40b4907-61d2-4ec7-b611-cb01d399c8d0" containerName="glance-httpd" containerID="cri-o://328becf5ebf16384ad2244583014f70780407728394585b140c94f7a4e16e55b" gracePeriod=30 Dec 06 06:18:43 crc kubenswrapper[4809]: I1206 06:18:43.102383 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"395a8d1c-6b2e-4499-9aec-dfb78bf272d8","Type":"ContainerStarted","Data":"c2bca4f5af1027b0f259ca193346b4d777593b4bae812815d36d0377087ff047"} Dec 06 06:18:43 crc kubenswrapper[4809]: I1206 06:18:43.104903 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6pcg" event={"ID":"d97f700e-65e3-41e3-a053-a9e97bd9857f","Type":"ContainerStarted","Data":"01971dc44037bca1545b095e771917450d2253226bb1feb09ead2036525b64fd"} Dec 06 06:18:43 crc kubenswrapper[4809]: I1206 06:18:43.988543 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:18:45 crc kubenswrapper[4809]: I1206 06:18:45.133078 4809 generic.go:334] "Generic (PLEG): container finished" podID="c40b4907-61d2-4ec7-b611-cb01d399c8d0" containerID="e66f47a441764738daccb70723b5c733763a3cad98cdd4e042a369cc6b7bce32" exitCode=143 Dec 06 06:18:45 crc kubenswrapper[4809]: I1206 06:18:45.133145 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c40b4907-61d2-4ec7-b611-cb01d399c8d0","Type":"ContainerDied","Data":"e66f47a441764738daccb70723b5c733763a3cad98cdd4e042a369cc6b7bce32"} Dec 06 06:18:46 crc kubenswrapper[4809]: I1206 06:18:46.144672 4809 generic.go:334] "Generic (PLEG): container finished" podID="dc087d56-1746-4ae2-a3d1-7a44f3084390" containerID="76b82b56f3e1173ce4aabac3674d4645c6199e126e35b484ef68d60197933ced" exitCode=0 Dec 06 06:18:46 crc kubenswrapper[4809]: I1206 06:18:46.144728 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dc087d56-1746-4ae2-a3d1-7a44f3084390","Type":"ContainerDied","Data":"76b82b56f3e1173ce4aabac3674d4645c6199e126e35b484ef68d60197933ced"} Dec 06 06:18:46 crc kubenswrapper[4809]: I1206 06:18:46.165331 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-v6pcg" podStartSLOduration=7.489705389 podStartE2EDuration="19.165312511s" podCreationTimestamp="2025-12-06 06:18:27 +0000 UTC" firstStartedPulling="2025-12-06 06:18:29.872348324 +0000 UTC m=+1634.761331266" lastFinishedPulling="2025-12-06 06:18:41.547955446 +0000 UTC m=+1646.436938388" observedRunningTime="2025-12-06 06:18:46.161662162 +0000 UTC m=+1651.050645104" watchObservedRunningTime="2025-12-06 06:18:46.165312511 +0000 UTC m=+1651.054295453" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.035088 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.144719 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc087d56-1746-4ae2-a3d1-7a44f3084390-combined-ca-bundle\") pod \"dc087d56-1746-4ae2-a3d1-7a44f3084390\" (UID: \"dc087d56-1746-4ae2-a3d1-7a44f3084390\") " Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.144892 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc087d56-1746-4ae2-a3d1-7a44f3084390-logs\") pod \"dc087d56-1746-4ae2-a3d1-7a44f3084390\" (UID: \"dc087d56-1746-4ae2-a3d1-7a44f3084390\") " Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.145019 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"dc087d56-1746-4ae2-a3d1-7a44f3084390\" (UID: \"dc087d56-1746-4ae2-a3d1-7a44f3084390\") " Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.145086 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc087d56-1746-4ae2-a3d1-7a44f3084390-public-tls-certs\") pod \"dc087d56-1746-4ae2-a3d1-7a44f3084390\" (UID: \"dc087d56-1746-4ae2-a3d1-7a44f3084390\") " Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.145110 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dc087d56-1746-4ae2-a3d1-7a44f3084390-httpd-run\") pod \"dc087d56-1746-4ae2-a3d1-7a44f3084390\" (UID: \"dc087d56-1746-4ae2-a3d1-7a44f3084390\") " Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.145164 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc087d56-1746-4ae2-a3d1-7a44f3084390-scripts\") pod \"dc087d56-1746-4ae2-a3d1-7a44f3084390\" (UID: \"dc087d56-1746-4ae2-a3d1-7a44f3084390\") " Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.145263 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gthwl\" (UniqueName: \"kubernetes.io/projected/dc087d56-1746-4ae2-a3d1-7a44f3084390-kube-api-access-gthwl\") pod \"dc087d56-1746-4ae2-a3d1-7a44f3084390\" (UID: \"dc087d56-1746-4ae2-a3d1-7a44f3084390\") " Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.145334 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc087d56-1746-4ae2-a3d1-7a44f3084390-config-data\") pod \"dc087d56-1746-4ae2-a3d1-7a44f3084390\" (UID: \"dc087d56-1746-4ae2-a3d1-7a44f3084390\") " Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.148527 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc087d56-1746-4ae2-a3d1-7a44f3084390-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "dc087d56-1746-4ae2-a3d1-7a44f3084390" (UID: "dc087d56-1746-4ae2-a3d1-7a44f3084390"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.148868 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc087d56-1746-4ae2-a3d1-7a44f3084390-logs" (OuterVolumeSpecName: "logs") pod "dc087d56-1746-4ae2-a3d1-7a44f3084390" (UID: "dc087d56-1746-4ae2-a3d1-7a44f3084390"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.161343 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "dc087d56-1746-4ae2-a3d1-7a44f3084390" (UID: "dc087d56-1746-4ae2-a3d1-7a44f3084390"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.164254 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc087d56-1746-4ae2-a3d1-7a44f3084390-kube-api-access-gthwl" (OuterVolumeSpecName: "kube-api-access-gthwl") pod "dc087d56-1746-4ae2-a3d1-7a44f3084390" (UID: "dc087d56-1746-4ae2-a3d1-7a44f3084390"). InnerVolumeSpecName "kube-api-access-gthwl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.170043 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc087d56-1746-4ae2-a3d1-7a44f3084390-scripts" (OuterVolumeSpecName: "scripts") pod "dc087d56-1746-4ae2-a3d1-7a44f3084390" (UID: "dc087d56-1746-4ae2-a3d1-7a44f3084390"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.204327 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.204348 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dc087d56-1746-4ae2-a3d1-7a44f3084390","Type":"ContainerDied","Data":"aa0efffb726b2380651dccfca520dab8405c3b056961bbd1a8f7ffa1b06ef98f"} Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.204392 4809 scope.go:117] "RemoveContainer" containerID="76b82b56f3e1173ce4aabac3674d4645c6199e126e35b484ef68d60197933ced" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.219671 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc087d56-1746-4ae2-a3d1-7a44f3084390-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dc087d56-1746-4ae2-a3d1-7a44f3084390" (UID: "dc087d56-1746-4ae2-a3d1-7a44f3084390"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.219872 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zx9rg" event={"ID":"d4adfb97-c4f0-4093-80c5-e0229644834c","Type":"ContainerStarted","Data":"cbcfe2679090af74eefb076ffaa4516d695958b4b6944d99b689cd324c47184d"} Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.231341 4809 generic.go:334] "Generic (PLEG): container finished" podID="c40b4907-61d2-4ec7-b611-cb01d399c8d0" containerID="328becf5ebf16384ad2244583014f70780407728394585b140c94f7a4e16e55b" exitCode=0 Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.231393 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c40b4907-61d2-4ec7-b611-cb01d399c8d0","Type":"ContainerDied","Data":"328becf5ebf16384ad2244583014f70780407728394585b140c94f7a4e16e55b"} Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.248993 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc087d56-1746-4ae2-a3d1-7a44f3084390-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.249027 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc087d56-1746-4ae2-a3d1-7a44f3084390-logs\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.249059 4809 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.249068 4809 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dc087d56-1746-4ae2-a3d1-7a44f3084390-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.249078 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc087d56-1746-4ae2-a3d1-7a44f3084390-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.249087 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gthwl\" (UniqueName: \"kubernetes.io/projected/dc087d56-1746-4ae2-a3d1-7a44f3084390-kube-api-access-gthwl\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.260591 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc087d56-1746-4ae2-a3d1-7a44f3084390-config-data" (OuterVolumeSpecName: "config-data") pod "dc087d56-1746-4ae2-a3d1-7a44f3084390" (UID: "dc087d56-1746-4ae2-a3d1-7a44f3084390"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.283595 4809 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.322133 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc087d56-1746-4ae2-a3d1-7a44f3084390-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "dc087d56-1746-4ae2-a3d1-7a44f3084390" (UID: "dc087d56-1746-4ae2-a3d1-7a44f3084390"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.351349 4809 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.351380 4809 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc087d56-1746-4ae2-a3d1-7a44f3084390-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.351391 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc087d56-1746-4ae2-a3d1-7a44f3084390-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.538892 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.548881 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.731047 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 06:18:47 crc kubenswrapper[4809]: E1206 06:18:47.731633 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc087d56-1746-4ae2-a3d1-7a44f3084390" containerName="glance-httpd" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.731651 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc087d56-1746-4ae2-a3d1-7a44f3084390" containerName="glance-httpd" Dec 06 06:18:47 crc kubenswrapper[4809]: E1206 06:18:47.731673 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc087d56-1746-4ae2-a3d1-7a44f3084390" containerName="glance-log" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.731680 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc087d56-1746-4ae2-a3d1-7a44f3084390" containerName="glance-log" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.732030 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc087d56-1746-4ae2-a3d1-7a44f3084390" containerName="glance-httpd" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.732068 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc087d56-1746-4ae2-a3d1-7a44f3084390" containerName="glance-log" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.732994 4809 scope.go:117] "RemoveContainer" containerID="eceb1dd6a956119561296e96f6dd57c1a36fc552142dab032ead78002d98f409" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.733566 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.738829 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.739026 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.760014 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.861983 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/807d7e0c-6c12-442b-8618-305e15340e02-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"807d7e0c-6c12-442b-8618-305e15340e02\") " pod="openstack/glance-default-external-api-0" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.862071 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"807d7e0c-6c12-442b-8618-305e15340e02\") " pod="openstack/glance-default-external-api-0" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.862158 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/807d7e0c-6c12-442b-8618-305e15340e02-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"807d7e0c-6c12-442b-8618-305e15340e02\") " pod="openstack/glance-default-external-api-0" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.862190 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lc47n\" (UniqueName: \"kubernetes.io/projected/807d7e0c-6c12-442b-8618-305e15340e02-kube-api-access-lc47n\") pod \"glance-default-external-api-0\" (UID: \"807d7e0c-6c12-442b-8618-305e15340e02\") " pod="openstack/glance-default-external-api-0" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.862283 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/807d7e0c-6c12-442b-8618-305e15340e02-config-data\") pod \"glance-default-external-api-0\" (UID: \"807d7e0c-6c12-442b-8618-305e15340e02\") " pod="openstack/glance-default-external-api-0" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.862420 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/807d7e0c-6c12-442b-8618-305e15340e02-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"807d7e0c-6c12-442b-8618-305e15340e02\") " pod="openstack/glance-default-external-api-0" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.862724 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/807d7e0c-6c12-442b-8618-305e15340e02-logs\") pod \"glance-default-external-api-0\" (UID: \"807d7e0c-6c12-442b-8618-305e15340e02\") " pod="openstack/glance-default-external-api-0" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.862861 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/807d7e0c-6c12-442b-8618-305e15340e02-scripts\") pod \"glance-default-external-api-0\" (UID: \"807d7e0c-6c12-442b-8618-305e15340e02\") " pod="openstack/glance-default-external-api-0" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.964941 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/807d7e0c-6c12-442b-8618-305e15340e02-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"807d7e0c-6c12-442b-8618-305e15340e02\") " pod="openstack/glance-default-external-api-0" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.965002 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"807d7e0c-6c12-442b-8618-305e15340e02\") " pod="openstack/glance-default-external-api-0" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.965076 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/807d7e0c-6c12-442b-8618-305e15340e02-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"807d7e0c-6c12-442b-8618-305e15340e02\") " pod="openstack/glance-default-external-api-0" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.965098 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lc47n\" (UniqueName: \"kubernetes.io/projected/807d7e0c-6c12-442b-8618-305e15340e02-kube-api-access-lc47n\") pod \"glance-default-external-api-0\" (UID: \"807d7e0c-6c12-442b-8618-305e15340e02\") " pod="openstack/glance-default-external-api-0" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.965139 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/807d7e0c-6c12-442b-8618-305e15340e02-config-data\") pod \"glance-default-external-api-0\" (UID: \"807d7e0c-6c12-442b-8618-305e15340e02\") " pod="openstack/glance-default-external-api-0" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.965196 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/807d7e0c-6c12-442b-8618-305e15340e02-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"807d7e0c-6c12-442b-8618-305e15340e02\") " pod="openstack/glance-default-external-api-0" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.965261 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/807d7e0c-6c12-442b-8618-305e15340e02-logs\") pod \"glance-default-external-api-0\" (UID: \"807d7e0c-6c12-442b-8618-305e15340e02\") " pod="openstack/glance-default-external-api-0" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.965286 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/807d7e0c-6c12-442b-8618-305e15340e02-scripts\") pod \"glance-default-external-api-0\" (UID: \"807d7e0c-6c12-442b-8618-305e15340e02\") " pod="openstack/glance-default-external-api-0" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.966612 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"807d7e0c-6c12-442b-8618-305e15340e02\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-external-api-0" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.966800 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/807d7e0c-6c12-442b-8618-305e15340e02-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"807d7e0c-6c12-442b-8618-305e15340e02\") " pod="openstack/glance-default-external-api-0" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.966990 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/807d7e0c-6c12-442b-8618-305e15340e02-logs\") pod \"glance-default-external-api-0\" (UID: \"807d7e0c-6c12-442b-8618-305e15340e02\") " pod="openstack/glance-default-external-api-0" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.972788 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/807d7e0c-6c12-442b-8618-305e15340e02-scripts\") pod \"glance-default-external-api-0\" (UID: \"807d7e0c-6c12-442b-8618-305e15340e02\") " pod="openstack/glance-default-external-api-0" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.973821 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/807d7e0c-6c12-442b-8618-305e15340e02-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"807d7e0c-6c12-442b-8618-305e15340e02\") " pod="openstack/glance-default-external-api-0" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.975961 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/807d7e0c-6c12-442b-8618-305e15340e02-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"807d7e0c-6c12-442b-8618-305e15340e02\") " pod="openstack/glance-default-external-api-0" Dec 06 06:18:47 crc kubenswrapper[4809]: I1206 06:18:47.976477 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/807d7e0c-6c12-442b-8618-305e15340e02-config-data\") pod \"glance-default-external-api-0\" (UID: \"807d7e0c-6c12-442b-8618-305e15340e02\") " pod="openstack/glance-default-external-api-0" Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.016743 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lc47n\" (UniqueName: \"kubernetes.io/projected/807d7e0c-6c12-442b-8618-305e15340e02-kube-api-access-lc47n\") pod \"glance-default-external-api-0\" (UID: \"807d7e0c-6c12-442b-8618-305e15340e02\") " pod="openstack/glance-default-external-api-0" Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.017943 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.047180 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"807d7e0c-6c12-442b-8618-305e15340e02\") " pod="openstack/glance-default-external-api-0" Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.169283 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c40b4907-61d2-4ec7-b611-cb01d399c8d0-httpd-run\") pod \"c40b4907-61d2-4ec7-b611-cb01d399c8d0\" (UID: \"c40b4907-61d2-4ec7-b611-cb01d399c8d0\") " Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.169372 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c40b4907-61d2-4ec7-b611-cb01d399c8d0-config-data\") pod \"c40b4907-61d2-4ec7-b611-cb01d399c8d0\" (UID: \"c40b4907-61d2-4ec7-b611-cb01d399c8d0\") " Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.169474 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c40b4907-61d2-4ec7-b611-cb01d399c8d0-combined-ca-bundle\") pod \"c40b4907-61d2-4ec7-b611-cb01d399c8d0\" (UID: \"c40b4907-61d2-4ec7-b611-cb01d399c8d0\") " Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.169540 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jg5f7\" (UniqueName: \"kubernetes.io/projected/c40b4907-61d2-4ec7-b611-cb01d399c8d0-kube-api-access-jg5f7\") pod \"c40b4907-61d2-4ec7-b611-cb01d399c8d0\" (UID: \"c40b4907-61d2-4ec7-b611-cb01d399c8d0\") " Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.169718 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"c40b4907-61d2-4ec7-b611-cb01d399c8d0\" (UID: \"c40b4907-61d2-4ec7-b611-cb01d399c8d0\") " Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.169847 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c40b4907-61d2-4ec7-b611-cb01d399c8d0-logs\") pod \"c40b4907-61d2-4ec7-b611-cb01d399c8d0\" (UID: \"c40b4907-61d2-4ec7-b611-cb01d399c8d0\") " Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.169884 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c40b4907-61d2-4ec7-b611-cb01d399c8d0-internal-tls-certs\") pod \"c40b4907-61d2-4ec7-b611-cb01d399c8d0\" (UID: \"c40b4907-61d2-4ec7-b611-cb01d399c8d0\") " Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.169912 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c40b4907-61d2-4ec7-b611-cb01d399c8d0-scripts\") pod \"c40b4907-61d2-4ec7-b611-cb01d399c8d0\" (UID: \"c40b4907-61d2-4ec7-b611-cb01d399c8d0\") " Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.171745 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c40b4907-61d2-4ec7-b611-cb01d399c8d0-logs" (OuterVolumeSpecName: "logs") pod "c40b4907-61d2-4ec7-b611-cb01d399c8d0" (UID: "c40b4907-61d2-4ec7-b611-cb01d399c8d0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.172021 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c40b4907-61d2-4ec7-b611-cb01d399c8d0-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "c40b4907-61d2-4ec7-b611-cb01d399c8d0" (UID: "c40b4907-61d2-4ec7-b611-cb01d399c8d0"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.194813 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c40b4907-61d2-4ec7-b611-cb01d399c8d0-kube-api-access-jg5f7" (OuterVolumeSpecName: "kube-api-access-jg5f7") pod "c40b4907-61d2-4ec7-b611-cb01d399c8d0" (UID: "c40b4907-61d2-4ec7-b611-cb01d399c8d0"). InnerVolumeSpecName "kube-api-access-jg5f7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.200109 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "c40b4907-61d2-4ec7-b611-cb01d399c8d0" (UID: "c40b4907-61d2-4ec7-b611-cb01d399c8d0"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.219414 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c40b4907-61d2-4ec7-b611-cb01d399c8d0-scripts" (OuterVolumeSpecName: "scripts") pod "c40b4907-61d2-4ec7-b611-cb01d399c8d0" (UID: "c40b4907-61d2-4ec7-b611-cb01d399c8d0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.273815 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jg5f7\" (UniqueName: \"kubernetes.io/projected/c40b4907-61d2-4ec7-b611-cb01d399c8d0-kube-api-access-jg5f7\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.273890 4809 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.273919 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c40b4907-61d2-4ec7-b611-cb01d399c8d0-logs\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.273940 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c40b4907-61d2-4ec7-b611-cb01d399c8d0-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.273950 4809 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c40b4907-61d2-4ec7-b611-cb01d399c8d0-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.288406 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-v6pcg" Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.288447 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-v6pcg" Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.312266 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.359147 4809 generic.go:334] "Generic (PLEG): container finished" podID="d4adfb97-c4f0-4093-80c5-e0229644834c" containerID="cbcfe2679090af74eefb076ffaa4516d695958b4b6944d99b689cd324c47184d" exitCode=0 Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.359227 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zx9rg" event={"ID":"d4adfb97-c4f0-4093-80c5-e0229644834c","Type":"ContainerDied","Data":"cbcfe2679090af74eefb076ffaa4516d695958b4b6944d99b689cd324c47184d"} Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.365314 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c40b4907-61d2-4ec7-b611-cb01d399c8d0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c40b4907-61d2-4ec7-b611-cb01d399c8d0" (UID: "c40b4907-61d2-4ec7-b611-cb01d399c8d0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.376599 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c40b4907-61d2-4ec7-b611-cb01d399c8d0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.379705 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.380903 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c40b4907-61d2-4ec7-b611-cb01d399c8d0","Type":"ContainerDied","Data":"7487ebf3d482625a35a71ae604e34b91a246b0fef4380fb68114da4b9bf4c673"} Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.380981 4809 scope.go:117] "RemoveContainer" containerID="328becf5ebf16384ad2244583014f70780407728394585b140c94f7a4e16e55b" Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.404857 4809 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.412418 4809 generic.go:334] "Generic (PLEG): container finished" podID="515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a" containerID="1b4497e7e985b3dcecad685e6cbc0fe4e3ac63269895ab1f86489595826907ef" exitCode=0 Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.412469 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7c94c94748-57b42" event={"ID":"515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a","Type":"ContainerDied","Data":"1b4497e7e985b3dcecad685e6cbc0fe4e3ac63269895ab1f86489595826907ef"} Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.438431 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-v6pcg" Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.449477 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c40b4907-61d2-4ec7-b611-cb01d399c8d0-config-data" (OuterVolumeSpecName: "config-data") pod "c40b4907-61d2-4ec7-b611-cb01d399c8d0" (UID: "c40b4907-61d2-4ec7-b611-cb01d399c8d0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.465784 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c40b4907-61d2-4ec7-b611-cb01d399c8d0-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "c40b4907-61d2-4ec7-b611-cb01d399c8d0" (UID: "c40b4907-61d2-4ec7-b611-cb01d399c8d0"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.479683 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c40b4907-61d2-4ec7-b611-cb01d399c8d0-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.479721 4809 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.479730 4809 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c40b4907-61d2-4ec7-b611-cb01d399c8d0-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.558424 4809 scope.go:117] "RemoveContainer" containerID="e66f47a441764738daccb70723b5c733763a3cad98cdd4e042a369cc6b7bce32" Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.770049 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.793266 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.827013 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 06:18:48 crc kubenswrapper[4809]: E1206 06:18:48.827664 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c40b4907-61d2-4ec7-b611-cb01d399c8d0" containerName="glance-httpd" Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.827692 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c40b4907-61d2-4ec7-b611-cb01d399c8d0" containerName="glance-httpd" Dec 06 06:18:48 crc kubenswrapper[4809]: E1206 06:18:48.827719 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c40b4907-61d2-4ec7-b611-cb01d399c8d0" containerName="glance-log" Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.827728 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c40b4907-61d2-4ec7-b611-cb01d399c8d0" containerName="glance-log" Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.828170 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="c40b4907-61d2-4ec7-b611-cb01d399c8d0" containerName="glance-log" Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.828202 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="c40b4907-61d2-4ec7-b611-cb01d399c8d0" containerName="glance-httpd" Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.829861 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.839331 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.839508 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.866032 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.893251 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fdb3b6c-da8d-4493-b931-a55561dfb2ae-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0fdb3b6c-da8d-4493-b931-a55561dfb2ae\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.893365 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0fdb3b6c-da8d-4493-b931-a55561dfb2ae-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0fdb3b6c-da8d-4493-b931-a55561dfb2ae\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.893396 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mk4sr\" (UniqueName: \"kubernetes.io/projected/0fdb3b6c-da8d-4493-b931-a55561dfb2ae-kube-api-access-mk4sr\") pod \"glance-default-internal-api-0\" (UID: \"0fdb3b6c-da8d-4493-b931-a55561dfb2ae\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.893411 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fdb3b6c-da8d-4493-b931-a55561dfb2ae-logs\") pod \"glance-default-internal-api-0\" (UID: \"0fdb3b6c-da8d-4493-b931-a55561dfb2ae\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.893442 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"0fdb3b6c-da8d-4493-b931-a55561dfb2ae\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.893463 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0fdb3b6c-da8d-4493-b931-a55561dfb2ae-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"0fdb3b6c-da8d-4493-b931-a55561dfb2ae\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.893482 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fdb3b6c-da8d-4493-b931-a55561dfb2ae-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0fdb3b6c-da8d-4493-b931-a55561dfb2ae\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.893546 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fdb3b6c-da8d-4493-b931-a55561dfb2ae-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0fdb3b6c-da8d-4493-b931-a55561dfb2ae\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.921318 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7c94c94748-57b42" Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.997269 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-brq7t\" (UniqueName: \"kubernetes.io/projected/515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a-kube-api-access-brq7t\") pod \"515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a\" (UID: \"515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a\") " Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.997425 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a-config-data-custom\") pod \"515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a\" (UID: \"515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a\") " Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.997592 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a-combined-ca-bundle\") pod \"515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a\" (UID: \"515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a\") " Dec 06 06:18:48 crc kubenswrapper[4809]: I1206 06:18:48.997622 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a-config-data\") pod \"515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a\" (UID: \"515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a\") " Dec 06 06:18:49 crc kubenswrapper[4809]: I1206 06:18:49.002144 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0fdb3b6c-da8d-4493-b931-a55561dfb2ae-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0fdb3b6c-da8d-4493-b931-a55561dfb2ae\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:18:49 crc kubenswrapper[4809]: I1206 06:18:49.002237 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mk4sr\" (UniqueName: \"kubernetes.io/projected/0fdb3b6c-da8d-4493-b931-a55561dfb2ae-kube-api-access-mk4sr\") pod \"glance-default-internal-api-0\" (UID: \"0fdb3b6c-da8d-4493-b931-a55561dfb2ae\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:18:49 crc kubenswrapper[4809]: I1206 06:18:49.002270 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fdb3b6c-da8d-4493-b931-a55561dfb2ae-logs\") pod \"glance-default-internal-api-0\" (UID: \"0fdb3b6c-da8d-4493-b931-a55561dfb2ae\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:18:49 crc kubenswrapper[4809]: I1206 06:18:49.002360 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"0fdb3b6c-da8d-4493-b931-a55561dfb2ae\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:18:49 crc kubenswrapper[4809]: I1206 06:18:49.002407 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0fdb3b6c-da8d-4493-b931-a55561dfb2ae-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"0fdb3b6c-da8d-4493-b931-a55561dfb2ae\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:18:49 crc kubenswrapper[4809]: I1206 06:18:49.002437 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fdb3b6c-da8d-4493-b931-a55561dfb2ae-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0fdb3b6c-da8d-4493-b931-a55561dfb2ae\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:18:49 crc kubenswrapper[4809]: I1206 06:18:49.002609 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fdb3b6c-da8d-4493-b931-a55561dfb2ae-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0fdb3b6c-da8d-4493-b931-a55561dfb2ae\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:18:49 crc kubenswrapper[4809]: I1206 06:18:49.002799 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fdb3b6c-da8d-4493-b931-a55561dfb2ae-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0fdb3b6c-da8d-4493-b931-a55561dfb2ae\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:18:49 crc kubenswrapper[4809]: I1206 06:18:49.011766 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0fdb3b6c-da8d-4493-b931-a55561dfb2ae-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0fdb3b6c-da8d-4493-b931-a55561dfb2ae\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:18:49 crc kubenswrapper[4809]: I1206 06:18:49.015495 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fdb3b6c-da8d-4493-b931-a55561dfb2ae-logs\") pod \"glance-default-internal-api-0\" (UID: \"0fdb3b6c-da8d-4493-b931-a55561dfb2ae\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:18:49 crc kubenswrapper[4809]: I1206 06:18:49.015654 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"0fdb3b6c-da8d-4493-b931-a55561dfb2ae\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-internal-api-0" Dec 06 06:18:49 crc kubenswrapper[4809]: I1206 06:18:49.025554 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0fdb3b6c-da8d-4493-b931-a55561dfb2ae-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"0fdb3b6c-da8d-4493-b931-a55561dfb2ae\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:18:49 crc kubenswrapper[4809]: I1206 06:18:49.034466 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a-kube-api-access-brq7t" (OuterVolumeSpecName: "kube-api-access-brq7t") pod "515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a" (UID: "515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a"). InnerVolumeSpecName "kube-api-access-brq7t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:18:49 crc kubenswrapper[4809]: I1206 06:18:49.050020 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a" (UID: "515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:18:49 crc kubenswrapper[4809]: I1206 06:18:49.051353 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fdb3b6c-da8d-4493-b931-a55561dfb2ae-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0fdb3b6c-da8d-4493-b931-a55561dfb2ae\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:18:49 crc kubenswrapper[4809]: I1206 06:18:49.052740 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fdb3b6c-da8d-4493-b931-a55561dfb2ae-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0fdb3b6c-da8d-4493-b931-a55561dfb2ae\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:18:49 crc kubenswrapper[4809]: I1206 06:18:49.053787 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fdb3b6c-da8d-4493-b931-a55561dfb2ae-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0fdb3b6c-da8d-4493-b931-a55561dfb2ae\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:18:49 crc kubenswrapper[4809]: I1206 06:18:49.063777 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mk4sr\" (UniqueName: \"kubernetes.io/projected/0fdb3b6c-da8d-4493-b931-a55561dfb2ae-kube-api-access-mk4sr\") pod \"glance-default-internal-api-0\" (UID: \"0fdb3b6c-da8d-4493-b931-a55561dfb2ae\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:18:49 crc kubenswrapper[4809]: I1206 06:18:49.108415 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-brq7t\" (UniqueName: \"kubernetes.io/projected/515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a-kube-api-access-brq7t\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:49 crc kubenswrapper[4809]: I1206 06:18:49.108450 4809 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:49 crc kubenswrapper[4809]: I1206 06:18:49.118648 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"0fdb3b6c-da8d-4493-b931-a55561dfb2ae\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:18:49 crc kubenswrapper[4809]: I1206 06:18:49.166907 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 06:18:49 crc kubenswrapper[4809]: I1206 06:18:49.178083 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a" (UID: "515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:18:49 crc kubenswrapper[4809]: I1206 06:18:49.211874 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:49 crc kubenswrapper[4809]: I1206 06:18:49.215917 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 06:18:49 crc kubenswrapper[4809]: I1206 06:18:49.236919 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a-config-data" (OuterVolumeSpecName: "config-data") pod "515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a" (UID: "515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:18:49 crc kubenswrapper[4809]: I1206 06:18:49.315019 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:49 crc kubenswrapper[4809]: I1206 06:18:49.421431 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c40b4907-61d2-4ec7-b611-cb01d399c8d0" path="/var/lib/kubelet/pods/c40b4907-61d2-4ec7-b611-cb01d399c8d0/volumes" Dec 06 06:18:49 crc kubenswrapper[4809]: I1206 06:18:49.451901 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7c94c94748-57b42" Dec 06 06:18:49 crc kubenswrapper[4809]: I1206 06:18:49.453827 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc087d56-1746-4ae2-a3d1-7a44f3084390" path="/var/lib/kubelet/pods/dc087d56-1746-4ae2-a3d1-7a44f3084390/volumes" Dec 06 06:18:49 crc kubenswrapper[4809]: I1206 06:18:49.514462 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7c94c94748-57b42" event={"ID":"515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a","Type":"ContainerDied","Data":"b0902d6876565ea7f2bc92882c1c0668113eb3beb5b752b48269e172803040a3"} Dec 06 06:18:49 crc kubenswrapper[4809]: I1206 06:18:49.514532 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"807d7e0c-6c12-442b-8618-305e15340e02","Type":"ContainerStarted","Data":"10d32b1fe329683dc8e687c574bafffab61a8ba8c5ba17bdb3854989b9a5b04c"} Dec 06 06:18:49 crc kubenswrapper[4809]: I1206 06:18:49.514552 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zx9rg" event={"ID":"d4adfb97-c4f0-4093-80c5-e0229644834c","Type":"ContainerStarted","Data":"518a0ed359ea197b4316903bbfd2d5ac37c11c7bc98ce51fc9b8ba3c39154423"} Dec 06 06:18:49 crc kubenswrapper[4809]: I1206 06:18:49.514575 4809 scope.go:117] "RemoveContainer" containerID="1b4497e7e985b3dcecad685e6cbc0fe4e3ac63269895ab1f86489595826907ef" Dec 06 06:18:49 crc kubenswrapper[4809]: I1206 06:18:49.522755 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"395a8d1c-6b2e-4499-9aec-dfb78bf272d8","Type":"ContainerStarted","Data":"adaff5f7480aeb10e1464676ffcb994398fc022e3838b6b67c975f6f73eb3824"} Dec 06 06:18:49 crc kubenswrapper[4809]: I1206 06:18:49.538054 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zx9rg" podStartSLOduration=9.693828072 podStartE2EDuration="17.537635974s" podCreationTimestamp="2025-12-06 06:18:32 +0000 UTC" firstStartedPulling="2025-12-06 06:18:41.051752089 +0000 UTC m=+1645.940735031" lastFinishedPulling="2025-12-06 06:18:48.895559991 +0000 UTC m=+1653.784542933" observedRunningTime="2025-12-06 06:18:49.527653754 +0000 UTC m=+1654.416636696" watchObservedRunningTime="2025-12-06 06:18:49.537635974 +0000 UTC m=+1654.426618916" Dec 06 06:18:49 crc kubenswrapper[4809]: I1206 06:18:49.591484 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-v6pcg" Dec 06 06:18:49 crc kubenswrapper[4809]: I1206 06:18:49.958952 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 06:18:50 crc kubenswrapper[4809]: I1206 06:18:50.539141 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"395a8d1c-6b2e-4499-9aec-dfb78bf272d8","Type":"ContainerStarted","Data":"73dd23b14e07077f2f91b598925c47338403483b252f273fab0f0c549ff24fd0"} Dec 06 06:18:50 crc kubenswrapper[4809]: I1206 06:18:50.539789 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"395a8d1c-6b2e-4499-9aec-dfb78bf272d8","Type":"ContainerStarted","Data":"0e23cc297bbd882a9475ed237aa11168d54243e3db05dd882af4da2f6244a6a1"} Dec 06 06:18:50 crc kubenswrapper[4809]: I1206 06:18:50.544105 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"807d7e0c-6c12-442b-8618-305e15340e02","Type":"ContainerStarted","Data":"cf377d6ed7ae29ec78a07e4cab27d2a05fac06483627ac80701a84725ec530ef"} Dec 06 06:18:50 crc kubenswrapper[4809]: I1206 06:18:50.547476 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0fdb3b6c-da8d-4493-b931-a55561dfb2ae","Type":"ContainerStarted","Data":"7d78e7bf986713a690e266ebbdce46cedfb4e588b634329d998680d429a68c23"} Dec 06 06:18:50 crc kubenswrapper[4809]: I1206 06:18:50.608572 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-v6pcg"] Dec 06 06:18:51 crc kubenswrapper[4809]: I1206 06:18:51.559453 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0fdb3b6c-da8d-4493-b931-a55561dfb2ae","Type":"ContainerStarted","Data":"ee5144af46ff972c57b0682dc75e0b1cbabe15a11c2ce821b26ff0f1de33e1e9"} Dec 06 06:18:51 crc kubenswrapper[4809]: I1206 06:18:51.562517 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"807d7e0c-6c12-442b-8618-305e15340e02","Type":"ContainerStarted","Data":"1d9819bee3da9d84c7bd71ac681302160f05c0bbc802fc840e20a1f770c7a112"} Dec 06 06:18:51 crc kubenswrapper[4809]: I1206 06:18:51.564573 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-v6pcg" podUID="d97f700e-65e3-41e3-a053-a9e97bd9857f" containerName="registry-server" containerID="cri-o://01971dc44037bca1545b095e771917450d2253226bb1feb09ead2036525b64fd" gracePeriod=2 Dec 06 06:18:51 crc kubenswrapper[4809]: I1206 06:18:51.589684 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.589440571 podStartE2EDuration="4.589440571s" podCreationTimestamp="2025-12-06 06:18:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:18:51.588443144 +0000 UTC m=+1656.477426086" watchObservedRunningTime="2025-12-06 06:18:51.589440571 +0000 UTC m=+1656.478423523" Dec 06 06:18:52 crc kubenswrapper[4809]: I1206 06:18:52.575820 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0fdb3b6c-da8d-4493-b931-a55561dfb2ae","Type":"ContainerStarted","Data":"dabe234a96b5c022185ca873f0667325dfd13d41ab07be16e20089d251c44a2f"} Dec 06 06:18:52 crc kubenswrapper[4809]: I1206 06:18:52.578648 4809 generic.go:334] "Generic (PLEG): container finished" podID="d97f700e-65e3-41e3-a053-a9e97bd9857f" containerID="01971dc44037bca1545b095e771917450d2253226bb1feb09ead2036525b64fd" exitCode=0 Dec 06 06:18:52 crc kubenswrapper[4809]: I1206 06:18:52.578718 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6pcg" event={"ID":"d97f700e-65e3-41e3-a053-a9e97bd9857f","Type":"ContainerDied","Data":"01971dc44037bca1545b095e771917450d2253226bb1feb09ead2036525b64fd"} Dec 06 06:18:52 crc kubenswrapper[4809]: I1206 06:18:52.610869 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.610853877 podStartE2EDuration="4.610853877s" podCreationTimestamp="2025-12-06 06:18:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:18:52.600086027 +0000 UTC m=+1657.489068979" watchObservedRunningTime="2025-12-06 06:18:52.610853877 +0000 UTC m=+1657.499836819" Dec 06 06:18:52 crc kubenswrapper[4809]: I1206 06:18:52.664275 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zx9rg" Dec 06 06:18:52 crc kubenswrapper[4809]: I1206 06:18:52.664331 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zx9rg" Dec 06 06:18:52 crc kubenswrapper[4809]: I1206 06:18:52.713657 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zx9rg" Dec 06 06:18:54 crc kubenswrapper[4809]: I1206 06:18:54.188501 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v6pcg" Dec 06 06:18:54 crc kubenswrapper[4809]: I1206 06:18:54.356566 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d97f700e-65e3-41e3-a053-a9e97bd9857f-catalog-content\") pod \"d97f700e-65e3-41e3-a053-a9e97bd9857f\" (UID: \"d97f700e-65e3-41e3-a053-a9e97bd9857f\") " Dec 06 06:18:54 crc kubenswrapper[4809]: I1206 06:18:54.356830 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bwr5s\" (UniqueName: \"kubernetes.io/projected/d97f700e-65e3-41e3-a053-a9e97bd9857f-kube-api-access-bwr5s\") pod \"d97f700e-65e3-41e3-a053-a9e97bd9857f\" (UID: \"d97f700e-65e3-41e3-a053-a9e97bd9857f\") " Dec 06 06:18:54 crc kubenswrapper[4809]: I1206 06:18:54.357070 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d97f700e-65e3-41e3-a053-a9e97bd9857f-utilities\") pod \"d97f700e-65e3-41e3-a053-a9e97bd9857f\" (UID: \"d97f700e-65e3-41e3-a053-a9e97bd9857f\") " Dec 06 06:18:54 crc kubenswrapper[4809]: I1206 06:18:54.359362 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d97f700e-65e3-41e3-a053-a9e97bd9857f-utilities" (OuterVolumeSpecName: "utilities") pod "d97f700e-65e3-41e3-a053-a9e97bd9857f" (UID: "d97f700e-65e3-41e3-a053-a9e97bd9857f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:18:54 crc kubenswrapper[4809]: I1206 06:18:54.365797 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d97f700e-65e3-41e3-a053-a9e97bd9857f-kube-api-access-bwr5s" (OuterVolumeSpecName: "kube-api-access-bwr5s") pod "d97f700e-65e3-41e3-a053-a9e97bd9857f" (UID: "d97f700e-65e3-41e3-a053-a9e97bd9857f"). InnerVolumeSpecName "kube-api-access-bwr5s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:18:54 crc kubenswrapper[4809]: I1206 06:18:54.388917 4809 scope.go:117] "RemoveContainer" containerID="9bc2dc88b3b9f309ce0169aa510326dd77022dece4c69cab3141df00fe5c0106" Dec 06 06:18:54 crc kubenswrapper[4809]: E1206 06:18:54.389333 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:18:54 crc kubenswrapper[4809]: I1206 06:18:54.421646 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d97f700e-65e3-41e3-a053-a9e97bd9857f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d97f700e-65e3-41e3-a053-a9e97bd9857f" (UID: "d97f700e-65e3-41e3-a053-a9e97bd9857f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:18:54 crc kubenswrapper[4809]: I1206 06:18:54.460635 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bwr5s\" (UniqueName: \"kubernetes.io/projected/d97f700e-65e3-41e3-a053-a9e97bd9857f-kube-api-access-bwr5s\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:54 crc kubenswrapper[4809]: I1206 06:18:54.460671 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d97f700e-65e3-41e3-a053-a9e97bd9857f-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:54 crc kubenswrapper[4809]: I1206 06:18:54.460680 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d97f700e-65e3-41e3-a053-a9e97bd9857f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:54 crc kubenswrapper[4809]: I1206 06:18:54.605236 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6pcg" event={"ID":"d97f700e-65e3-41e3-a053-a9e97bd9857f","Type":"ContainerDied","Data":"a9dca06c39aa1bc14016f04104c55bdb8599488fe076e01ed59cdb8a2e4f6184"} Dec 06 06:18:54 crc kubenswrapper[4809]: I1206 06:18:54.605340 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v6pcg" Dec 06 06:18:54 crc kubenswrapper[4809]: I1206 06:18:54.605646 4809 scope.go:117] "RemoveContainer" containerID="01971dc44037bca1545b095e771917450d2253226bb1feb09ead2036525b64fd" Dec 06 06:18:54 crc kubenswrapper[4809]: I1206 06:18:54.655873 4809 scope.go:117] "RemoveContainer" containerID="71495a991e495292715145b998352a921331f8b2e67b9803873b6ba262ecd42d" Dec 06 06:18:54 crc kubenswrapper[4809]: I1206 06:18:54.667972 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-v6pcg"] Dec 06 06:18:54 crc kubenswrapper[4809]: I1206 06:18:54.679163 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-v6pcg"] Dec 06 06:18:54 crc kubenswrapper[4809]: I1206 06:18:54.682021 4809 scope.go:117] "RemoveContainer" containerID="51eff57acd4a5b48273d67a5859417435e2ab837f35dcb56b80ea46a0e230ef4" Dec 06 06:18:55 crc kubenswrapper[4809]: I1206 06:18:55.405350 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d97f700e-65e3-41e3-a053-a9e97bd9857f" path="/var/lib/kubelet/pods/d97f700e-65e3-41e3-a053-a9e97bd9857f/volumes" Dec 06 06:18:57 crc kubenswrapper[4809]: I1206 06:18:57.644561 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"395a8d1c-6b2e-4499-9aec-dfb78bf272d8","Type":"ContainerStarted","Data":"8fca81cf46ee4798615c86c02e502e11f56f7b52f462cc24b53865a56ca46131"} Dec 06 06:18:57 crc kubenswrapper[4809]: I1206 06:18:57.645154 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 06:18:57 crc kubenswrapper[4809]: I1206 06:18:57.644799 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="395a8d1c-6b2e-4499-9aec-dfb78bf272d8" containerName="proxy-httpd" containerID="cri-o://8fca81cf46ee4798615c86c02e502e11f56f7b52f462cc24b53865a56ca46131" gracePeriod=30 Dec 06 06:18:57 crc kubenswrapper[4809]: I1206 06:18:57.644781 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="395a8d1c-6b2e-4499-9aec-dfb78bf272d8" containerName="ceilometer-central-agent" containerID="cri-o://adaff5f7480aeb10e1464676ffcb994398fc022e3838b6b67c975f6f73eb3824" gracePeriod=30 Dec 06 06:18:57 crc kubenswrapper[4809]: I1206 06:18:57.644967 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="395a8d1c-6b2e-4499-9aec-dfb78bf272d8" containerName="ceilometer-notification-agent" containerID="cri-o://0e23cc297bbd882a9475ed237aa11168d54243e3db05dd882af4da2f6244a6a1" gracePeriod=30 Dec 06 06:18:57 crc kubenswrapper[4809]: I1206 06:18:57.644841 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="395a8d1c-6b2e-4499-9aec-dfb78bf272d8" containerName="sg-core" containerID="cri-o://73dd23b14e07077f2f91b598925c47338403483b252f273fab0f0c549ff24fd0" gracePeriod=30 Dec 06 06:18:57 crc kubenswrapper[4809]: I1206 06:18:57.678066 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.354234248 podStartE2EDuration="16.678039977s" podCreationTimestamp="2025-12-06 06:18:41 +0000 UTC" firstStartedPulling="2025-12-06 06:18:42.142738923 +0000 UTC m=+1647.031721855" lastFinishedPulling="2025-12-06 06:18:56.466544642 +0000 UTC m=+1661.355527584" observedRunningTime="2025-12-06 06:18:57.670448603 +0000 UTC m=+1662.559431555" watchObservedRunningTime="2025-12-06 06:18:57.678039977 +0000 UTC m=+1662.567022909" Dec 06 06:18:58 crc kubenswrapper[4809]: I1206 06:18:58.313905 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 06 06:18:58 crc kubenswrapper[4809]: I1206 06:18:58.314279 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 06 06:18:58 crc kubenswrapper[4809]: I1206 06:18:58.360307 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 06 06:18:58 crc kubenswrapper[4809]: I1206 06:18:58.368588 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 06 06:18:58 crc kubenswrapper[4809]: I1206 06:18:58.658511 4809 generic.go:334] "Generic (PLEG): container finished" podID="395a8d1c-6b2e-4499-9aec-dfb78bf272d8" containerID="8fca81cf46ee4798615c86c02e502e11f56f7b52f462cc24b53865a56ca46131" exitCode=0 Dec 06 06:18:58 crc kubenswrapper[4809]: I1206 06:18:58.658831 4809 generic.go:334] "Generic (PLEG): container finished" podID="395a8d1c-6b2e-4499-9aec-dfb78bf272d8" containerID="73dd23b14e07077f2f91b598925c47338403483b252f273fab0f0c549ff24fd0" exitCode=2 Dec 06 06:18:58 crc kubenswrapper[4809]: I1206 06:18:58.658592 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"395a8d1c-6b2e-4499-9aec-dfb78bf272d8","Type":"ContainerDied","Data":"8fca81cf46ee4798615c86c02e502e11f56f7b52f462cc24b53865a56ca46131"} Dec 06 06:18:58 crc kubenswrapper[4809]: I1206 06:18:58.658884 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"395a8d1c-6b2e-4499-9aec-dfb78bf272d8","Type":"ContainerDied","Data":"73dd23b14e07077f2f91b598925c47338403483b252f273fab0f0c549ff24fd0"} Dec 06 06:18:58 crc kubenswrapper[4809]: I1206 06:18:58.658900 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"395a8d1c-6b2e-4499-9aec-dfb78bf272d8","Type":"ContainerDied","Data":"0e23cc297bbd882a9475ed237aa11168d54243e3db05dd882af4da2f6244a6a1"} Dec 06 06:18:58 crc kubenswrapper[4809]: I1206 06:18:58.658840 4809 generic.go:334] "Generic (PLEG): container finished" podID="395a8d1c-6b2e-4499-9aec-dfb78bf272d8" containerID="0e23cc297bbd882a9475ed237aa11168d54243e3db05dd882af4da2f6244a6a1" exitCode=0 Dec 06 06:18:58 crc kubenswrapper[4809]: I1206 06:18:58.658922 4809 generic.go:334] "Generic (PLEG): container finished" podID="395a8d1c-6b2e-4499-9aec-dfb78bf272d8" containerID="adaff5f7480aeb10e1464676ffcb994398fc022e3838b6b67c975f6f73eb3824" exitCode=0 Dec 06 06:18:58 crc kubenswrapper[4809]: I1206 06:18:58.659047 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"395a8d1c-6b2e-4499-9aec-dfb78bf272d8","Type":"ContainerDied","Data":"adaff5f7480aeb10e1464676ffcb994398fc022e3838b6b67c975f6f73eb3824"} Dec 06 06:18:58 crc kubenswrapper[4809]: I1206 06:18:58.659452 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 06 06:18:58 crc kubenswrapper[4809]: I1206 06:18:58.659778 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 06 06:18:58 crc kubenswrapper[4809]: I1206 06:18:58.814349 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:18:58 crc kubenswrapper[4809]: I1206 06:18:58.958827 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/395a8d1c-6b2e-4499-9aec-dfb78bf272d8-config-data\") pod \"395a8d1c-6b2e-4499-9aec-dfb78bf272d8\" (UID: \"395a8d1c-6b2e-4499-9aec-dfb78bf272d8\") " Dec 06 06:18:58 crc kubenswrapper[4809]: I1206 06:18:58.958990 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/395a8d1c-6b2e-4499-9aec-dfb78bf272d8-run-httpd\") pod \"395a8d1c-6b2e-4499-9aec-dfb78bf272d8\" (UID: \"395a8d1c-6b2e-4499-9aec-dfb78bf272d8\") " Dec 06 06:18:58 crc kubenswrapper[4809]: I1206 06:18:58.959092 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-chtzh\" (UniqueName: \"kubernetes.io/projected/395a8d1c-6b2e-4499-9aec-dfb78bf272d8-kube-api-access-chtzh\") pod \"395a8d1c-6b2e-4499-9aec-dfb78bf272d8\" (UID: \"395a8d1c-6b2e-4499-9aec-dfb78bf272d8\") " Dec 06 06:18:58 crc kubenswrapper[4809]: I1206 06:18:58.959155 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/395a8d1c-6b2e-4499-9aec-dfb78bf272d8-log-httpd\") pod \"395a8d1c-6b2e-4499-9aec-dfb78bf272d8\" (UID: \"395a8d1c-6b2e-4499-9aec-dfb78bf272d8\") " Dec 06 06:18:58 crc kubenswrapper[4809]: I1206 06:18:58.959184 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/395a8d1c-6b2e-4499-9aec-dfb78bf272d8-sg-core-conf-yaml\") pod \"395a8d1c-6b2e-4499-9aec-dfb78bf272d8\" (UID: \"395a8d1c-6b2e-4499-9aec-dfb78bf272d8\") " Dec 06 06:18:58 crc kubenswrapper[4809]: I1206 06:18:58.959216 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/395a8d1c-6b2e-4499-9aec-dfb78bf272d8-scripts\") pod \"395a8d1c-6b2e-4499-9aec-dfb78bf272d8\" (UID: \"395a8d1c-6b2e-4499-9aec-dfb78bf272d8\") " Dec 06 06:18:58 crc kubenswrapper[4809]: I1206 06:18:58.959292 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/395a8d1c-6b2e-4499-9aec-dfb78bf272d8-combined-ca-bundle\") pod \"395a8d1c-6b2e-4499-9aec-dfb78bf272d8\" (UID: \"395a8d1c-6b2e-4499-9aec-dfb78bf272d8\") " Dec 06 06:18:58 crc kubenswrapper[4809]: I1206 06:18:58.959856 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/395a8d1c-6b2e-4499-9aec-dfb78bf272d8-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "395a8d1c-6b2e-4499-9aec-dfb78bf272d8" (UID: "395a8d1c-6b2e-4499-9aec-dfb78bf272d8"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:18:58 crc kubenswrapper[4809]: I1206 06:18:58.960068 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/395a8d1c-6b2e-4499-9aec-dfb78bf272d8-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "395a8d1c-6b2e-4499-9aec-dfb78bf272d8" (UID: "395a8d1c-6b2e-4499-9aec-dfb78bf272d8"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:18:58 crc kubenswrapper[4809]: I1206 06:18:58.965597 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/395a8d1c-6b2e-4499-9aec-dfb78bf272d8-scripts" (OuterVolumeSpecName: "scripts") pod "395a8d1c-6b2e-4499-9aec-dfb78bf272d8" (UID: "395a8d1c-6b2e-4499-9aec-dfb78bf272d8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:18:58 crc kubenswrapper[4809]: I1206 06:18:58.966184 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/395a8d1c-6b2e-4499-9aec-dfb78bf272d8-kube-api-access-chtzh" (OuterVolumeSpecName: "kube-api-access-chtzh") pod "395a8d1c-6b2e-4499-9aec-dfb78bf272d8" (UID: "395a8d1c-6b2e-4499-9aec-dfb78bf272d8"). InnerVolumeSpecName "kube-api-access-chtzh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.007419 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/395a8d1c-6b2e-4499-9aec-dfb78bf272d8-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "395a8d1c-6b2e-4499-9aec-dfb78bf272d8" (UID: "395a8d1c-6b2e-4499-9aec-dfb78bf272d8"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.061887 4809 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/395a8d1c-6b2e-4499-9aec-dfb78bf272d8-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.061924 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-chtzh\" (UniqueName: \"kubernetes.io/projected/395a8d1c-6b2e-4499-9aec-dfb78bf272d8-kube-api-access-chtzh\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.061966 4809 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/395a8d1c-6b2e-4499-9aec-dfb78bf272d8-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.061976 4809 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/395a8d1c-6b2e-4499-9aec-dfb78bf272d8-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.061987 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/395a8d1c-6b2e-4499-9aec-dfb78bf272d8-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.095145 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/395a8d1c-6b2e-4499-9aec-dfb78bf272d8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "395a8d1c-6b2e-4499-9aec-dfb78bf272d8" (UID: "395a8d1c-6b2e-4499-9aec-dfb78bf272d8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.104840 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/395a8d1c-6b2e-4499-9aec-dfb78bf272d8-config-data" (OuterVolumeSpecName: "config-data") pod "395a8d1c-6b2e-4499-9aec-dfb78bf272d8" (UID: "395a8d1c-6b2e-4499-9aec-dfb78bf272d8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.164038 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/395a8d1c-6b2e-4499-9aec-dfb78bf272d8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.164081 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/395a8d1c-6b2e-4499-9aec-dfb78bf272d8-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.167416 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.167461 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.230211 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.267581 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.672630 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"395a8d1c-6b2e-4499-9aec-dfb78bf272d8","Type":"ContainerDied","Data":"c2bca4f5af1027b0f259ca193346b4d777593b4bae812815d36d0377087ff047"} Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.673793 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.673898 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.673913 4809 scope.go:117] "RemoveContainer" containerID="8fca81cf46ee4798615c86c02e502e11f56f7b52f462cc24b53865a56ca46131" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.672763 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.709737 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.709816 4809 scope.go:117] "RemoveContainer" containerID="73dd23b14e07077f2f91b598925c47338403483b252f273fab0f0c549ff24fd0" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.723481 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.750023 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:18:59 crc kubenswrapper[4809]: E1206 06:18:59.750843 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="395a8d1c-6b2e-4499-9aec-dfb78bf272d8" containerName="ceilometer-notification-agent" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.750958 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="395a8d1c-6b2e-4499-9aec-dfb78bf272d8" containerName="ceilometer-notification-agent" Dec 06 06:18:59 crc kubenswrapper[4809]: E1206 06:18:59.751051 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d97f700e-65e3-41e3-a053-a9e97bd9857f" containerName="registry-server" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.751135 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d97f700e-65e3-41e3-a053-a9e97bd9857f" containerName="registry-server" Dec 06 06:18:59 crc kubenswrapper[4809]: E1206 06:18:59.751225 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d97f700e-65e3-41e3-a053-a9e97bd9857f" containerName="extract-content" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.751310 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d97f700e-65e3-41e3-a053-a9e97bd9857f" containerName="extract-content" Dec 06 06:18:59 crc kubenswrapper[4809]: E1206 06:18:59.751403 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a" containerName="heat-engine" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.751487 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a" containerName="heat-engine" Dec 06 06:18:59 crc kubenswrapper[4809]: E1206 06:18:59.751576 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="395a8d1c-6b2e-4499-9aec-dfb78bf272d8" containerName="ceilometer-central-agent" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.751647 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="395a8d1c-6b2e-4499-9aec-dfb78bf272d8" containerName="ceilometer-central-agent" Dec 06 06:18:59 crc kubenswrapper[4809]: E1206 06:18:59.751741 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="395a8d1c-6b2e-4499-9aec-dfb78bf272d8" containerName="proxy-httpd" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.751815 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="395a8d1c-6b2e-4499-9aec-dfb78bf272d8" containerName="proxy-httpd" Dec 06 06:18:59 crc kubenswrapper[4809]: E1206 06:18:59.751920 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="395a8d1c-6b2e-4499-9aec-dfb78bf272d8" containerName="sg-core" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.752028 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="395a8d1c-6b2e-4499-9aec-dfb78bf272d8" containerName="sg-core" Dec 06 06:18:59 crc kubenswrapper[4809]: E1206 06:18:59.752124 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d97f700e-65e3-41e3-a053-a9e97bd9857f" containerName="extract-utilities" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.752209 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d97f700e-65e3-41e3-a053-a9e97bd9857f" containerName="extract-utilities" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.752638 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="395a8d1c-6b2e-4499-9aec-dfb78bf272d8" containerName="proxy-httpd" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.752733 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="395a8d1c-6b2e-4499-9aec-dfb78bf272d8" containerName="sg-core" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.752841 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="395a8d1c-6b2e-4499-9aec-dfb78bf272d8" containerName="ceilometer-central-agent" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.752960 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="d97f700e-65e3-41e3-a053-a9e97bd9857f" containerName="registry-server" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.753067 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a" containerName="heat-engine" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.753147 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="395a8d1c-6b2e-4499-9aec-dfb78bf272d8" containerName="ceilometer-notification-agent" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.756039 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.760770 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.761908 4809 scope.go:117] "RemoveContainer" containerID="0e23cc297bbd882a9475ed237aa11168d54243e3db05dd882af4da2f6244a6a1" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.762331 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.778551 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.802124 4809 scope.go:117] "RemoveContainer" containerID="adaff5f7480aeb10e1464676ffcb994398fc022e3838b6b67c975f6f73eb3824" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.857858 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:18:59 crc kubenswrapper[4809]: E1206 06:18:59.858753 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle config-data kube-api-access-8lv8c log-httpd run-httpd scripts sg-core-conf-yaml], unattached volumes=[], failed to process volumes=[combined-ca-bundle config-data kube-api-access-8lv8c log-httpd run-httpd scripts sg-core-conf-yaml]: context canceled" pod="openstack/ceilometer-0" podUID="9904b8e9-9383-4afc-bd5d-21f33a7993c4" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.880313 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lv8c\" (UniqueName: \"kubernetes.io/projected/9904b8e9-9383-4afc-bd5d-21f33a7993c4-kube-api-access-8lv8c\") pod \"ceilometer-0\" (UID: \"9904b8e9-9383-4afc-bd5d-21f33a7993c4\") " pod="openstack/ceilometer-0" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.880581 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9904b8e9-9383-4afc-bd5d-21f33a7993c4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9904b8e9-9383-4afc-bd5d-21f33a7993c4\") " pod="openstack/ceilometer-0" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.880679 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9904b8e9-9383-4afc-bd5d-21f33a7993c4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9904b8e9-9383-4afc-bd5d-21f33a7993c4\") " pod="openstack/ceilometer-0" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.880845 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9904b8e9-9383-4afc-bd5d-21f33a7993c4-log-httpd\") pod \"ceilometer-0\" (UID: \"9904b8e9-9383-4afc-bd5d-21f33a7993c4\") " pod="openstack/ceilometer-0" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.881075 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9904b8e9-9383-4afc-bd5d-21f33a7993c4-scripts\") pod \"ceilometer-0\" (UID: \"9904b8e9-9383-4afc-bd5d-21f33a7993c4\") " pod="openstack/ceilometer-0" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.881243 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9904b8e9-9383-4afc-bd5d-21f33a7993c4-config-data\") pod \"ceilometer-0\" (UID: \"9904b8e9-9383-4afc-bd5d-21f33a7993c4\") " pod="openstack/ceilometer-0" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.881364 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9904b8e9-9383-4afc-bd5d-21f33a7993c4-run-httpd\") pod \"ceilometer-0\" (UID: \"9904b8e9-9383-4afc-bd5d-21f33a7993c4\") " pod="openstack/ceilometer-0" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.982887 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9904b8e9-9383-4afc-bd5d-21f33a7993c4-log-httpd\") pod \"ceilometer-0\" (UID: \"9904b8e9-9383-4afc-bd5d-21f33a7993c4\") " pod="openstack/ceilometer-0" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.983017 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9904b8e9-9383-4afc-bd5d-21f33a7993c4-scripts\") pod \"ceilometer-0\" (UID: \"9904b8e9-9383-4afc-bd5d-21f33a7993c4\") " pod="openstack/ceilometer-0" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.983083 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9904b8e9-9383-4afc-bd5d-21f33a7993c4-config-data\") pod \"ceilometer-0\" (UID: \"9904b8e9-9383-4afc-bd5d-21f33a7993c4\") " pod="openstack/ceilometer-0" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.983135 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9904b8e9-9383-4afc-bd5d-21f33a7993c4-run-httpd\") pod \"ceilometer-0\" (UID: \"9904b8e9-9383-4afc-bd5d-21f33a7993c4\") " pod="openstack/ceilometer-0" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.983219 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lv8c\" (UniqueName: \"kubernetes.io/projected/9904b8e9-9383-4afc-bd5d-21f33a7993c4-kube-api-access-8lv8c\") pod \"ceilometer-0\" (UID: \"9904b8e9-9383-4afc-bd5d-21f33a7993c4\") " pod="openstack/ceilometer-0" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.983308 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9904b8e9-9383-4afc-bd5d-21f33a7993c4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9904b8e9-9383-4afc-bd5d-21f33a7993c4\") " pod="openstack/ceilometer-0" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.983345 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9904b8e9-9383-4afc-bd5d-21f33a7993c4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9904b8e9-9383-4afc-bd5d-21f33a7993c4\") " pod="openstack/ceilometer-0" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.983434 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9904b8e9-9383-4afc-bd5d-21f33a7993c4-log-httpd\") pod \"ceilometer-0\" (UID: \"9904b8e9-9383-4afc-bd5d-21f33a7993c4\") " pod="openstack/ceilometer-0" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.983691 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9904b8e9-9383-4afc-bd5d-21f33a7993c4-run-httpd\") pod \"ceilometer-0\" (UID: \"9904b8e9-9383-4afc-bd5d-21f33a7993c4\") " pod="openstack/ceilometer-0" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.987416 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9904b8e9-9383-4afc-bd5d-21f33a7993c4-scripts\") pod \"ceilometer-0\" (UID: \"9904b8e9-9383-4afc-bd5d-21f33a7993c4\") " pod="openstack/ceilometer-0" Dec 06 06:18:59 crc kubenswrapper[4809]: I1206 06:18:59.987865 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9904b8e9-9383-4afc-bd5d-21f33a7993c4-config-data\") pod \"ceilometer-0\" (UID: \"9904b8e9-9383-4afc-bd5d-21f33a7993c4\") " pod="openstack/ceilometer-0" Dec 06 06:19:00 crc kubenswrapper[4809]: I1206 06:19:00.004765 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9904b8e9-9383-4afc-bd5d-21f33a7993c4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9904b8e9-9383-4afc-bd5d-21f33a7993c4\") " pod="openstack/ceilometer-0" Dec 06 06:19:00 crc kubenswrapper[4809]: I1206 06:19:00.005115 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9904b8e9-9383-4afc-bd5d-21f33a7993c4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9904b8e9-9383-4afc-bd5d-21f33a7993c4\") " pod="openstack/ceilometer-0" Dec 06 06:19:00 crc kubenswrapper[4809]: I1206 06:19:00.015599 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lv8c\" (UniqueName: \"kubernetes.io/projected/9904b8e9-9383-4afc-bd5d-21f33a7993c4-kube-api-access-8lv8c\") pod \"ceilometer-0\" (UID: \"9904b8e9-9383-4afc-bd5d-21f33a7993c4\") " pod="openstack/ceilometer-0" Dec 06 06:19:00 crc kubenswrapper[4809]: I1206 06:19:00.688347 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:19:00 crc kubenswrapper[4809]: I1206 06:19:00.710130 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:19:00 crc kubenswrapper[4809]: I1206 06:19:00.800923 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9904b8e9-9383-4afc-bd5d-21f33a7993c4-config-data\") pod \"9904b8e9-9383-4afc-bd5d-21f33a7993c4\" (UID: \"9904b8e9-9383-4afc-bd5d-21f33a7993c4\") " Dec 06 06:19:00 crc kubenswrapper[4809]: I1206 06:19:00.800995 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9904b8e9-9383-4afc-bd5d-21f33a7993c4-scripts\") pod \"9904b8e9-9383-4afc-bd5d-21f33a7993c4\" (UID: \"9904b8e9-9383-4afc-bd5d-21f33a7993c4\") " Dec 06 06:19:00 crc kubenswrapper[4809]: I1206 06:19:00.801078 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9904b8e9-9383-4afc-bd5d-21f33a7993c4-log-httpd\") pod \"9904b8e9-9383-4afc-bd5d-21f33a7993c4\" (UID: \"9904b8e9-9383-4afc-bd5d-21f33a7993c4\") " Dec 06 06:19:00 crc kubenswrapper[4809]: I1206 06:19:00.801213 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9904b8e9-9383-4afc-bd5d-21f33a7993c4-sg-core-conf-yaml\") pod \"9904b8e9-9383-4afc-bd5d-21f33a7993c4\" (UID: \"9904b8e9-9383-4afc-bd5d-21f33a7993c4\") " Dec 06 06:19:00 crc kubenswrapper[4809]: I1206 06:19:00.801267 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8lv8c\" (UniqueName: \"kubernetes.io/projected/9904b8e9-9383-4afc-bd5d-21f33a7993c4-kube-api-access-8lv8c\") pod \"9904b8e9-9383-4afc-bd5d-21f33a7993c4\" (UID: \"9904b8e9-9383-4afc-bd5d-21f33a7993c4\") " Dec 06 06:19:00 crc kubenswrapper[4809]: I1206 06:19:00.801363 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9904b8e9-9383-4afc-bd5d-21f33a7993c4-combined-ca-bundle\") pod \"9904b8e9-9383-4afc-bd5d-21f33a7993c4\" (UID: \"9904b8e9-9383-4afc-bd5d-21f33a7993c4\") " Dec 06 06:19:00 crc kubenswrapper[4809]: I1206 06:19:00.801398 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9904b8e9-9383-4afc-bd5d-21f33a7993c4-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "9904b8e9-9383-4afc-bd5d-21f33a7993c4" (UID: "9904b8e9-9383-4afc-bd5d-21f33a7993c4"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:19:00 crc kubenswrapper[4809]: I1206 06:19:00.801404 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9904b8e9-9383-4afc-bd5d-21f33a7993c4-run-httpd\") pod \"9904b8e9-9383-4afc-bd5d-21f33a7993c4\" (UID: \"9904b8e9-9383-4afc-bd5d-21f33a7993c4\") " Dec 06 06:19:00 crc kubenswrapper[4809]: I1206 06:19:00.801589 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9904b8e9-9383-4afc-bd5d-21f33a7993c4-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "9904b8e9-9383-4afc-bd5d-21f33a7993c4" (UID: "9904b8e9-9383-4afc-bd5d-21f33a7993c4"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:19:00 crc kubenswrapper[4809]: I1206 06:19:00.802206 4809 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9904b8e9-9383-4afc-bd5d-21f33a7993c4-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:00 crc kubenswrapper[4809]: I1206 06:19:00.802226 4809 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9904b8e9-9383-4afc-bd5d-21f33a7993c4-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:00 crc kubenswrapper[4809]: I1206 06:19:00.806599 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9904b8e9-9383-4afc-bd5d-21f33a7993c4-scripts" (OuterVolumeSpecName: "scripts") pod "9904b8e9-9383-4afc-bd5d-21f33a7993c4" (UID: "9904b8e9-9383-4afc-bd5d-21f33a7993c4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:19:00 crc kubenswrapper[4809]: I1206 06:19:00.807113 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9904b8e9-9383-4afc-bd5d-21f33a7993c4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9904b8e9-9383-4afc-bd5d-21f33a7993c4" (UID: "9904b8e9-9383-4afc-bd5d-21f33a7993c4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:19:00 crc kubenswrapper[4809]: I1206 06:19:00.807322 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9904b8e9-9383-4afc-bd5d-21f33a7993c4-config-data" (OuterVolumeSpecName: "config-data") pod "9904b8e9-9383-4afc-bd5d-21f33a7993c4" (UID: "9904b8e9-9383-4afc-bd5d-21f33a7993c4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:19:00 crc kubenswrapper[4809]: I1206 06:19:00.808086 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9904b8e9-9383-4afc-bd5d-21f33a7993c4-kube-api-access-8lv8c" (OuterVolumeSpecName: "kube-api-access-8lv8c") pod "9904b8e9-9383-4afc-bd5d-21f33a7993c4" (UID: "9904b8e9-9383-4afc-bd5d-21f33a7993c4"). InnerVolumeSpecName "kube-api-access-8lv8c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:19:00 crc kubenswrapper[4809]: I1206 06:19:00.808389 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9904b8e9-9383-4afc-bd5d-21f33a7993c4-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "9904b8e9-9383-4afc-bd5d-21f33a7993c4" (UID: "9904b8e9-9383-4afc-bd5d-21f33a7993c4"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:19:00 crc kubenswrapper[4809]: I1206 06:19:00.904422 4809 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9904b8e9-9383-4afc-bd5d-21f33a7993c4-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:00 crc kubenswrapper[4809]: I1206 06:19:00.904459 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8lv8c\" (UniqueName: \"kubernetes.io/projected/9904b8e9-9383-4afc-bd5d-21f33a7993c4-kube-api-access-8lv8c\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:00 crc kubenswrapper[4809]: I1206 06:19:00.904471 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9904b8e9-9383-4afc-bd5d-21f33a7993c4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:00 crc kubenswrapper[4809]: I1206 06:19:00.904479 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9904b8e9-9383-4afc-bd5d-21f33a7993c4-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:00 crc kubenswrapper[4809]: I1206 06:19:00.904488 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9904b8e9-9383-4afc-bd5d-21f33a7993c4-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:01 crc kubenswrapper[4809]: I1206 06:19:01.094647 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 06 06:19:01 crc kubenswrapper[4809]: I1206 06:19:01.094757 4809 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 06:19:01 crc kubenswrapper[4809]: I1206 06:19:01.096376 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 06 06:19:01 crc kubenswrapper[4809]: I1206 06:19:01.402483 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="395a8d1c-6b2e-4499-9aec-dfb78bf272d8" path="/var/lib/kubelet/pods/395a8d1c-6b2e-4499-9aec-dfb78bf272d8/volumes" Dec 06 06:19:01 crc kubenswrapper[4809]: I1206 06:19:01.698611 4809 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 06:19:01 crc kubenswrapper[4809]: I1206 06:19:01.699228 4809 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 06:19:01 crc kubenswrapper[4809]: I1206 06:19:01.698637 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:19:01 crc kubenswrapper[4809]: I1206 06:19:01.740204 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 06 06:19:01 crc kubenswrapper[4809]: I1206 06:19:01.783388 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:19:01 crc kubenswrapper[4809]: I1206 06:19:01.800782 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:19:01 crc kubenswrapper[4809]: I1206 06:19:01.807440 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 06 06:19:01 crc kubenswrapper[4809]: I1206 06:19:01.821220 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:19:01 crc kubenswrapper[4809]: I1206 06:19:01.837031 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:19:01 crc kubenswrapper[4809]: I1206 06:19:01.845873 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 06:19:01 crc kubenswrapper[4809]: I1206 06:19:01.851713 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 06:19:01 crc kubenswrapper[4809]: I1206 06:19:01.878819 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:19:01 crc kubenswrapper[4809]: I1206 06:19:01.941919 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/438fc849-1432-45d3-8709-4b9e4734e56f-config-data\") pod \"ceilometer-0\" (UID: \"438fc849-1432-45d3-8709-4b9e4734e56f\") " pod="openstack/ceilometer-0" Dec 06 06:19:01 crc kubenswrapper[4809]: I1206 06:19:01.942228 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/438fc849-1432-45d3-8709-4b9e4734e56f-scripts\") pod \"ceilometer-0\" (UID: \"438fc849-1432-45d3-8709-4b9e4734e56f\") " pod="openstack/ceilometer-0" Dec 06 06:19:01 crc kubenswrapper[4809]: I1206 06:19:01.942388 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/438fc849-1432-45d3-8709-4b9e4734e56f-log-httpd\") pod \"ceilometer-0\" (UID: \"438fc849-1432-45d3-8709-4b9e4734e56f\") " pod="openstack/ceilometer-0" Dec 06 06:19:01 crc kubenswrapper[4809]: I1206 06:19:01.942457 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/438fc849-1432-45d3-8709-4b9e4734e56f-run-httpd\") pod \"ceilometer-0\" (UID: \"438fc849-1432-45d3-8709-4b9e4734e56f\") " pod="openstack/ceilometer-0" Dec 06 06:19:01 crc kubenswrapper[4809]: I1206 06:19:01.942585 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/438fc849-1432-45d3-8709-4b9e4734e56f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"438fc849-1432-45d3-8709-4b9e4734e56f\") " pod="openstack/ceilometer-0" Dec 06 06:19:01 crc kubenswrapper[4809]: I1206 06:19:01.942800 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bmv8\" (UniqueName: \"kubernetes.io/projected/438fc849-1432-45d3-8709-4b9e4734e56f-kube-api-access-8bmv8\") pod \"ceilometer-0\" (UID: \"438fc849-1432-45d3-8709-4b9e4734e56f\") " pod="openstack/ceilometer-0" Dec 06 06:19:01 crc kubenswrapper[4809]: I1206 06:19:01.943157 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/438fc849-1432-45d3-8709-4b9e4734e56f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"438fc849-1432-45d3-8709-4b9e4734e56f\") " pod="openstack/ceilometer-0" Dec 06 06:19:02 crc kubenswrapper[4809]: I1206 06:19:02.045621 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/438fc849-1432-45d3-8709-4b9e4734e56f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"438fc849-1432-45d3-8709-4b9e4734e56f\") " pod="openstack/ceilometer-0" Dec 06 06:19:02 crc kubenswrapper[4809]: I1206 06:19:02.045722 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bmv8\" (UniqueName: \"kubernetes.io/projected/438fc849-1432-45d3-8709-4b9e4734e56f-kube-api-access-8bmv8\") pod \"ceilometer-0\" (UID: \"438fc849-1432-45d3-8709-4b9e4734e56f\") " pod="openstack/ceilometer-0" Dec 06 06:19:02 crc kubenswrapper[4809]: I1206 06:19:02.045808 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/438fc849-1432-45d3-8709-4b9e4734e56f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"438fc849-1432-45d3-8709-4b9e4734e56f\") " pod="openstack/ceilometer-0" Dec 06 06:19:02 crc kubenswrapper[4809]: I1206 06:19:02.045862 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/438fc849-1432-45d3-8709-4b9e4734e56f-config-data\") pod \"ceilometer-0\" (UID: \"438fc849-1432-45d3-8709-4b9e4734e56f\") " pod="openstack/ceilometer-0" Dec 06 06:19:02 crc kubenswrapper[4809]: I1206 06:19:02.046047 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/438fc849-1432-45d3-8709-4b9e4734e56f-scripts\") pod \"ceilometer-0\" (UID: \"438fc849-1432-45d3-8709-4b9e4734e56f\") " pod="openstack/ceilometer-0" Dec 06 06:19:02 crc kubenswrapper[4809]: I1206 06:19:02.046082 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/438fc849-1432-45d3-8709-4b9e4734e56f-log-httpd\") pod \"ceilometer-0\" (UID: \"438fc849-1432-45d3-8709-4b9e4734e56f\") " pod="openstack/ceilometer-0" Dec 06 06:19:02 crc kubenswrapper[4809]: I1206 06:19:02.046110 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/438fc849-1432-45d3-8709-4b9e4734e56f-run-httpd\") pod \"ceilometer-0\" (UID: \"438fc849-1432-45d3-8709-4b9e4734e56f\") " pod="openstack/ceilometer-0" Dec 06 06:19:02 crc kubenswrapper[4809]: I1206 06:19:02.046575 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/438fc849-1432-45d3-8709-4b9e4734e56f-run-httpd\") pod \"ceilometer-0\" (UID: \"438fc849-1432-45d3-8709-4b9e4734e56f\") " pod="openstack/ceilometer-0" Dec 06 06:19:02 crc kubenswrapper[4809]: I1206 06:19:02.046840 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/438fc849-1432-45d3-8709-4b9e4734e56f-log-httpd\") pod \"ceilometer-0\" (UID: \"438fc849-1432-45d3-8709-4b9e4734e56f\") " pod="openstack/ceilometer-0" Dec 06 06:19:02 crc kubenswrapper[4809]: I1206 06:19:02.063804 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/438fc849-1432-45d3-8709-4b9e4734e56f-scripts\") pod \"ceilometer-0\" (UID: \"438fc849-1432-45d3-8709-4b9e4734e56f\") " pod="openstack/ceilometer-0" Dec 06 06:19:02 crc kubenswrapper[4809]: I1206 06:19:02.068653 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/438fc849-1432-45d3-8709-4b9e4734e56f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"438fc849-1432-45d3-8709-4b9e4734e56f\") " pod="openstack/ceilometer-0" Dec 06 06:19:02 crc kubenswrapper[4809]: I1206 06:19:02.069201 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/438fc849-1432-45d3-8709-4b9e4734e56f-config-data\") pod \"ceilometer-0\" (UID: \"438fc849-1432-45d3-8709-4b9e4734e56f\") " pod="openstack/ceilometer-0" Dec 06 06:19:02 crc kubenswrapper[4809]: I1206 06:19:02.069789 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/438fc849-1432-45d3-8709-4b9e4734e56f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"438fc849-1432-45d3-8709-4b9e4734e56f\") " pod="openstack/ceilometer-0" Dec 06 06:19:02 crc kubenswrapper[4809]: I1206 06:19:02.071515 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bmv8\" (UniqueName: \"kubernetes.io/projected/438fc849-1432-45d3-8709-4b9e4734e56f-kube-api-access-8bmv8\") pod \"ceilometer-0\" (UID: \"438fc849-1432-45d3-8709-4b9e4734e56f\") " pod="openstack/ceilometer-0" Dec 06 06:19:02 crc kubenswrapper[4809]: I1206 06:19:02.171565 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:19:02 crc kubenswrapper[4809]: I1206 06:19:02.780089 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:19:02 crc kubenswrapper[4809]: I1206 06:19:02.783404 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zx9rg" Dec 06 06:19:02 crc kubenswrapper[4809]: I1206 06:19:02.856208 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zx9rg"] Dec 06 06:19:03 crc kubenswrapper[4809]: I1206 06:19:03.337092 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:19:03 crc kubenswrapper[4809]: I1206 06:19:03.405590 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9904b8e9-9383-4afc-bd5d-21f33a7993c4" path="/var/lib/kubelet/pods/9904b8e9-9383-4afc-bd5d-21f33a7993c4/volumes" Dec 06 06:19:03 crc kubenswrapper[4809]: I1206 06:19:03.752220 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"438fc849-1432-45d3-8709-4b9e4734e56f","Type":"ContainerStarted","Data":"b3951cb22277a2e28993378fa15999ef2ef04c53df89c85b8d0947ff920deaf7"} Dec 06 06:19:03 crc kubenswrapper[4809]: I1206 06:19:03.752459 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zx9rg" podUID="d4adfb97-c4f0-4093-80c5-e0229644834c" containerName="registry-server" containerID="cri-o://518a0ed359ea197b4316903bbfd2d5ac37c11c7bc98ce51fc9b8ba3c39154423" gracePeriod=2 Dec 06 06:19:04 crc kubenswrapper[4809]: I1206 06:19:04.768034 4809 generic.go:334] "Generic (PLEG): container finished" podID="d4adfb97-c4f0-4093-80c5-e0229644834c" containerID="518a0ed359ea197b4316903bbfd2d5ac37c11c7bc98ce51fc9b8ba3c39154423" exitCode=0 Dec 06 06:19:04 crc kubenswrapper[4809]: I1206 06:19:04.768148 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zx9rg" event={"ID":"d4adfb97-c4f0-4093-80c5-e0229644834c","Type":"ContainerDied","Data":"518a0ed359ea197b4316903bbfd2d5ac37c11c7bc98ce51fc9b8ba3c39154423"} Dec 06 06:19:04 crc kubenswrapper[4809]: I1206 06:19:04.770553 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"438fc849-1432-45d3-8709-4b9e4734e56f","Type":"ContainerStarted","Data":"2eaf3473d57f772f14f96c219113e3f7b1dc1bf47459cd506632caa4b1b48dc1"} Dec 06 06:19:05 crc kubenswrapper[4809]: I1206 06:19:05.083632 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zx9rg" Dec 06 06:19:05 crc kubenswrapper[4809]: I1206 06:19:05.253973 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7s27g\" (UniqueName: \"kubernetes.io/projected/d4adfb97-c4f0-4093-80c5-e0229644834c-kube-api-access-7s27g\") pod \"d4adfb97-c4f0-4093-80c5-e0229644834c\" (UID: \"d4adfb97-c4f0-4093-80c5-e0229644834c\") " Dec 06 06:19:05 crc kubenswrapper[4809]: I1206 06:19:05.254637 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4adfb97-c4f0-4093-80c5-e0229644834c-catalog-content\") pod \"d4adfb97-c4f0-4093-80c5-e0229644834c\" (UID: \"d4adfb97-c4f0-4093-80c5-e0229644834c\") " Dec 06 06:19:05 crc kubenswrapper[4809]: I1206 06:19:05.257304 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4adfb97-c4f0-4093-80c5-e0229644834c-utilities\") pod \"d4adfb97-c4f0-4093-80c5-e0229644834c\" (UID: \"d4adfb97-c4f0-4093-80c5-e0229644834c\") " Dec 06 06:19:05 crc kubenswrapper[4809]: I1206 06:19:05.263059 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4adfb97-c4f0-4093-80c5-e0229644834c-utilities" (OuterVolumeSpecName: "utilities") pod "d4adfb97-c4f0-4093-80c5-e0229644834c" (UID: "d4adfb97-c4f0-4093-80c5-e0229644834c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:19:05 crc kubenswrapper[4809]: I1206 06:19:05.265350 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4adfb97-c4f0-4093-80c5-e0229644834c-kube-api-access-7s27g" (OuterVolumeSpecName: "kube-api-access-7s27g") pod "d4adfb97-c4f0-4093-80c5-e0229644834c" (UID: "d4adfb97-c4f0-4093-80c5-e0229644834c"). InnerVolumeSpecName "kube-api-access-7s27g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:19:05 crc kubenswrapper[4809]: I1206 06:19:05.279113 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4adfb97-c4f0-4093-80c5-e0229644834c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d4adfb97-c4f0-4093-80c5-e0229644834c" (UID: "d4adfb97-c4f0-4093-80c5-e0229644834c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:19:05 crc kubenswrapper[4809]: I1206 06:19:05.360781 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4adfb97-c4f0-4093-80c5-e0229644834c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:05 crc kubenswrapper[4809]: I1206 06:19:05.360816 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4adfb97-c4f0-4093-80c5-e0229644834c-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:05 crc kubenswrapper[4809]: I1206 06:19:05.360830 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7s27g\" (UniqueName: \"kubernetes.io/projected/d4adfb97-c4f0-4093-80c5-e0229644834c-kube-api-access-7s27g\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:05 crc kubenswrapper[4809]: I1206 06:19:05.783543 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"438fc849-1432-45d3-8709-4b9e4734e56f","Type":"ContainerStarted","Data":"214b66e1c2371010ad89f60894fff66ea2452e3abb99cdccc6dca247c7d5028e"} Dec 06 06:19:05 crc kubenswrapper[4809]: I1206 06:19:05.786354 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zx9rg" event={"ID":"d4adfb97-c4f0-4093-80c5-e0229644834c","Type":"ContainerDied","Data":"f101c865677027413942eb36929ac28a4f61696de2f786ee882cb7c0ce2f4bc5"} Dec 06 06:19:05 crc kubenswrapper[4809]: I1206 06:19:05.786394 4809 scope.go:117] "RemoveContainer" containerID="518a0ed359ea197b4316903bbfd2d5ac37c11c7bc98ce51fc9b8ba3c39154423" Dec 06 06:19:05 crc kubenswrapper[4809]: I1206 06:19:05.786518 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zx9rg" Dec 06 06:19:05 crc kubenswrapper[4809]: I1206 06:19:05.812948 4809 scope.go:117] "RemoveContainer" containerID="cbcfe2679090af74eefb076ffaa4516d695958b4b6944d99b689cd324c47184d" Dec 06 06:19:05 crc kubenswrapper[4809]: I1206 06:19:05.823807 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zx9rg"] Dec 06 06:19:05 crc kubenswrapper[4809]: I1206 06:19:05.845656 4809 scope.go:117] "RemoveContainer" containerID="7c94743d940eabfd47a0a92465458b1b3d9348c604c3a433c1da4fe15ed1eb61" Dec 06 06:19:05 crc kubenswrapper[4809]: I1206 06:19:05.845735 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zx9rg"] Dec 06 06:19:06 crc kubenswrapper[4809]: I1206 06:19:06.390361 4809 scope.go:117] "RemoveContainer" containerID="9bc2dc88b3b9f309ce0169aa510326dd77022dece4c69cab3141df00fe5c0106" Dec 06 06:19:06 crc kubenswrapper[4809]: E1206 06:19:06.394785 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:19:06 crc kubenswrapper[4809]: I1206 06:19:06.797123 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"438fc849-1432-45d3-8709-4b9e4734e56f","Type":"ContainerStarted","Data":"7acc7ef16f31e7c4b28b16816753e06fe96f5a8c7956bb11f29cce88ea77c9e9"} Dec 06 06:19:07 crc kubenswrapper[4809]: I1206 06:19:07.418734 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4adfb97-c4f0-4093-80c5-e0229644834c" path="/var/lib/kubelet/pods/d4adfb97-c4f0-4093-80c5-e0229644834c/volumes" Dec 06 06:19:08 crc kubenswrapper[4809]: I1206 06:19:08.034633 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="53615cf8-5ea5-4d28-a4e7-1491e2b33744" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.211:3000/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 06:19:08 crc kubenswrapper[4809]: I1206 06:19:08.826581 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"438fc849-1432-45d3-8709-4b9e4734e56f","Type":"ContainerStarted","Data":"81af9828c93bb80ee07a5d7ec2bdb431c342fcf222dafec261f249d39e43a7e6"} Dec 06 06:19:08 crc kubenswrapper[4809]: I1206 06:19:08.826923 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="438fc849-1432-45d3-8709-4b9e4734e56f" containerName="ceilometer-notification-agent" containerID="cri-o://214b66e1c2371010ad89f60894fff66ea2452e3abb99cdccc6dca247c7d5028e" gracePeriod=30 Dec 06 06:19:08 crc kubenswrapper[4809]: I1206 06:19:08.826989 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 06:19:08 crc kubenswrapper[4809]: I1206 06:19:08.826894 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="438fc849-1432-45d3-8709-4b9e4734e56f" containerName="proxy-httpd" containerID="cri-o://81af9828c93bb80ee07a5d7ec2bdb431c342fcf222dafec261f249d39e43a7e6" gracePeriod=30 Dec 06 06:19:08 crc kubenswrapper[4809]: I1206 06:19:08.826906 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="438fc849-1432-45d3-8709-4b9e4734e56f" containerName="ceilometer-central-agent" containerID="cri-o://2eaf3473d57f772f14f96c219113e3f7b1dc1bf47459cd506632caa4b1b48dc1" gracePeriod=30 Dec 06 06:19:08 crc kubenswrapper[4809]: I1206 06:19:08.826923 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="438fc849-1432-45d3-8709-4b9e4734e56f" containerName="sg-core" containerID="cri-o://7acc7ef16f31e7c4b28b16816753e06fe96f5a8c7956bb11f29cce88ea77c9e9" gracePeriod=30 Dec 06 06:19:08 crc kubenswrapper[4809]: I1206 06:19:08.872650 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.514230574 podStartE2EDuration="7.87263342s" podCreationTimestamp="2025-12-06 06:19:01 +0000 UTC" firstStartedPulling="2025-12-06 06:19:02.783346765 +0000 UTC m=+1667.672329707" lastFinishedPulling="2025-12-06 06:19:08.141749611 +0000 UTC m=+1673.030732553" observedRunningTime="2025-12-06 06:19:08.865851168 +0000 UTC m=+1673.754834110" watchObservedRunningTime="2025-12-06 06:19:08.87263342 +0000 UTC m=+1673.761616362" Dec 06 06:19:09 crc kubenswrapper[4809]: E1206 06:19:09.568014 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod438fc849_1432_45d3_8709_4b9e4734e56f.slice/crio-conmon-214b66e1c2371010ad89f60894fff66ea2452e3abb99cdccc6dca247c7d5028e.scope\": RecentStats: unable to find data in memory cache]" Dec 06 06:19:09 crc kubenswrapper[4809]: I1206 06:19:09.838136 4809 generic.go:334] "Generic (PLEG): container finished" podID="13e0242d-c42e-4aad-95a2-62de95d87e42" containerID="2a58fd22495f3af98ef23bc24a54ae7cfc770faff52f1f4baae10ae5065326b0" exitCode=0 Dec 06 06:19:09 crc kubenswrapper[4809]: I1206 06:19:09.838223 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-vt99m" event={"ID":"13e0242d-c42e-4aad-95a2-62de95d87e42","Type":"ContainerDied","Data":"2a58fd22495f3af98ef23bc24a54ae7cfc770faff52f1f4baae10ae5065326b0"} Dec 06 06:19:09 crc kubenswrapper[4809]: I1206 06:19:09.842505 4809 generic.go:334] "Generic (PLEG): container finished" podID="438fc849-1432-45d3-8709-4b9e4734e56f" containerID="81af9828c93bb80ee07a5d7ec2bdb431c342fcf222dafec261f249d39e43a7e6" exitCode=0 Dec 06 06:19:09 crc kubenswrapper[4809]: I1206 06:19:09.842545 4809 generic.go:334] "Generic (PLEG): container finished" podID="438fc849-1432-45d3-8709-4b9e4734e56f" containerID="7acc7ef16f31e7c4b28b16816753e06fe96f5a8c7956bb11f29cce88ea77c9e9" exitCode=2 Dec 06 06:19:09 crc kubenswrapper[4809]: I1206 06:19:09.842554 4809 generic.go:334] "Generic (PLEG): container finished" podID="438fc849-1432-45d3-8709-4b9e4734e56f" containerID="214b66e1c2371010ad89f60894fff66ea2452e3abb99cdccc6dca247c7d5028e" exitCode=0 Dec 06 06:19:09 crc kubenswrapper[4809]: I1206 06:19:09.842580 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"438fc849-1432-45d3-8709-4b9e4734e56f","Type":"ContainerDied","Data":"81af9828c93bb80ee07a5d7ec2bdb431c342fcf222dafec261f249d39e43a7e6"} Dec 06 06:19:09 crc kubenswrapper[4809]: I1206 06:19:09.842619 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"438fc849-1432-45d3-8709-4b9e4734e56f","Type":"ContainerDied","Data":"7acc7ef16f31e7c4b28b16816753e06fe96f5a8c7956bb11f29cce88ea77c9e9"} Dec 06 06:19:09 crc kubenswrapper[4809]: I1206 06:19:09.842645 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"438fc849-1432-45d3-8709-4b9e4734e56f","Type":"ContainerDied","Data":"214b66e1c2371010ad89f60894fff66ea2452e3abb99cdccc6dca247c7d5028e"} Dec 06 06:19:11 crc kubenswrapper[4809]: I1206 06:19:11.366743 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-vt99m" Dec 06 06:19:11 crc kubenswrapper[4809]: I1206 06:19:11.517654 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13e0242d-c42e-4aad-95a2-62de95d87e42-combined-ca-bundle\") pod \"13e0242d-c42e-4aad-95a2-62de95d87e42\" (UID: \"13e0242d-c42e-4aad-95a2-62de95d87e42\") " Dec 06 06:19:11 crc kubenswrapper[4809]: I1206 06:19:11.518319 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13e0242d-c42e-4aad-95a2-62de95d87e42-scripts\") pod \"13e0242d-c42e-4aad-95a2-62de95d87e42\" (UID: \"13e0242d-c42e-4aad-95a2-62de95d87e42\") " Dec 06 06:19:11 crc kubenswrapper[4809]: I1206 06:19:11.518474 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4bxt9\" (UniqueName: \"kubernetes.io/projected/13e0242d-c42e-4aad-95a2-62de95d87e42-kube-api-access-4bxt9\") pod \"13e0242d-c42e-4aad-95a2-62de95d87e42\" (UID: \"13e0242d-c42e-4aad-95a2-62de95d87e42\") " Dec 06 06:19:11 crc kubenswrapper[4809]: I1206 06:19:11.518576 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13e0242d-c42e-4aad-95a2-62de95d87e42-config-data\") pod \"13e0242d-c42e-4aad-95a2-62de95d87e42\" (UID: \"13e0242d-c42e-4aad-95a2-62de95d87e42\") " Dec 06 06:19:11 crc kubenswrapper[4809]: I1206 06:19:11.523031 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13e0242d-c42e-4aad-95a2-62de95d87e42-kube-api-access-4bxt9" (OuterVolumeSpecName: "kube-api-access-4bxt9") pod "13e0242d-c42e-4aad-95a2-62de95d87e42" (UID: "13e0242d-c42e-4aad-95a2-62de95d87e42"). InnerVolumeSpecName "kube-api-access-4bxt9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:19:11 crc kubenswrapper[4809]: I1206 06:19:11.523739 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13e0242d-c42e-4aad-95a2-62de95d87e42-scripts" (OuterVolumeSpecName: "scripts") pod "13e0242d-c42e-4aad-95a2-62de95d87e42" (UID: "13e0242d-c42e-4aad-95a2-62de95d87e42"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:19:11 crc kubenswrapper[4809]: I1206 06:19:11.621539 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4bxt9\" (UniqueName: \"kubernetes.io/projected/13e0242d-c42e-4aad-95a2-62de95d87e42-kube-api-access-4bxt9\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:11 crc kubenswrapper[4809]: I1206 06:19:11.621577 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13e0242d-c42e-4aad-95a2-62de95d87e42-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:11 crc kubenswrapper[4809]: I1206 06:19:11.864570 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-vt99m" event={"ID":"13e0242d-c42e-4aad-95a2-62de95d87e42","Type":"ContainerDied","Data":"5feeb302e32bc6c326fbdd2d33d8d85d6cabddb04bc953da8551a17a3e501fb8"} Dec 06 06:19:11 crc kubenswrapper[4809]: I1206 06:19:11.864627 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5feeb302e32bc6c326fbdd2d33d8d85d6cabddb04bc953da8551a17a3e501fb8" Dec 06 06:19:11 crc kubenswrapper[4809]: I1206 06:19:11.864649 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-vt99m" Dec 06 06:19:11 crc kubenswrapper[4809]: I1206 06:19:11.881327 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13e0242d-c42e-4aad-95a2-62de95d87e42-config-data" (OuterVolumeSpecName: "config-data") pod "13e0242d-c42e-4aad-95a2-62de95d87e42" (UID: "13e0242d-c42e-4aad-95a2-62de95d87e42"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:19:11 crc kubenswrapper[4809]: I1206 06:19:11.882971 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13e0242d-c42e-4aad-95a2-62de95d87e42-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "13e0242d-c42e-4aad-95a2-62de95d87e42" (UID: "13e0242d-c42e-4aad-95a2-62de95d87e42"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:19:11 crc kubenswrapper[4809]: I1206 06:19:11.928464 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13e0242d-c42e-4aad-95a2-62de95d87e42-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:11 crc kubenswrapper[4809]: I1206 06:19:11.928510 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13e0242d-c42e-4aad-95a2-62de95d87e42-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:11 crc kubenswrapper[4809]: I1206 06:19:11.994436 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 06 06:19:11 crc kubenswrapper[4809]: E1206 06:19:11.994908 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13e0242d-c42e-4aad-95a2-62de95d87e42" containerName="nova-cell0-conductor-db-sync" Dec 06 06:19:11 crc kubenswrapper[4809]: I1206 06:19:11.994950 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="13e0242d-c42e-4aad-95a2-62de95d87e42" containerName="nova-cell0-conductor-db-sync" Dec 06 06:19:11 crc kubenswrapper[4809]: E1206 06:19:11.994973 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4adfb97-c4f0-4093-80c5-e0229644834c" containerName="extract-utilities" Dec 06 06:19:11 crc kubenswrapper[4809]: I1206 06:19:11.994983 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4adfb97-c4f0-4093-80c5-e0229644834c" containerName="extract-utilities" Dec 06 06:19:11 crc kubenswrapper[4809]: E1206 06:19:11.995000 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4adfb97-c4f0-4093-80c5-e0229644834c" containerName="registry-server" Dec 06 06:19:11 crc kubenswrapper[4809]: I1206 06:19:11.995007 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4adfb97-c4f0-4093-80c5-e0229644834c" containerName="registry-server" Dec 06 06:19:11 crc kubenswrapper[4809]: E1206 06:19:11.995015 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4adfb97-c4f0-4093-80c5-e0229644834c" containerName="extract-content" Dec 06 06:19:11 crc kubenswrapper[4809]: I1206 06:19:11.995021 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4adfb97-c4f0-4093-80c5-e0229644834c" containerName="extract-content" Dec 06 06:19:11 crc kubenswrapper[4809]: I1206 06:19:11.995242 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="13e0242d-c42e-4aad-95a2-62de95d87e42" containerName="nova-cell0-conductor-db-sync" Dec 06 06:19:11 crc kubenswrapper[4809]: I1206 06:19:11.995266 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4adfb97-c4f0-4093-80c5-e0229644834c" containerName="registry-server" Dec 06 06:19:11 crc kubenswrapper[4809]: I1206 06:19:11.997113 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 06 06:19:12 crc kubenswrapper[4809]: I1206 06:19:12.013472 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 06 06:19:12 crc kubenswrapper[4809]: I1206 06:19:12.133810 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-md7cc\" (UniqueName: \"kubernetes.io/projected/fc561fa6-2f1c-41b6-8b04-b8aecf53d25a-kube-api-access-md7cc\") pod \"nova-cell0-conductor-0\" (UID: \"fc561fa6-2f1c-41b6-8b04-b8aecf53d25a\") " pod="openstack/nova-cell0-conductor-0" Dec 06 06:19:12 crc kubenswrapper[4809]: I1206 06:19:12.134422 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc561fa6-2f1c-41b6-8b04-b8aecf53d25a-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"fc561fa6-2f1c-41b6-8b04-b8aecf53d25a\") " pod="openstack/nova-cell0-conductor-0" Dec 06 06:19:12 crc kubenswrapper[4809]: I1206 06:19:12.134515 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc561fa6-2f1c-41b6-8b04-b8aecf53d25a-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"fc561fa6-2f1c-41b6-8b04-b8aecf53d25a\") " pod="openstack/nova-cell0-conductor-0" Dec 06 06:19:12 crc kubenswrapper[4809]: I1206 06:19:12.237068 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-md7cc\" (UniqueName: \"kubernetes.io/projected/fc561fa6-2f1c-41b6-8b04-b8aecf53d25a-kube-api-access-md7cc\") pod \"nova-cell0-conductor-0\" (UID: \"fc561fa6-2f1c-41b6-8b04-b8aecf53d25a\") " pod="openstack/nova-cell0-conductor-0" Dec 06 06:19:12 crc kubenswrapper[4809]: I1206 06:19:12.237163 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc561fa6-2f1c-41b6-8b04-b8aecf53d25a-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"fc561fa6-2f1c-41b6-8b04-b8aecf53d25a\") " pod="openstack/nova-cell0-conductor-0" Dec 06 06:19:12 crc kubenswrapper[4809]: I1206 06:19:12.237187 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc561fa6-2f1c-41b6-8b04-b8aecf53d25a-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"fc561fa6-2f1c-41b6-8b04-b8aecf53d25a\") " pod="openstack/nova-cell0-conductor-0" Dec 06 06:19:12 crc kubenswrapper[4809]: I1206 06:19:12.241786 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc561fa6-2f1c-41b6-8b04-b8aecf53d25a-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"fc561fa6-2f1c-41b6-8b04-b8aecf53d25a\") " pod="openstack/nova-cell0-conductor-0" Dec 06 06:19:12 crc kubenswrapper[4809]: I1206 06:19:12.242145 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc561fa6-2f1c-41b6-8b04-b8aecf53d25a-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"fc561fa6-2f1c-41b6-8b04-b8aecf53d25a\") " pod="openstack/nova-cell0-conductor-0" Dec 06 06:19:12 crc kubenswrapper[4809]: I1206 06:19:12.255188 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-md7cc\" (UniqueName: \"kubernetes.io/projected/fc561fa6-2f1c-41b6-8b04-b8aecf53d25a-kube-api-access-md7cc\") pod \"nova-cell0-conductor-0\" (UID: \"fc561fa6-2f1c-41b6-8b04-b8aecf53d25a\") " pod="openstack/nova-cell0-conductor-0" Dec 06 06:19:12 crc kubenswrapper[4809]: I1206 06:19:12.315752 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 06 06:19:12 crc kubenswrapper[4809]: I1206 06:19:12.772942 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 06 06:19:12 crc kubenswrapper[4809]: I1206 06:19:12.886134 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"fc561fa6-2f1c-41b6-8b04-b8aecf53d25a","Type":"ContainerStarted","Data":"8b7dd1940269a82f24b6e7432745149184834f0efd3bfa35d692383b4ccf9231"} Dec 06 06:19:13 crc kubenswrapper[4809]: I1206 06:19:13.897248 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"fc561fa6-2f1c-41b6-8b04-b8aecf53d25a","Type":"ContainerStarted","Data":"a5a46fd606605307406543c7e97daa22a4d4abd77d99480a8605107d3c65f201"} Dec 06 06:19:13 crc kubenswrapper[4809]: I1206 06:19:13.897600 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 06 06:19:13 crc kubenswrapper[4809]: I1206 06:19:13.923220 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.923200702 podStartE2EDuration="2.923200702s" podCreationTimestamp="2025-12-06 06:19:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:19:13.918459254 +0000 UTC m=+1678.807442196" watchObservedRunningTime="2025-12-06 06:19:13.923200702 +0000 UTC m=+1678.812183644" Dec 06 06:19:16 crc kubenswrapper[4809]: I1206 06:19:16.954355 4809 generic.go:334] "Generic (PLEG): container finished" podID="438fc849-1432-45d3-8709-4b9e4734e56f" containerID="2eaf3473d57f772f14f96c219113e3f7b1dc1bf47459cd506632caa4b1b48dc1" exitCode=0 Dec 06 06:19:16 crc kubenswrapper[4809]: I1206 06:19:16.954435 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"438fc849-1432-45d3-8709-4b9e4734e56f","Type":"ContainerDied","Data":"2eaf3473d57f772f14f96c219113e3f7b1dc1bf47459cd506632caa4b1b48dc1"} Dec 06 06:19:16 crc kubenswrapper[4809]: I1206 06:19:16.954855 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"438fc849-1432-45d3-8709-4b9e4734e56f","Type":"ContainerDied","Data":"b3951cb22277a2e28993378fa15999ef2ef04c53df89c85b8d0947ff920deaf7"} Dec 06 06:19:16 crc kubenswrapper[4809]: I1206 06:19:16.954871 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b3951cb22277a2e28993378fa15999ef2ef04c53df89c85b8d0947ff920deaf7" Dec 06 06:19:17 crc kubenswrapper[4809]: I1206 06:19:17.030428 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:19:17 crc kubenswrapper[4809]: I1206 06:19:17.156127 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8bmv8\" (UniqueName: \"kubernetes.io/projected/438fc849-1432-45d3-8709-4b9e4734e56f-kube-api-access-8bmv8\") pod \"438fc849-1432-45d3-8709-4b9e4734e56f\" (UID: \"438fc849-1432-45d3-8709-4b9e4734e56f\") " Dec 06 06:19:17 crc kubenswrapper[4809]: I1206 06:19:17.156195 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/438fc849-1432-45d3-8709-4b9e4734e56f-config-data\") pod \"438fc849-1432-45d3-8709-4b9e4734e56f\" (UID: \"438fc849-1432-45d3-8709-4b9e4734e56f\") " Dec 06 06:19:17 crc kubenswrapper[4809]: I1206 06:19:17.156244 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/438fc849-1432-45d3-8709-4b9e4734e56f-run-httpd\") pod \"438fc849-1432-45d3-8709-4b9e4734e56f\" (UID: \"438fc849-1432-45d3-8709-4b9e4734e56f\") " Dec 06 06:19:17 crc kubenswrapper[4809]: I1206 06:19:17.156332 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/438fc849-1432-45d3-8709-4b9e4734e56f-log-httpd\") pod \"438fc849-1432-45d3-8709-4b9e4734e56f\" (UID: \"438fc849-1432-45d3-8709-4b9e4734e56f\") " Dec 06 06:19:17 crc kubenswrapper[4809]: I1206 06:19:17.156378 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/438fc849-1432-45d3-8709-4b9e4734e56f-sg-core-conf-yaml\") pod \"438fc849-1432-45d3-8709-4b9e4734e56f\" (UID: \"438fc849-1432-45d3-8709-4b9e4734e56f\") " Dec 06 06:19:17 crc kubenswrapper[4809]: I1206 06:19:17.156450 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/438fc849-1432-45d3-8709-4b9e4734e56f-combined-ca-bundle\") pod \"438fc849-1432-45d3-8709-4b9e4734e56f\" (UID: \"438fc849-1432-45d3-8709-4b9e4734e56f\") " Dec 06 06:19:17 crc kubenswrapper[4809]: I1206 06:19:17.156473 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/438fc849-1432-45d3-8709-4b9e4734e56f-scripts\") pod \"438fc849-1432-45d3-8709-4b9e4734e56f\" (UID: \"438fc849-1432-45d3-8709-4b9e4734e56f\") " Dec 06 06:19:17 crc kubenswrapper[4809]: I1206 06:19:17.156642 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/438fc849-1432-45d3-8709-4b9e4734e56f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "438fc849-1432-45d3-8709-4b9e4734e56f" (UID: "438fc849-1432-45d3-8709-4b9e4734e56f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:19:17 crc kubenswrapper[4809]: I1206 06:19:17.156916 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/438fc849-1432-45d3-8709-4b9e4734e56f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "438fc849-1432-45d3-8709-4b9e4734e56f" (UID: "438fc849-1432-45d3-8709-4b9e4734e56f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:19:17 crc kubenswrapper[4809]: I1206 06:19:17.157219 4809 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/438fc849-1432-45d3-8709-4b9e4734e56f-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:17 crc kubenswrapper[4809]: I1206 06:19:17.157234 4809 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/438fc849-1432-45d3-8709-4b9e4734e56f-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:17 crc kubenswrapper[4809]: I1206 06:19:17.162358 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/438fc849-1432-45d3-8709-4b9e4734e56f-kube-api-access-8bmv8" (OuterVolumeSpecName: "kube-api-access-8bmv8") pod "438fc849-1432-45d3-8709-4b9e4734e56f" (UID: "438fc849-1432-45d3-8709-4b9e4734e56f"). InnerVolumeSpecName "kube-api-access-8bmv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:19:17 crc kubenswrapper[4809]: I1206 06:19:17.174145 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/438fc849-1432-45d3-8709-4b9e4734e56f-scripts" (OuterVolumeSpecName: "scripts") pod "438fc849-1432-45d3-8709-4b9e4734e56f" (UID: "438fc849-1432-45d3-8709-4b9e4734e56f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:19:17 crc kubenswrapper[4809]: I1206 06:19:17.192747 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/438fc849-1432-45d3-8709-4b9e4734e56f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "438fc849-1432-45d3-8709-4b9e4734e56f" (UID: "438fc849-1432-45d3-8709-4b9e4734e56f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:19:17 crc kubenswrapper[4809]: I1206 06:19:17.249073 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/438fc849-1432-45d3-8709-4b9e4734e56f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "438fc849-1432-45d3-8709-4b9e4734e56f" (UID: "438fc849-1432-45d3-8709-4b9e4734e56f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:19:17 crc kubenswrapper[4809]: I1206 06:19:17.258896 4809 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/438fc849-1432-45d3-8709-4b9e4734e56f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:17 crc kubenswrapper[4809]: I1206 06:19:17.258974 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/438fc849-1432-45d3-8709-4b9e4734e56f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:17 crc kubenswrapper[4809]: I1206 06:19:17.258987 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/438fc849-1432-45d3-8709-4b9e4734e56f-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:17 crc kubenswrapper[4809]: I1206 06:19:17.258997 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8bmv8\" (UniqueName: \"kubernetes.io/projected/438fc849-1432-45d3-8709-4b9e4734e56f-kube-api-access-8bmv8\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:17 crc kubenswrapper[4809]: I1206 06:19:17.274519 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/438fc849-1432-45d3-8709-4b9e4734e56f-config-data" (OuterVolumeSpecName: "config-data") pod "438fc849-1432-45d3-8709-4b9e4734e56f" (UID: "438fc849-1432-45d3-8709-4b9e4734e56f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:19:17 crc kubenswrapper[4809]: I1206 06:19:17.361294 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/438fc849-1432-45d3-8709-4b9e4734e56f-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:17 crc kubenswrapper[4809]: I1206 06:19:17.964513 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:19:18 crc kubenswrapper[4809]: I1206 06:19:18.005739 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:19:18 crc kubenswrapper[4809]: I1206 06:19:18.018001 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:19:18 crc kubenswrapper[4809]: I1206 06:19:18.030892 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:19:18 crc kubenswrapper[4809]: E1206 06:19:18.031417 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="438fc849-1432-45d3-8709-4b9e4734e56f" containerName="ceilometer-notification-agent" Dec 06 06:19:18 crc kubenswrapper[4809]: I1206 06:19:18.031447 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="438fc849-1432-45d3-8709-4b9e4734e56f" containerName="ceilometer-notification-agent" Dec 06 06:19:18 crc kubenswrapper[4809]: E1206 06:19:18.031469 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="438fc849-1432-45d3-8709-4b9e4734e56f" containerName="sg-core" Dec 06 06:19:18 crc kubenswrapper[4809]: I1206 06:19:18.031479 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="438fc849-1432-45d3-8709-4b9e4734e56f" containerName="sg-core" Dec 06 06:19:18 crc kubenswrapper[4809]: E1206 06:19:18.031507 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="438fc849-1432-45d3-8709-4b9e4734e56f" containerName="ceilometer-central-agent" Dec 06 06:19:18 crc kubenswrapper[4809]: I1206 06:19:18.031515 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="438fc849-1432-45d3-8709-4b9e4734e56f" containerName="ceilometer-central-agent" Dec 06 06:19:18 crc kubenswrapper[4809]: E1206 06:19:18.031531 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="438fc849-1432-45d3-8709-4b9e4734e56f" containerName="proxy-httpd" Dec 06 06:19:18 crc kubenswrapper[4809]: I1206 06:19:18.031539 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="438fc849-1432-45d3-8709-4b9e4734e56f" containerName="proxy-httpd" Dec 06 06:19:18 crc kubenswrapper[4809]: I1206 06:19:18.031818 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="438fc849-1432-45d3-8709-4b9e4734e56f" containerName="sg-core" Dec 06 06:19:18 crc kubenswrapper[4809]: I1206 06:19:18.031842 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="438fc849-1432-45d3-8709-4b9e4734e56f" containerName="proxy-httpd" Dec 06 06:19:18 crc kubenswrapper[4809]: I1206 06:19:18.031868 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="438fc849-1432-45d3-8709-4b9e4734e56f" containerName="ceilometer-notification-agent" Dec 06 06:19:18 crc kubenswrapper[4809]: I1206 06:19:18.031891 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="438fc849-1432-45d3-8709-4b9e4734e56f" containerName="ceilometer-central-agent" Dec 06 06:19:18 crc kubenswrapper[4809]: I1206 06:19:18.033980 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:19:18 crc kubenswrapper[4809]: I1206 06:19:18.042251 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 06:19:18 crc kubenswrapper[4809]: I1206 06:19:18.043134 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 06:19:18 crc kubenswrapper[4809]: I1206 06:19:18.060306 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:19:18 crc kubenswrapper[4809]: I1206 06:19:18.178421 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9170461-4a1f-43da-be11-12310d9b2574-log-httpd\") pod \"ceilometer-0\" (UID: \"a9170461-4a1f-43da-be11-12310d9b2574\") " pod="openstack/ceilometer-0" Dec 06 06:19:18 crc kubenswrapper[4809]: I1206 06:19:18.178541 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a9170461-4a1f-43da-be11-12310d9b2574-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a9170461-4a1f-43da-be11-12310d9b2574\") " pod="openstack/ceilometer-0" Dec 06 06:19:18 crc kubenswrapper[4809]: I1206 06:19:18.178573 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9170461-4a1f-43da-be11-12310d9b2574-run-httpd\") pod \"ceilometer-0\" (UID: \"a9170461-4a1f-43da-be11-12310d9b2574\") " pod="openstack/ceilometer-0" Dec 06 06:19:18 crc kubenswrapper[4809]: I1206 06:19:18.178621 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9170461-4a1f-43da-be11-12310d9b2574-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a9170461-4a1f-43da-be11-12310d9b2574\") " pod="openstack/ceilometer-0" Dec 06 06:19:18 crc kubenswrapper[4809]: I1206 06:19:18.178670 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hjhg\" (UniqueName: \"kubernetes.io/projected/a9170461-4a1f-43da-be11-12310d9b2574-kube-api-access-4hjhg\") pod \"ceilometer-0\" (UID: \"a9170461-4a1f-43da-be11-12310d9b2574\") " pod="openstack/ceilometer-0" Dec 06 06:19:18 crc kubenswrapper[4809]: I1206 06:19:18.178783 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9170461-4a1f-43da-be11-12310d9b2574-scripts\") pod \"ceilometer-0\" (UID: \"a9170461-4a1f-43da-be11-12310d9b2574\") " pod="openstack/ceilometer-0" Dec 06 06:19:18 crc kubenswrapper[4809]: I1206 06:19:18.178841 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9170461-4a1f-43da-be11-12310d9b2574-config-data\") pod \"ceilometer-0\" (UID: \"a9170461-4a1f-43da-be11-12310d9b2574\") " pod="openstack/ceilometer-0" Dec 06 06:19:18 crc kubenswrapper[4809]: I1206 06:19:18.281125 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hjhg\" (UniqueName: \"kubernetes.io/projected/a9170461-4a1f-43da-be11-12310d9b2574-kube-api-access-4hjhg\") pod \"ceilometer-0\" (UID: \"a9170461-4a1f-43da-be11-12310d9b2574\") " pod="openstack/ceilometer-0" Dec 06 06:19:18 crc kubenswrapper[4809]: I1206 06:19:18.281305 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9170461-4a1f-43da-be11-12310d9b2574-scripts\") pod \"ceilometer-0\" (UID: \"a9170461-4a1f-43da-be11-12310d9b2574\") " pod="openstack/ceilometer-0" Dec 06 06:19:18 crc kubenswrapper[4809]: I1206 06:19:18.281360 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9170461-4a1f-43da-be11-12310d9b2574-config-data\") pod \"ceilometer-0\" (UID: \"a9170461-4a1f-43da-be11-12310d9b2574\") " pod="openstack/ceilometer-0" Dec 06 06:19:18 crc kubenswrapper[4809]: I1206 06:19:18.281415 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9170461-4a1f-43da-be11-12310d9b2574-log-httpd\") pod \"ceilometer-0\" (UID: \"a9170461-4a1f-43da-be11-12310d9b2574\") " pod="openstack/ceilometer-0" Dec 06 06:19:18 crc kubenswrapper[4809]: I1206 06:19:18.281475 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a9170461-4a1f-43da-be11-12310d9b2574-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a9170461-4a1f-43da-be11-12310d9b2574\") " pod="openstack/ceilometer-0" Dec 06 06:19:18 crc kubenswrapper[4809]: I1206 06:19:18.281510 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9170461-4a1f-43da-be11-12310d9b2574-run-httpd\") pod \"ceilometer-0\" (UID: \"a9170461-4a1f-43da-be11-12310d9b2574\") " pod="openstack/ceilometer-0" Dec 06 06:19:18 crc kubenswrapper[4809]: I1206 06:19:18.282052 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9170461-4a1f-43da-be11-12310d9b2574-log-httpd\") pod \"ceilometer-0\" (UID: \"a9170461-4a1f-43da-be11-12310d9b2574\") " pod="openstack/ceilometer-0" Dec 06 06:19:18 crc kubenswrapper[4809]: I1206 06:19:18.282161 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9170461-4a1f-43da-be11-12310d9b2574-run-httpd\") pod \"ceilometer-0\" (UID: \"a9170461-4a1f-43da-be11-12310d9b2574\") " pod="openstack/ceilometer-0" Dec 06 06:19:18 crc kubenswrapper[4809]: I1206 06:19:18.282280 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9170461-4a1f-43da-be11-12310d9b2574-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a9170461-4a1f-43da-be11-12310d9b2574\") " pod="openstack/ceilometer-0" Dec 06 06:19:18 crc kubenswrapper[4809]: I1206 06:19:18.296922 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a9170461-4a1f-43da-be11-12310d9b2574-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a9170461-4a1f-43da-be11-12310d9b2574\") " pod="openstack/ceilometer-0" Dec 06 06:19:18 crc kubenswrapper[4809]: I1206 06:19:18.299083 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9170461-4a1f-43da-be11-12310d9b2574-scripts\") pod \"ceilometer-0\" (UID: \"a9170461-4a1f-43da-be11-12310d9b2574\") " pod="openstack/ceilometer-0" Dec 06 06:19:18 crc kubenswrapper[4809]: I1206 06:19:18.299562 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9170461-4a1f-43da-be11-12310d9b2574-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a9170461-4a1f-43da-be11-12310d9b2574\") " pod="openstack/ceilometer-0" Dec 06 06:19:18 crc kubenswrapper[4809]: I1206 06:19:18.300034 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9170461-4a1f-43da-be11-12310d9b2574-config-data\") pod \"ceilometer-0\" (UID: \"a9170461-4a1f-43da-be11-12310d9b2574\") " pod="openstack/ceilometer-0" Dec 06 06:19:18 crc kubenswrapper[4809]: I1206 06:19:18.304902 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hjhg\" (UniqueName: \"kubernetes.io/projected/a9170461-4a1f-43da-be11-12310d9b2574-kube-api-access-4hjhg\") pod \"ceilometer-0\" (UID: \"a9170461-4a1f-43da-be11-12310d9b2574\") " pod="openstack/ceilometer-0" Dec 06 06:19:18 crc kubenswrapper[4809]: I1206 06:19:18.377991 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:19:18 crc kubenswrapper[4809]: I1206 06:19:18.978569 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:19:18 crc kubenswrapper[4809]: W1206 06:19:18.985140 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda9170461_4a1f_43da_be11_12310d9b2574.slice/crio-64a4628b5a00fd07a1219a7d59cc7304f0a9c0db0073c3acdca7e0141f01e5b5 WatchSource:0}: Error finding container 64a4628b5a00fd07a1219a7d59cc7304f0a9c0db0073c3acdca7e0141f01e5b5: Status 404 returned error can't find the container with id 64a4628b5a00fd07a1219a7d59cc7304f0a9c0db0073c3acdca7e0141f01e5b5 Dec 06 06:19:19 crc kubenswrapper[4809]: I1206 06:19:19.389020 4809 scope.go:117] "RemoveContainer" containerID="9bc2dc88b3b9f309ce0169aa510326dd77022dece4c69cab3141df00fe5c0106" Dec 06 06:19:19 crc kubenswrapper[4809]: E1206 06:19:19.389839 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:19:19 crc kubenswrapper[4809]: I1206 06:19:19.403482 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="438fc849-1432-45d3-8709-4b9e4734e56f" path="/var/lib/kubelet/pods/438fc849-1432-45d3-8709-4b9e4734e56f/volumes" Dec 06 06:19:19 crc kubenswrapper[4809]: I1206 06:19:19.535741 4809 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a] : Timed out while waiting for systemd to remove kubepods-besteffort-pod515fc75f_1ad6_41e0_bfd5_4a6e27d75d7a.slice" Dec 06 06:19:19 crc kubenswrapper[4809]: E1206 06:19:19.535800 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort pod515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a] : unable to destroy cgroup paths for cgroup [kubepods besteffort pod515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a] : Timed out while waiting for systemd to remove kubepods-besteffort-pod515fc75f_1ad6_41e0_bfd5_4a6e27d75d7a.slice" pod="openstack/heat-engine-7c94c94748-57b42" podUID="515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a" Dec 06 06:19:19 crc kubenswrapper[4809]: I1206 06:19:19.562505 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-ld8fs"] Dec 06 06:19:19 crc kubenswrapper[4809]: I1206 06:19:19.564406 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-ld8fs" Dec 06 06:19:19 crc kubenswrapper[4809]: I1206 06:19:19.582984 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-ld8fs"] Dec 06 06:19:19 crc kubenswrapper[4809]: I1206 06:19:19.658967 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-133e-account-create-update-cpzqq"] Dec 06 06:19:19 crc kubenswrapper[4809]: I1206 06:19:19.660743 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-133e-account-create-update-cpzqq" Dec 06 06:19:19 crc kubenswrapper[4809]: I1206 06:19:19.663910 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Dec 06 06:19:19 crc kubenswrapper[4809]: I1206 06:19:19.691172 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-133e-account-create-update-cpzqq"] Dec 06 06:19:19 crc kubenswrapper[4809]: I1206 06:19:19.715567 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjttx\" (UniqueName: \"kubernetes.io/projected/62e3de39-6323-4837-b746-b2b51b7d8e6c-kube-api-access-rjttx\") pod \"aodh-133e-account-create-update-cpzqq\" (UID: \"62e3de39-6323-4837-b746-b2b51b7d8e6c\") " pod="openstack/aodh-133e-account-create-update-cpzqq" Dec 06 06:19:19 crc kubenswrapper[4809]: I1206 06:19:19.715671 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84tqg\" (UniqueName: \"kubernetes.io/projected/353be8e6-45e2-4731-bdf0-62a5260adf96-kube-api-access-84tqg\") pod \"aodh-db-create-ld8fs\" (UID: \"353be8e6-45e2-4731-bdf0-62a5260adf96\") " pod="openstack/aodh-db-create-ld8fs" Dec 06 06:19:19 crc kubenswrapper[4809]: I1206 06:19:19.715952 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/62e3de39-6323-4837-b746-b2b51b7d8e6c-operator-scripts\") pod \"aodh-133e-account-create-update-cpzqq\" (UID: \"62e3de39-6323-4837-b746-b2b51b7d8e6c\") " pod="openstack/aodh-133e-account-create-update-cpzqq" Dec 06 06:19:19 crc kubenswrapper[4809]: I1206 06:19:19.716157 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/353be8e6-45e2-4731-bdf0-62a5260adf96-operator-scripts\") pod \"aodh-db-create-ld8fs\" (UID: \"353be8e6-45e2-4731-bdf0-62a5260adf96\") " pod="openstack/aodh-db-create-ld8fs" Dec 06 06:19:19 crc kubenswrapper[4809]: I1206 06:19:19.818603 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjttx\" (UniqueName: \"kubernetes.io/projected/62e3de39-6323-4837-b746-b2b51b7d8e6c-kube-api-access-rjttx\") pod \"aodh-133e-account-create-update-cpzqq\" (UID: \"62e3de39-6323-4837-b746-b2b51b7d8e6c\") " pod="openstack/aodh-133e-account-create-update-cpzqq" Dec 06 06:19:19 crc kubenswrapper[4809]: I1206 06:19:19.818703 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84tqg\" (UniqueName: \"kubernetes.io/projected/353be8e6-45e2-4731-bdf0-62a5260adf96-kube-api-access-84tqg\") pod \"aodh-db-create-ld8fs\" (UID: \"353be8e6-45e2-4731-bdf0-62a5260adf96\") " pod="openstack/aodh-db-create-ld8fs" Dec 06 06:19:19 crc kubenswrapper[4809]: I1206 06:19:19.818773 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/62e3de39-6323-4837-b746-b2b51b7d8e6c-operator-scripts\") pod \"aodh-133e-account-create-update-cpzqq\" (UID: \"62e3de39-6323-4837-b746-b2b51b7d8e6c\") " pod="openstack/aodh-133e-account-create-update-cpzqq" Dec 06 06:19:19 crc kubenswrapper[4809]: I1206 06:19:19.818828 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/353be8e6-45e2-4731-bdf0-62a5260adf96-operator-scripts\") pod \"aodh-db-create-ld8fs\" (UID: \"353be8e6-45e2-4731-bdf0-62a5260adf96\") " pod="openstack/aodh-db-create-ld8fs" Dec 06 06:19:19 crc kubenswrapper[4809]: I1206 06:19:19.819504 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/353be8e6-45e2-4731-bdf0-62a5260adf96-operator-scripts\") pod \"aodh-db-create-ld8fs\" (UID: \"353be8e6-45e2-4731-bdf0-62a5260adf96\") " pod="openstack/aodh-db-create-ld8fs" Dec 06 06:19:19 crc kubenswrapper[4809]: I1206 06:19:19.819916 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/62e3de39-6323-4837-b746-b2b51b7d8e6c-operator-scripts\") pod \"aodh-133e-account-create-update-cpzqq\" (UID: \"62e3de39-6323-4837-b746-b2b51b7d8e6c\") " pod="openstack/aodh-133e-account-create-update-cpzqq" Dec 06 06:19:19 crc kubenswrapper[4809]: I1206 06:19:19.837516 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84tqg\" (UniqueName: \"kubernetes.io/projected/353be8e6-45e2-4731-bdf0-62a5260adf96-kube-api-access-84tqg\") pod \"aodh-db-create-ld8fs\" (UID: \"353be8e6-45e2-4731-bdf0-62a5260adf96\") " pod="openstack/aodh-db-create-ld8fs" Dec 06 06:19:19 crc kubenswrapper[4809]: I1206 06:19:19.837554 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjttx\" (UniqueName: \"kubernetes.io/projected/62e3de39-6323-4837-b746-b2b51b7d8e6c-kube-api-access-rjttx\") pod \"aodh-133e-account-create-update-cpzqq\" (UID: \"62e3de39-6323-4837-b746-b2b51b7d8e6c\") " pod="openstack/aodh-133e-account-create-update-cpzqq" Dec 06 06:19:19 crc kubenswrapper[4809]: I1206 06:19:19.907984 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-ld8fs" Dec 06 06:19:19 crc kubenswrapper[4809]: I1206 06:19:19.990707 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-133e-account-create-update-cpzqq" Dec 06 06:19:20 crc kubenswrapper[4809]: I1206 06:19:20.007112 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7c94c94748-57b42" Dec 06 06:19:20 crc kubenswrapper[4809]: I1206 06:19:20.007101 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9170461-4a1f-43da-be11-12310d9b2574","Type":"ContainerStarted","Data":"e697b371434aa54fc302370afd097c1a6a1ab6c62b417790b09e40130cd2672e"} Dec 06 06:19:20 crc kubenswrapper[4809]: I1206 06:19:20.007266 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9170461-4a1f-43da-be11-12310d9b2574","Type":"ContainerStarted","Data":"64a4628b5a00fd07a1219a7d59cc7304f0a9c0db0073c3acdca7e0141f01e5b5"} Dec 06 06:19:20 crc kubenswrapper[4809]: I1206 06:19:20.085299 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-7c94c94748-57b42"] Dec 06 06:19:20 crc kubenswrapper[4809]: I1206 06:19:20.122588 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-7c94c94748-57b42"] Dec 06 06:19:20 crc kubenswrapper[4809]: W1206 06:19:20.515319 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod353be8e6_45e2_4731_bdf0_62a5260adf96.slice/crio-be06baabca79a43302fab63eced084413493c36af4933e458cb57a8a7b760776 WatchSource:0}: Error finding container be06baabca79a43302fab63eced084413493c36af4933e458cb57a8a7b760776: Status 404 returned error can't find the container with id be06baabca79a43302fab63eced084413493c36af4933e458cb57a8a7b760776 Dec 06 06:19:20 crc kubenswrapper[4809]: I1206 06:19:20.518776 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-ld8fs"] Dec 06 06:19:20 crc kubenswrapper[4809]: I1206 06:19:20.697970 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-133e-account-create-update-cpzqq"] Dec 06 06:19:20 crc kubenswrapper[4809]: W1206 06:19:20.701157 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod62e3de39_6323_4837_b746_b2b51b7d8e6c.slice/crio-0285011a124f4f765f99a7e438b1c1cce0dd5fbe98972efa07fef43ebd0397c3 WatchSource:0}: Error finding container 0285011a124f4f765f99a7e438b1c1cce0dd5fbe98972efa07fef43ebd0397c3: Status 404 returned error can't find the container with id 0285011a124f4f765f99a7e438b1c1cce0dd5fbe98972efa07fef43ebd0397c3 Dec 06 06:19:21 crc kubenswrapper[4809]: I1206 06:19:21.029123 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9170461-4a1f-43da-be11-12310d9b2574","Type":"ContainerStarted","Data":"db6f189fb52dcee256ce499fe10fafdebc05ca8f3f7a49453b7e2a39cf2ca71c"} Dec 06 06:19:21 crc kubenswrapper[4809]: I1206 06:19:21.031522 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-ld8fs" event={"ID":"353be8e6-45e2-4731-bdf0-62a5260adf96","Type":"ContainerStarted","Data":"be06baabca79a43302fab63eced084413493c36af4933e458cb57a8a7b760776"} Dec 06 06:19:21 crc kubenswrapper[4809]: I1206 06:19:21.035435 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-133e-account-create-update-cpzqq" event={"ID":"62e3de39-6323-4837-b746-b2b51b7d8e6c","Type":"ContainerStarted","Data":"6ff2f5a5901ee62a2d83e64767490215843e438ad6b3ed23aad29389d4960353"} Dec 06 06:19:21 crc kubenswrapper[4809]: I1206 06:19:21.035494 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-133e-account-create-update-cpzqq" event={"ID":"62e3de39-6323-4837-b746-b2b51b7d8e6c","Type":"ContainerStarted","Data":"0285011a124f4f765f99a7e438b1c1cce0dd5fbe98972efa07fef43ebd0397c3"} Dec 06 06:19:21 crc kubenswrapper[4809]: I1206 06:19:21.055203 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-create-ld8fs" podStartSLOduration=2.055183889 podStartE2EDuration="2.055183889s" podCreationTimestamp="2025-12-06 06:19:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:19:21.054789588 +0000 UTC m=+1685.943772530" watchObservedRunningTime="2025-12-06 06:19:21.055183889 +0000 UTC m=+1685.944166841" Dec 06 06:19:21 crc kubenswrapper[4809]: I1206 06:19:21.085236 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-133e-account-create-update-cpzqq" podStartSLOduration=2.085211809 podStartE2EDuration="2.085211809s" podCreationTimestamp="2025-12-06 06:19:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:19:21.07413383 +0000 UTC m=+1685.963116772" watchObservedRunningTime="2025-12-06 06:19:21.085211809 +0000 UTC m=+1685.974194751" Dec 06 06:19:21 crc kubenswrapper[4809]: I1206 06:19:21.415387 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a" path="/var/lib/kubelet/pods/515fc75f-1ad6-41e0-bfd5-4a6e27d75d7a/volumes" Dec 06 06:19:22 crc kubenswrapper[4809]: I1206 06:19:22.048449 4809 generic.go:334] "Generic (PLEG): container finished" podID="353be8e6-45e2-4731-bdf0-62a5260adf96" containerID="71c474112cf4eb39fccdbfb9f48003b4719be0290686f158fed9729a0af7ebad" exitCode=0 Dec 06 06:19:22 crc kubenswrapper[4809]: I1206 06:19:22.048626 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-ld8fs" event={"ID":"353be8e6-45e2-4731-bdf0-62a5260adf96","Type":"ContainerDied","Data":"71c474112cf4eb39fccdbfb9f48003b4719be0290686f158fed9729a0af7ebad"} Dec 06 06:19:22 crc kubenswrapper[4809]: I1206 06:19:22.050869 4809 generic.go:334] "Generic (PLEG): container finished" podID="62e3de39-6323-4837-b746-b2b51b7d8e6c" containerID="6ff2f5a5901ee62a2d83e64767490215843e438ad6b3ed23aad29389d4960353" exitCode=0 Dec 06 06:19:22 crc kubenswrapper[4809]: I1206 06:19:22.050945 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-133e-account-create-update-cpzqq" event={"ID":"62e3de39-6323-4837-b746-b2b51b7d8e6c","Type":"ContainerDied","Data":"6ff2f5a5901ee62a2d83e64767490215843e438ad6b3ed23aad29389d4960353"} Dec 06 06:19:22 crc kubenswrapper[4809]: I1206 06:19:22.053290 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9170461-4a1f-43da-be11-12310d9b2574","Type":"ContainerStarted","Data":"c1b0d9118ab641766dd069635e54219a8078aba3d89dd6a49c1a4633398b7b99"} Dec 06 06:19:22 crc kubenswrapper[4809]: I1206 06:19:22.346783 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 06 06:19:22 crc kubenswrapper[4809]: I1206 06:19:22.830791 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-x8zvs"] Dec 06 06:19:22 crc kubenswrapper[4809]: I1206 06:19:22.833144 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-x8zvs" Dec 06 06:19:22 crc kubenswrapper[4809]: I1206 06:19:22.837254 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 06 06:19:22 crc kubenswrapper[4809]: I1206 06:19:22.840027 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 06 06:19:22 crc kubenswrapper[4809]: I1206 06:19:22.854129 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-x8zvs"] Dec 06 06:19:22 crc kubenswrapper[4809]: I1206 06:19:22.903716 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dad7df76-c6ae-4d0c-aef7-63ca9efb77a7-config-data\") pod \"nova-cell0-cell-mapping-x8zvs\" (UID: \"dad7df76-c6ae-4d0c-aef7-63ca9efb77a7\") " pod="openstack/nova-cell0-cell-mapping-x8zvs" Dec 06 06:19:22 crc kubenswrapper[4809]: I1206 06:19:22.903838 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dad7df76-c6ae-4d0c-aef7-63ca9efb77a7-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-x8zvs\" (UID: \"dad7df76-c6ae-4d0c-aef7-63ca9efb77a7\") " pod="openstack/nova-cell0-cell-mapping-x8zvs" Dec 06 06:19:22 crc kubenswrapper[4809]: I1206 06:19:22.903887 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dad7df76-c6ae-4d0c-aef7-63ca9efb77a7-scripts\") pod \"nova-cell0-cell-mapping-x8zvs\" (UID: \"dad7df76-c6ae-4d0c-aef7-63ca9efb77a7\") " pod="openstack/nova-cell0-cell-mapping-x8zvs" Dec 06 06:19:22 crc kubenswrapper[4809]: I1206 06:19:22.903926 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5q64s\" (UniqueName: \"kubernetes.io/projected/dad7df76-c6ae-4d0c-aef7-63ca9efb77a7-kube-api-access-5q64s\") pod \"nova-cell0-cell-mapping-x8zvs\" (UID: \"dad7df76-c6ae-4d0c-aef7-63ca9efb77a7\") " pod="openstack/nova-cell0-cell-mapping-x8zvs" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.006108 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dad7df76-c6ae-4d0c-aef7-63ca9efb77a7-config-data\") pod \"nova-cell0-cell-mapping-x8zvs\" (UID: \"dad7df76-c6ae-4d0c-aef7-63ca9efb77a7\") " pod="openstack/nova-cell0-cell-mapping-x8zvs" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.006226 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dad7df76-c6ae-4d0c-aef7-63ca9efb77a7-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-x8zvs\" (UID: \"dad7df76-c6ae-4d0c-aef7-63ca9efb77a7\") " pod="openstack/nova-cell0-cell-mapping-x8zvs" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.006280 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dad7df76-c6ae-4d0c-aef7-63ca9efb77a7-scripts\") pod \"nova-cell0-cell-mapping-x8zvs\" (UID: \"dad7df76-c6ae-4d0c-aef7-63ca9efb77a7\") " pod="openstack/nova-cell0-cell-mapping-x8zvs" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.006318 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5q64s\" (UniqueName: \"kubernetes.io/projected/dad7df76-c6ae-4d0c-aef7-63ca9efb77a7-kube-api-access-5q64s\") pod \"nova-cell0-cell-mapping-x8zvs\" (UID: \"dad7df76-c6ae-4d0c-aef7-63ca9efb77a7\") " pod="openstack/nova-cell0-cell-mapping-x8zvs" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.020036 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dad7df76-c6ae-4d0c-aef7-63ca9efb77a7-scripts\") pod \"nova-cell0-cell-mapping-x8zvs\" (UID: \"dad7df76-c6ae-4d0c-aef7-63ca9efb77a7\") " pod="openstack/nova-cell0-cell-mapping-x8zvs" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.034119 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dad7df76-c6ae-4d0c-aef7-63ca9efb77a7-config-data\") pod \"nova-cell0-cell-mapping-x8zvs\" (UID: \"dad7df76-c6ae-4d0c-aef7-63ca9efb77a7\") " pod="openstack/nova-cell0-cell-mapping-x8zvs" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.046810 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5q64s\" (UniqueName: \"kubernetes.io/projected/dad7df76-c6ae-4d0c-aef7-63ca9efb77a7-kube-api-access-5q64s\") pod \"nova-cell0-cell-mapping-x8zvs\" (UID: \"dad7df76-c6ae-4d0c-aef7-63ca9efb77a7\") " pod="openstack/nova-cell0-cell-mapping-x8zvs" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.056537 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dad7df76-c6ae-4d0c-aef7-63ca9efb77a7-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-x8zvs\" (UID: \"dad7df76-c6ae-4d0c-aef7-63ca9efb77a7\") " pod="openstack/nova-cell0-cell-mapping-x8zvs" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.087157 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.089023 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.094752 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9170461-4a1f-43da-be11-12310d9b2574","Type":"ContainerStarted","Data":"d094101e80ac7e9b6b8cfe518288078b17e1d05fcf9784669c5e643c32b4b06e"} Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.094952 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.121164 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.122818 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.138333 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.145177 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.160778 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-x8zvs" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.167542 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.218739 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b09b87b-a70c-43d1-936d-0758596953aa-config-data\") pod \"nova-metadata-0\" (UID: \"2b09b87b-a70c-43d1-936d-0758596953aa\") " pod="openstack/nova-metadata-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.218862 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b09b87b-a70c-43d1-936d-0758596953aa-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2b09b87b-a70c-43d1-936d-0758596953aa\") " pod="openstack/nova-metadata-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.218949 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b09b87b-a70c-43d1-936d-0758596953aa-logs\") pod \"nova-metadata-0\" (UID: \"2b09b87b-a70c-43d1-936d-0758596953aa\") " pod="openstack/nova-metadata-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.219033 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdlxd\" (UniqueName: \"kubernetes.io/projected/1063337d-4069-4953-87af-504675fd2680-kube-api-access-gdlxd\") pod \"nova-api-0\" (UID: \"1063337d-4069-4953-87af-504675fd2680\") " pod="openstack/nova-api-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.219105 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g62t4\" (UniqueName: \"kubernetes.io/projected/2b09b87b-a70c-43d1-936d-0758596953aa-kube-api-access-g62t4\") pod \"nova-metadata-0\" (UID: \"2b09b87b-a70c-43d1-936d-0758596953aa\") " pod="openstack/nova-metadata-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.219149 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1063337d-4069-4953-87af-504675fd2680-config-data\") pod \"nova-api-0\" (UID: \"1063337d-4069-4953-87af-504675fd2680\") " pod="openstack/nova-api-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.219390 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1063337d-4069-4953-87af-504675fd2680-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1063337d-4069-4953-87af-504675fd2680\") " pod="openstack/nova-api-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.219428 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1063337d-4069-4953-87af-504675fd2680-logs\") pod \"nova-api-0\" (UID: \"1063337d-4069-4953-87af-504675fd2680\") " pod="openstack/nova-api-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.232028 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.260890 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.262348 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.310431 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.348302 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vcbp\" (UniqueName: \"kubernetes.io/projected/d1eb5524-8750-4c35-8342-e93796d69d35-kube-api-access-8vcbp\") pod \"nova-scheduler-0\" (UID: \"d1eb5524-8750-4c35-8342-e93796d69d35\") " pod="openstack/nova-scheduler-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.348581 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1063337d-4069-4953-87af-504675fd2680-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1063337d-4069-4953-87af-504675fd2680\") " pod="openstack/nova-api-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.348636 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1063337d-4069-4953-87af-504675fd2680-logs\") pod \"nova-api-0\" (UID: \"1063337d-4069-4953-87af-504675fd2680\") " pod="openstack/nova-api-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.348674 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b09b87b-a70c-43d1-936d-0758596953aa-config-data\") pod \"nova-metadata-0\" (UID: \"2b09b87b-a70c-43d1-936d-0758596953aa\") " pod="openstack/nova-metadata-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.348761 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b09b87b-a70c-43d1-936d-0758596953aa-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2b09b87b-a70c-43d1-936d-0758596953aa\") " pod="openstack/nova-metadata-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.348849 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1eb5524-8750-4c35-8342-e93796d69d35-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d1eb5524-8750-4c35-8342-e93796d69d35\") " pod="openstack/nova-scheduler-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.348883 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b09b87b-a70c-43d1-936d-0758596953aa-logs\") pod \"nova-metadata-0\" (UID: \"2b09b87b-a70c-43d1-936d-0758596953aa\") " pod="openstack/nova-metadata-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.348994 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdlxd\" (UniqueName: \"kubernetes.io/projected/1063337d-4069-4953-87af-504675fd2680-kube-api-access-gdlxd\") pod \"nova-api-0\" (UID: \"1063337d-4069-4953-87af-504675fd2680\") " pod="openstack/nova-api-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.349077 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g62t4\" (UniqueName: \"kubernetes.io/projected/2b09b87b-a70c-43d1-936d-0758596953aa-kube-api-access-g62t4\") pod \"nova-metadata-0\" (UID: \"2b09b87b-a70c-43d1-936d-0758596953aa\") " pod="openstack/nova-metadata-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.349132 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1063337d-4069-4953-87af-504675fd2680-config-data\") pod \"nova-api-0\" (UID: \"1063337d-4069-4953-87af-504675fd2680\") " pod="openstack/nova-api-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.349305 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1eb5524-8750-4c35-8342-e93796d69d35-config-data\") pod \"nova-scheduler-0\" (UID: \"d1eb5524-8750-4c35-8342-e93796d69d35\") " pod="openstack/nova-scheduler-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.358337 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1063337d-4069-4953-87af-504675fd2680-logs\") pod \"nova-api-0\" (UID: \"1063337d-4069-4953-87af-504675fd2680\") " pod="openstack/nova-api-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.361160 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b09b87b-a70c-43d1-936d-0758596953aa-logs\") pod \"nova-metadata-0\" (UID: \"2b09b87b-a70c-43d1-936d-0758596953aa\") " pod="openstack/nova-metadata-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.391675 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b09b87b-a70c-43d1-936d-0758596953aa-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2b09b87b-a70c-43d1-936d-0758596953aa\") " pod="openstack/nova-metadata-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.401653 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1063337d-4069-4953-87af-504675fd2680-config-data\") pod \"nova-api-0\" (UID: \"1063337d-4069-4953-87af-504675fd2680\") " pod="openstack/nova-api-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.423719 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdlxd\" (UniqueName: \"kubernetes.io/projected/1063337d-4069-4953-87af-504675fd2680-kube-api-access-gdlxd\") pod \"nova-api-0\" (UID: \"1063337d-4069-4953-87af-504675fd2680\") " pod="openstack/nova-api-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.424920 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g62t4\" (UniqueName: \"kubernetes.io/projected/2b09b87b-a70c-43d1-936d-0758596953aa-kube-api-access-g62t4\") pod \"nova-metadata-0\" (UID: \"2b09b87b-a70c-43d1-936d-0758596953aa\") " pod="openstack/nova-metadata-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.451696 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b09b87b-a70c-43d1-936d-0758596953aa-config-data\") pod \"nova-metadata-0\" (UID: \"2b09b87b-a70c-43d1-936d-0758596953aa\") " pod="openstack/nova-metadata-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.452310 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.222160384 podStartE2EDuration="6.452287841s" podCreationTimestamp="2025-12-06 06:19:17 +0000 UTC" firstStartedPulling="2025-12-06 06:19:18.987692829 +0000 UTC m=+1683.876675771" lastFinishedPulling="2025-12-06 06:19:22.217820286 +0000 UTC m=+1687.106803228" observedRunningTime="2025-12-06 06:19:23.232636505 +0000 UTC m=+1688.121619447" watchObservedRunningTime="2025-12-06 06:19:23.452287841 +0000 UTC m=+1688.341270783" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.454887 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1eb5524-8750-4c35-8342-e93796d69d35-config-data\") pod \"nova-scheduler-0\" (UID: \"d1eb5524-8750-4c35-8342-e93796d69d35\") " pod="openstack/nova-scheduler-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.455037 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1063337d-4069-4953-87af-504675fd2680-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1063337d-4069-4953-87af-504675fd2680\") " pod="openstack/nova-api-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.470727 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vcbp\" (UniqueName: \"kubernetes.io/projected/d1eb5524-8750-4c35-8342-e93796d69d35-kube-api-access-8vcbp\") pod \"nova-scheduler-0\" (UID: \"d1eb5524-8750-4c35-8342-e93796d69d35\") " pod="openstack/nova-scheduler-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.471081 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1eb5524-8750-4c35-8342-e93796d69d35-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d1eb5524-8750-4c35-8342-e93796d69d35\") " pod="openstack/nova-scheduler-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.479099 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.484482 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1eb5524-8750-4c35-8342-e93796d69d35-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d1eb5524-8750-4c35-8342-e93796d69d35\") " pod="openstack/nova-scheduler-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.489584 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1eb5524-8750-4c35-8342-e93796d69d35-config-data\") pod \"nova-scheduler-0\" (UID: \"d1eb5524-8750-4c35-8342-e93796d69d35\") " pod="openstack/nova-scheduler-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.503765 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-fg9jq"] Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.511184 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-fg9jq" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.543284 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.544544 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vcbp\" (UniqueName: \"kubernetes.io/projected/d1eb5524-8750-4c35-8342-e93796d69d35-kube-api-access-8vcbp\") pod \"nova-scheduler-0\" (UID: \"d1eb5524-8750-4c35-8342-e93796d69d35\") " pod="openstack/nova-scheduler-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.563599 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.619961 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.622097 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.634644 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.643536 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.651425 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.678406 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-fg9jq"] Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.711021 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37c89662-6b35-4e73-bf37-f6bb01978c5f-ovsdbserver-nb\") pod \"dnsmasq-dns-9b86998b5-fg9jq\" (UID: \"37c89662-6b35-4e73-bf37-f6bb01978c5f\") " pod="openstack/dnsmasq-dns-9b86998b5-fg9jq" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.711179 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/37c89662-6b35-4e73-bf37-f6bb01978c5f-dns-swift-storage-0\") pod \"dnsmasq-dns-9b86998b5-fg9jq\" (UID: \"37c89662-6b35-4e73-bf37-f6bb01978c5f\") " pod="openstack/dnsmasq-dns-9b86998b5-fg9jq" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.711369 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27mh5\" (UniqueName: \"kubernetes.io/projected/37c89662-6b35-4e73-bf37-f6bb01978c5f-kube-api-access-27mh5\") pod \"dnsmasq-dns-9b86998b5-fg9jq\" (UID: \"37c89662-6b35-4e73-bf37-f6bb01978c5f\") " pod="openstack/dnsmasq-dns-9b86998b5-fg9jq" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.711418 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37c89662-6b35-4e73-bf37-f6bb01978c5f-config\") pod \"dnsmasq-dns-9b86998b5-fg9jq\" (UID: \"37c89662-6b35-4e73-bf37-f6bb01978c5f\") " pod="openstack/dnsmasq-dns-9b86998b5-fg9jq" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.711601 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37c89662-6b35-4e73-bf37-f6bb01978c5f-ovsdbserver-sb\") pod \"dnsmasq-dns-9b86998b5-fg9jq\" (UID: \"37c89662-6b35-4e73-bf37-f6bb01978c5f\") " pod="openstack/dnsmasq-dns-9b86998b5-fg9jq" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.711646 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37c89662-6b35-4e73-bf37-f6bb01978c5f-dns-svc\") pod \"dnsmasq-dns-9b86998b5-fg9jq\" (UID: \"37c89662-6b35-4e73-bf37-f6bb01978c5f\") " pod="openstack/dnsmasq-dns-9b86998b5-fg9jq" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.817608 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37c89662-6b35-4e73-bf37-f6bb01978c5f-ovsdbserver-nb\") pod \"dnsmasq-dns-9b86998b5-fg9jq\" (UID: \"37c89662-6b35-4e73-bf37-f6bb01978c5f\") " pod="openstack/dnsmasq-dns-9b86998b5-fg9jq" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.818168 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/37c89662-6b35-4e73-bf37-f6bb01978c5f-dns-swift-storage-0\") pod \"dnsmasq-dns-9b86998b5-fg9jq\" (UID: \"37c89662-6b35-4e73-bf37-f6bb01978c5f\") " pod="openstack/dnsmasq-dns-9b86998b5-fg9jq" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.818862 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37c89662-6b35-4e73-bf37-f6bb01978c5f-ovsdbserver-nb\") pod \"dnsmasq-dns-9b86998b5-fg9jq\" (UID: \"37c89662-6b35-4e73-bf37-f6bb01978c5f\") " pod="openstack/dnsmasq-dns-9b86998b5-fg9jq" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.819084 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/37c89662-6b35-4e73-bf37-f6bb01978c5f-dns-swift-storage-0\") pod \"dnsmasq-dns-9b86998b5-fg9jq\" (UID: \"37c89662-6b35-4e73-bf37-f6bb01978c5f\") " pod="openstack/dnsmasq-dns-9b86998b5-fg9jq" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.819555 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6980646-cff8-45b6-a26b-53d6282f26a9-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"b6980646-cff8-45b6-a26b-53d6282f26a9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.819651 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27mh5\" (UniqueName: \"kubernetes.io/projected/37c89662-6b35-4e73-bf37-f6bb01978c5f-kube-api-access-27mh5\") pod \"dnsmasq-dns-9b86998b5-fg9jq\" (UID: \"37c89662-6b35-4e73-bf37-f6bb01978c5f\") " pod="openstack/dnsmasq-dns-9b86998b5-fg9jq" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.819694 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37c89662-6b35-4e73-bf37-f6bb01978c5f-config\") pod \"dnsmasq-dns-9b86998b5-fg9jq\" (UID: \"37c89662-6b35-4e73-bf37-f6bb01978c5f\") " pod="openstack/dnsmasq-dns-9b86998b5-fg9jq" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.819858 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2kdb\" (UniqueName: \"kubernetes.io/projected/b6980646-cff8-45b6-a26b-53d6282f26a9-kube-api-access-w2kdb\") pod \"nova-cell1-novncproxy-0\" (UID: \"b6980646-cff8-45b6-a26b-53d6282f26a9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.820625 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37c89662-6b35-4e73-bf37-f6bb01978c5f-ovsdbserver-sb\") pod \"dnsmasq-dns-9b86998b5-fg9jq\" (UID: \"37c89662-6b35-4e73-bf37-f6bb01978c5f\") " pod="openstack/dnsmasq-dns-9b86998b5-fg9jq" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.819925 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37c89662-6b35-4e73-bf37-f6bb01978c5f-ovsdbserver-sb\") pod \"dnsmasq-dns-9b86998b5-fg9jq\" (UID: \"37c89662-6b35-4e73-bf37-f6bb01978c5f\") " pod="openstack/dnsmasq-dns-9b86998b5-fg9jq" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.823659 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37c89662-6b35-4e73-bf37-f6bb01978c5f-dns-svc\") pod \"dnsmasq-dns-9b86998b5-fg9jq\" (UID: \"37c89662-6b35-4e73-bf37-f6bb01978c5f\") " pod="openstack/dnsmasq-dns-9b86998b5-fg9jq" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.823734 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6980646-cff8-45b6-a26b-53d6282f26a9-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"b6980646-cff8-45b6-a26b-53d6282f26a9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.825283 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37c89662-6b35-4e73-bf37-f6bb01978c5f-dns-svc\") pod \"dnsmasq-dns-9b86998b5-fg9jq\" (UID: \"37c89662-6b35-4e73-bf37-f6bb01978c5f\") " pod="openstack/dnsmasq-dns-9b86998b5-fg9jq" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.830243 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37c89662-6b35-4e73-bf37-f6bb01978c5f-config\") pod \"dnsmasq-dns-9b86998b5-fg9jq\" (UID: \"37c89662-6b35-4e73-bf37-f6bb01978c5f\") " pod="openstack/dnsmasq-dns-9b86998b5-fg9jq" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.849118 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27mh5\" (UniqueName: \"kubernetes.io/projected/37c89662-6b35-4e73-bf37-f6bb01978c5f-kube-api-access-27mh5\") pod \"dnsmasq-dns-9b86998b5-fg9jq\" (UID: \"37c89662-6b35-4e73-bf37-f6bb01978c5f\") " pod="openstack/dnsmasq-dns-9b86998b5-fg9jq" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.927660 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6980646-cff8-45b6-a26b-53d6282f26a9-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"b6980646-cff8-45b6-a26b-53d6282f26a9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.927771 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2kdb\" (UniqueName: \"kubernetes.io/projected/b6980646-cff8-45b6-a26b-53d6282f26a9-kube-api-access-w2kdb\") pod \"nova-cell1-novncproxy-0\" (UID: \"b6980646-cff8-45b6-a26b-53d6282f26a9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.927837 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6980646-cff8-45b6-a26b-53d6282f26a9-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"b6980646-cff8-45b6-a26b-53d6282f26a9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.946543 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6980646-cff8-45b6-a26b-53d6282f26a9-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"b6980646-cff8-45b6-a26b-53d6282f26a9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.950135 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6980646-cff8-45b6-a26b-53d6282f26a9-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"b6980646-cff8-45b6-a26b-53d6282f26a9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.972026 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2kdb\" (UniqueName: \"kubernetes.io/projected/b6980646-cff8-45b6-a26b-53d6282f26a9-kube-api-access-w2kdb\") pod \"nova-cell1-novncproxy-0\" (UID: \"b6980646-cff8-45b6-a26b-53d6282f26a9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.975410 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-fg9jq" Dec 06 06:19:23 crc kubenswrapper[4809]: I1206 06:19:23.984750 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:19:24 crc kubenswrapper[4809]: I1206 06:19:24.459181 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-133e-account-create-update-cpzqq" Dec 06 06:19:24 crc kubenswrapper[4809]: I1206 06:19:24.484920 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-ld8fs" Dec 06 06:19:24 crc kubenswrapper[4809]: I1206 06:19:24.571256 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rjttx\" (UniqueName: \"kubernetes.io/projected/62e3de39-6323-4837-b746-b2b51b7d8e6c-kube-api-access-rjttx\") pod \"62e3de39-6323-4837-b746-b2b51b7d8e6c\" (UID: \"62e3de39-6323-4837-b746-b2b51b7d8e6c\") " Dec 06 06:19:24 crc kubenswrapper[4809]: I1206 06:19:24.571345 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/62e3de39-6323-4837-b746-b2b51b7d8e6c-operator-scripts\") pod \"62e3de39-6323-4837-b746-b2b51b7d8e6c\" (UID: \"62e3de39-6323-4837-b746-b2b51b7d8e6c\") " Dec 06 06:19:24 crc kubenswrapper[4809]: I1206 06:19:24.575642 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62e3de39-6323-4837-b746-b2b51b7d8e6c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "62e3de39-6323-4837-b746-b2b51b7d8e6c" (UID: "62e3de39-6323-4837-b746-b2b51b7d8e6c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:19:24 crc kubenswrapper[4809]: I1206 06:19:24.606037 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62e3de39-6323-4837-b746-b2b51b7d8e6c-kube-api-access-rjttx" (OuterVolumeSpecName: "kube-api-access-rjttx") pod "62e3de39-6323-4837-b746-b2b51b7d8e6c" (UID: "62e3de39-6323-4837-b746-b2b51b7d8e6c"). InnerVolumeSpecName "kube-api-access-rjttx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:19:24 crc kubenswrapper[4809]: I1206 06:19:24.675018 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-84tqg\" (UniqueName: \"kubernetes.io/projected/353be8e6-45e2-4731-bdf0-62a5260adf96-kube-api-access-84tqg\") pod \"353be8e6-45e2-4731-bdf0-62a5260adf96\" (UID: \"353be8e6-45e2-4731-bdf0-62a5260adf96\") " Dec 06 06:19:24 crc kubenswrapper[4809]: I1206 06:19:24.675125 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/353be8e6-45e2-4731-bdf0-62a5260adf96-operator-scripts\") pod \"353be8e6-45e2-4731-bdf0-62a5260adf96\" (UID: \"353be8e6-45e2-4731-bdf0-62a5260adf96\") " Dec 06 06:19:24 crc kubenswrapper[4809]: I1206 06:19:24.675883 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rjttx\" (UniqueName: \"kubernetes.io/projected/62e3de39-6323-4837-b746-b2b51b7d8e6c-kube-api-access-rjttx\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:24 crc kubenswrapper[4809]: I1206 06:19:24.675904 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/62e3de39-6323-4837-b746-b2b51b7d8e6c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:24 crc kubenswrapper[4809]: I1206 06:19:24.676584 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/353be8e6-45e2-4731-bdf0-62a5260adf96-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "353be8e6-45e2-4731-bdf0-62a5260adf96" (UID: "353be8e6-45e2-4731-bdf0-62a5260adf96"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:19:24 crc kubenswrapper[4809]: I1206 06:19:24.684735 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/353be8e6-45e2-4731-bdf0-62a5260adf96-kube-api-access-84tqg" (OuterVolumeSpecName: "kube-api-access-84tqg") pod "353be8e6-45e2-4731-bdf0-62a5260adf96" (UID: "353be8e6-45e2-4731-bdf0-62a5260adf96"). InnerVolumeSpecName "kube-api-access-84tqg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:19:24 crc kubenswrapper[4809]: I1206 06:19:24.778273 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-84tqg\" (UniqueName: \"kubernetes.io/projected/353be8e6-45e2-4731-bdf0-62a5260adf96-kube-api-access-84tqg\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:24 crc kubenswrapper[4809]: I1206 06:19:24.778297 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/353be8e6-45e2-4731-bdf0-62a5260adf96-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:25 crc kubenswrapper[4809]: I1206 06:19:25.157849 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-x8zvs"] Dec 06 06:19:25 crc kubenswrapper[4809]: I1206 06:19:25.202823 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 06:19:25 crc kubenswrapper[4809]: I1206 06:19:25.293986 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 06:19:25 crc kubenswrapper[4809]: I1206 06:19:25.301603 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 06:19:25 crc kubenswrapper[4809]: I1206 06:19:25.311679 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 06:19:25 crc kubenswrapper[4809]: I1206 06:19:25.318483 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-x8zvs" event={"ID":"dad7df76-c6ae-4d0c-aef7-63ca9efb77a7","Type":"ContainerStarted","Data":"08844096fda39abf61f3e0470d8015cffc93a65cb1f4549dbde5b178cc3e1012"} Dec 06 06:19:25 crc kubenswrapper[4809]: I1206 06:19:25.330799 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-ld8fs" event={"ID":"353be8e6-45e2-4731-bdf0-62a5260adf96","Type":"ContainerDied","Data":"be06baabca79a43302fab63eced084413493c36af4933e458cb57a8a7b760776"} Dec 06 06:19:25 crc kubenswrapper[4809]: I1206 06:19:25.330839 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be06baabca79a43302fab63eced084413493c36af4933e458cb57a8a7b760776" Dec 06 06:19:25 crc kubenswrapper[4809]: I1206 06:19:25.330917 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-ld8fs" Dec 06 06:19:25 crc kubenswrapper[4809]: W1206 06:19:25.342506 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb6980646_cff8_45b6_a26b_53d6282f26a9.slice/crio-c30e1938b5b8b25ef56f2066873a9ad7a9eff7cd7c024e5d375f8720edf838c5 WatchSource:0}: Error finding container c30e1938b5b8b25ef56f2066873a9ad7a9eff7cd7c024e5d375f8720edf838c5: Status 404 returned error can't find the container with id c30e1938b5b8b25ef56f2066873a9ad7a9eff7cd7c024e5d375f8720edf838c5 Dec 06 06:19:25 crc kubenswrapper[4809]: I1206 06:19:25.344502 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-133e-account-create-update-cpzqq" event={"ID":"62e3de39-6323-4837-b746-b2b51b7d8e6c","Type":"ContainerDied","Data":"0285011a124f4f765f99a7e438b1c1cce0dd5fbe98972efa07fef43ebd0397c3"} Dec 06 06:19:25 crc kubenswrapper[4809]: I1206 06:19:25.344545 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0285011a124f4f765f99a7e438b1c1cce0dd5fbe98972efa07fef43ebd0397c3" Dec 06 06:19:25 crc kubenswrapper[4809]: I1206 06:19:25.345281 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-133e-account-create-update-cpzqq" Dec 06 06:19:25 crc kubenswrapper[4809]: I1206 06:19:25.360818 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-fg9jq"] Dec 06 06:19:25 crc kubenswrapper[4809]: I1206 06:19:25.761004 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8jvh2"] Dec 06 06:19:25 crc kubenswrapper[4809]: E1206 06:19:25.761758 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="353be8e6-45e2-4731-bdf0-62a5260adf96" containerName="mariadb-database-create" Dec 06 06:19:25 crc kubenswrapper[4809]: I1206 06:19:25.761770 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="353be8e6-45e2-4731-bdf0-62a5260adf96" containerName="mariadb-database-create" Dec 06 06:19:25 crc kubenswrapper[4809]: E1206 06:19:25.761794 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62e3de39-6323-4837-b746-b2b51b7d8e6c" containerName="mariadb-account-create-update" Dec 06 06:19:25 crc kubenswrapper[4809]: I1206 06:19:25.761800 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="62e3de39-6323-4837-b746-b2b51b7d8e6c" containerName="mariadb-account-create-update" Dec 06 06:19:25 crc kubenswrapper[4809]: I1206 06:19:25.762193 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="62e3de39-6323-4837-b746-b2b51b7d8e6c" containerName="mariadb-account-create-update" Dec 06 06:19:25 crc kubenswrapper[4809]: I1206 06:19:25.762208 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="353be8e6-45e2-4731-bdf0-62a5260adf96" containerName="mariadb-database-create" Dec 06 06:19:25 crc kubenswrapper[4809]: I1206 06:19:25.763245 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8jvh2" Dec 06 06:19:25 crc kubenswrapper[4809]: I1206 06:19:25.766520 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 06 06:19:25 crc kubenswrapper[4809]: I1206 06:19:25.767043 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 06 06:19:25 crc kubenswrapper[4809]: I1206 06:19:25.792006 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8jvh2"] Dec 06 06:19:25 crc kubenswrapper[4809]: I1206 06:19:25.929946 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4af32e7-96e5-4f86-b271-6b4efcace0dc-config-data\") pod \"nova-cell1-conductor-db-sync-8jvh2\" (UID: \"d4af32e7-96e5-4f86-b271-6b4efcace0dc\") " pod="openstack/nova-cell1-conductor-db-sync-8jvh2" Dec 06 06:19:25 crc kubenswrapper[4809]: I1206 06:19:25.930397 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4af32e7-96e5-4f86-b271-6b4efcace0dc-scripts\") pod \"nova-cell1-conductor-db-sync-8jvh2\" (UID: \"d4af32e7-96e5-4f86-b271-6b4efcace0dc\") " pod="openstack/nova-cell1-conductor-db-sync-8jvh2" Dec 06 06:19:25 crc kubenswrapper[4809]: I1206 06:19:25.930441 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbhtf\" (UniqueName: \"kubernetes.io/projected/d4af32e7-96e5-4f86-b271-6b4efcace0dc-kube-api-access-gbhtf\") pod \"nova-cell1-conductor-db-sync-8jvh2\" (UID: \"d4af32e7-96e5-4f86-b271-6b4efcace0dc\") " pod="openstack/nova-cell1-conductor-db-sync-8jvh2" Dec 06 06:19:25 crc kubenswrapper[4809]: I1206 06:19:25.931866 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4af32e7-96e5-4f86-b271-6b4efcace0dc-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-8jvh2\" (UID: \"d4af32e7-96e5-4f86-b271-6b4efcace0dc\") " pod="openstack/nova-cell1-conductor-db-sync-8jvh2" Dec 06 06:19:26 crc kubenswrapper[4809]: I1206 06:19:26.035106 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4af32e7-96e5-4f86-b271-6b4efcace0dc-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-8jvh2\" (UID: \"d4af32e7-96e5-4f86-b271-6b4efcace0dc\") " pod="openstack/nova-cell1-conductor-db-sync-8jvh2" Dec 06 06:19:26 crc kubenswrapper[4809]: I1206 06:19:26.035228 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4af32e7-96e5-4f86-b271-6b4efcace0dc-config-data\") pod \"nova-cell1-conductor-db-sync-8jvh2\" (UID: \"d4af32e7-96e5-4f86-b271-6b4efcace0dc\") " pod="openstack/nova-cell1-conductor-db-sync-8jvh2" Dec 06 06:19:26 crc kubenswrapper[4809]: I1206 06:19:26.035317 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4af32e7-96e5-4f86-b271-6b4efcace0dc-scripts\") pod \"nova-cell1-conductor-db-sync-8jvh2\" (UID: \"d4af32e7-96e5-4f86-b271-6b4efcace0dc\") " pod="openstack/nova-cell1-conductor-db-sync-8jvh2" Dec 06 06:19:26 crc kubenswrapper[4809]: I1206 06:19:26.035355 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbhtf\" (UniqueName: \"kubernetes.io/projected/d4af32e7-96e5-4f86-b271-6b4efcace0dc-kube-api-access-gbhtf\") pod \"nova-cell1-conductor-db-sync-8jvh2\" (UID: \"d4af32e7-96e5-4f86-b271-6b4efcace0dc\") " pod="openstack/nova-cell1-conductor-db-sync-8jvh2" Dec 06 06:19:26 crc kubenswrapper[4809]: I1206 06:19:26.046715 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4af32e7-96e5-4f86-b271-6b4efcace0dc-config-data\") pod \"nova-cell1-conductor-db-sync-8jvh2\" (UID: \"d4af32e7-96e5-4f86-b271-6b4efcace0dc\") " pod="openstack/nova-cell1-conductor-db-sync-8jvh2" Dec 06 06:19:26 crc kubenswrapper[4809]: I1206 06:19:26.047071 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4af32e7-96e5-4f86-b271-6b4efcace0dc-scripts\") pod \"nova-cell1-conductor-db-sync-8jvh2\" (UID: \"d4af32e7-96e5-4f86-b271-6b4efcace0dc\") " pod="openstack/nova-cell1-conductor-db-sync-8jvh2" Dec 06 06:19:26 crc kubenswrapper[4809]: I1206 06:19:26.048903 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4af32e7-96e5-4f86-b271-6b4efcace0dc-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-8jvh2\" (UID: \"d4af32e7-96e5-4f86-b271-6b4efcace0dc\") " pod="openstack/nova-cell1-conductor-db-sync-8jvh2" Dec 06 06:19:26 crc kubenswrapper[4809]: I1206 06:19:26.062534 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbhtf\" (UniqueName: \"kubernetes.io/projected/d4af32e7-96e5-4f86-b271-6b4efcace0dc-kube-api-access-gbhtf\") pod \"nova-cell1-conductor-db-sync-8jvh2\" (UID: \"d4af32e7-96e5-4f86-b271-6b4efcace0dc\") " pod="openstack/nova-cell1-conductor-db-sync-8jvh2" Dec 06 06:19:26 crc kubenswrapper[4809]: I1206 06:19:26.090727 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8jvh2" Dec 06 06:19:26 crc kubenswrapper[4809]: I1206 06:19:26.367760 4809 generic.go:334] "Generic (PLEG): container finished" podID="37c89662-6b35-4e73-bf37-f6bb01978c5f" containerID="4608c06dd7d29f002c9151795ccc44dba1da393d792ee6eb2cdbba4b65ea7485" exitCode=0 Dec 06 06:19:26 crc kubenswrapper[4809]: I1206 06:19:26.369940 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-fg9jq" event={"ID":"37c89662-6b35-4e73-bf37-f6bb01978c5f","Type":"ContainerDied","Data":"4608c06dd7d29f002c9151795ccc44dba1da393d792ee6eb2cdbba4b65ea7485"} Dec 06 06:19:26 crc kubenswrapper[4809]: I1206 06:19:26.369976 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-fg9jq" event={"ID":"37c89662-6b35-4e73-bf37-f6bb01978c5f","Type":"ContainerStarted","Data":"a3b2aeabf80e82647cf1efb7c757498fecab574b5cc8fa45f0d14668ecea2846"} Dec 06 06:19:26 crc kubenswrapper[4809]: I1206 06:19:26.384139 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1063337d-4069-4953-87af-504675fd2680","Type":"ContainerStarted","Data":"5ba0e0184a3c76b27ac0fd21ddb91c91db328ddb1530ed2251e69c4a5c21a378"} Dec 06 06:19:26 crc kubenswrapper[4809]: I1206 06:19:26.402011 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2b09b87b-a70c-43d1-936d-0758596953aa","Type":"ContainerStarted","Data":"ed09ea49ce462f1c0c5f1b4e23e4e276708a81fa9aa242f26d82305d8cea4ba7"} Dec 06 06:19:26 crc kubenswrapper[4809]: I1206 06:19:26.429049 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d1eb5524-8750-4c35-8342-e93796d69d35","Type":"ContainerStarted","Data":"f41a50de85789e4ec57df7dd785e71ba62570ead8965f9259bf661263028b52a"} Dec 06 06:19:26 crc kubenswrapper[4809]: I1206 06:19:26.443527 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"b6980646-cff8-45b6-a26b-53d6282f26a9","Type":"ContainerStarted","Data":"c30e1938b5b8b25ef56f2066873a9ad7a9eff7cd7c024e5d375f8720edf838c5"} Dec 06 06:19:26 crc kubenswrapper[4809]: I1206 06:19:26.468405 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-x8zvs" event={"ID":"dad7df76-c6ae-4d0c-aef7-63ca9efb77a7","Type":"ContainerStarted","Data":"0c398009cd796b9757f33073f5ae46f28d2c1cb725130ab094f4dd9d64918527"} Dec 06 06:19:26 crc kubenswrapper[4809]: I1206 06:19:26.498135 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-x8zvs" podStartSLOduration=4.498115766 podStartE2EDuration="4.498115766s" podCreationTimestamp="2025-12-06 06:19:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:19:26.488821635 +0000 UTC m=+1691.377804577" watchObservedRunningTime="2025-12-06 06:19:26.498115766 +0000 UTC m=+1691.387098708" Dec 06 06:19:26 crc kubenswrapper[4809]: W1206 06:19:26.736216 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd4af32e7_96e5_4f86_b271_6b4efcace0dc.slice/crio-553b7c1b21cbc2fabc21f34229e517151f16b7342e9d9b4e8fc713118ef833ed WatchSource:0}: Error finding container 553b7c1b21cbc2fabc21f34229e517151f16b7342e9d9b4e8fc713118ef833ed: Status 404 returned error can't find the container with id 553b7c1b21cbc2fabc21f34229e517151f16b7342e9d9b4e8fc713118ef833ed Dec 06 06:19:26 crc kubenswrapper[4809]: I1206 06:19:26.753914 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8jvh2"] Dec 06 06:19:27 crc kubenswrapper[4809]: I1206 06:19:27.493652 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-fg9jq" event={"ID":"37c89662-6b35-4e73-bf37-f6bb01978c5f","Type":"ContainerStarted","Data":"54bc7c704b1b4e133a8e1806521b4431560c2ad2f82669b44524c18ffb5d443c"} Dec 06 06:19:27 crc kubenswrapper[4809]: I1206 06:19:27.494030 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-9b86998b5-fg9jq" Dec 06 06:19:27 crc kubenswrapper[4809]: I1206 06:19:27.516450 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8jvh2" event={"ID":"d4af32e7-96e5-4f86-b271-6b4efcace0dc","Type":"ContainerStarted","Data":"e31814af765d0bec9a5bad864050f9b02bb215972eb7f9ffc2d5fba01a2f13d8"} Dec 06 06:19:27 crc kubenswrapper[4809]: I1206 06:19:27.516491 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8jvh2" event={"ID":"d4af32e7-96e5-4f86-b271-6b4efcace0dc","Type":"ContainerStarted","Data":"553b7c1b21cbc2fabc21f34229e517151f16b7342e9d9b4e8fc713118ef833ed"} Dec 06 06:19:27 crc kubenswrapper[4809]: I1206 06:19:27.517783 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-9b86998b5-fg9jq" podStartSLOduration=4.5177462550000005 podStartE2EDuration="4.517746255s" podCreationTimestamp="2025-12-06 06:19:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:19:27.515449653 +0000 UTC m=+1692.404432615" watchObservedRunningTime="2025-12-06 06:19:27.517746255 +0000 UTC m=+1692.406729197" Dec 06 06:19:27 crc kubenswrapper[4809]: I1206 06:19:27.539236 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-8jvh2" podStartSLOduration=2.539213304 podStartE2EDuration="2.539213304s" podCreationTimestamp="2025-12-06 06:19:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:19:27.538727381 +0000 UTC m=+1692.427710323" watchObservedRunningTime="2025-12-06 06:19:27.539213304 +0000 UTC m=+1692.428196246" Dec 06 06:19:27 crc kubenswrapper[4809]: I1206 06:19:27.583985 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 06:19:27 crc kubenswrapper[4809]: I1206 06:19:27.618510 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 06:19:30 crc kubenswrapper[4809]: I1206 06:19:30.132884 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-xmngf"] Dec 06 06:19:30 crc kubenswrapper[4809]: I1206 06:19:30.135109 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-xmngf" Dec 06 06:19:30 crc kubenswrapper[4809]: I1206 06:19:30.143630 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 06 06:19:30 crc kubenswrapper[4809]: I1206 06:19:30.143858 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-f84tp" Dec 06 06:19:30 crc kubenswrapper[4809]: I1206 06:19:30.144028 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 06 06:19:30 crc kubenswrapper[4809]: I1206 06:19:30.144175 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 06 06:19:30 crc kubenswrapper[4809]: I1206 06:19:30.145756 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-xmngf"] Dec 06 06:19:30 crc kubenswrapper[4809]: I1206 06:19:30.267975 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68932bed-06eb-4c47-843c-a171ccbe23c3-combined-ca-bundle\") pod \"aodh-db-sync-xmngf\" (UID: \"68932bed-06eb-4c47-843c-a171ccbe23c3\") " pod="openstack/aodh-db-sync-xmngf" Dec 06 06:19:30 crc kubenswrapper[4809]: I1206 06:19:30.268381 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68932bed-06eb-4c47-843c-a171ccbe23c3-scripts\") pod \"aodh-db-sync-xmngf\" (UID: \"68932bed-06eb-4c47-843c-a171ccbe23c3\") " pod="openstack/aodh-db-sync-xmngf" Dec 06 06:19:30 crc kubenswrapper[4809]: I1206 06:19:30.268636 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kb8w8\" (UniqueName: \"kubernetes.io/projected/68932bed-06eb-4c47-843c-a171ccbe23c3-kube-api-access-kb8w8\") pod \"aodh-db-sync-xmngf\" (UID: \"68932bed-06eb-4c47-843c-a171ccbe23c3\") " pod="openstack/aodh-db-sync-xmngf" Dec 06 06:19:30 crc kubenswrapper[4809]: I1206 06:19:30.268670 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68932bed-06eb-4c47-843c-a171ccbe23c3-config-data\") pod \"aodh-db-sync-xmngf\" (UID: \"68932bed-06eb-4c47-843c-a171ccbe23c3\") " pod="openstack/aodh-db-sync-xmngf" Dec 06 06:19:30 crc kubenswrapper[4809]: I1206 06:19:30.370673 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68932bed-06eb-4c47-843c-a171ccbe23c3-combined-ca-bundle\") pod \"aodh-db-sync-xmngf\" (UID: \"68932bed-06eb-4c47-843c-a171ccbe23c3\") " pod="openstack/aodh-db-sync-xmngf" Dec 06 06:19:30 crc kubenswrapper[4809]: I1206 06:19:30.370877 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68932bed-06eb-4c47-843c-a171ccbe23c3-scripts\") pod \"aodh-db-sync-xmngf\" (UID: \"68932bed-06eb-4c47-843c-a171ccbe23c3\") " pod="openstack/aodh-db-sync-xmngf" Dec 06 06:19:30 crc kubenswrapper[4809]: I1206 06:19:30.371166 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kb8w8\" (UniqueName: \"kubernetes.io/projected/68932bed-06eb-4c47-843c-a171ccbe23c3-kube-api-access-kb8w8\") pod \"aodh-db-sync-xmngf\" (UID: \"68932bed-06eb-4c47-843c-a171ccbe23c3\") " pod="openstack/aodh-db-sync-xmngf" Dec 06 06:19:30 crc kubenswrapper[4809]: I1206 06:19:30.371231 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68932bed-06eb-4c47-843c-a171ccbe23c3-config-data\") pod \"aodh-db-sync-xmngf\" (UID: \"68932bed-06eb-4c47-843c-a171ccbe23c3\") " pod="openstack/aodh-db-sync-xmngf" Dec 06 06:19:30 crc kubenswrapper[4809]: I1206 06:19:30.374778 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68932bed-06eb-4c47-843c-a171ccbe23c3-scripts\") pod \"aodh-db-sync-xmngf\" (UID: \"68932bed-06eb-4c47-843c-a171ccbe23c3\") " pod="openstack/aodh-db-sync-xmngf" Dec 06 06:19:30 crc kubenswrapper[4809]: I1206 06:19:30.385918 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68932bed-06eb-4c47-843c-a171ccbe23c3-combined-ca-bundle\") pod \"aodh-db-sync-xmngf\" (UID: \"68932bed-06eb-4c47-843c-a171ccbe23c3\") " pod="openstack/aodh-db-sync-xmngf" Dec 06 06:19:30 crc kubenswrapper[4809]: I1206 06:19:30.386763 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68932bed-06eb-4c47-843c-a171ccbe23c3-config-data\") pod \"aodh-db-sync-xmngf\" (UID: \"68932bed-06eb-4c47-843c-a171ccbe23c3\") " pod="openstack/aodh-db-sync-xmngf" Dec 06 06:19:30 crc kubenswrapper[4809]: I1206 06:19:30.392506 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kb8w8\" (UniqueName: \"kubernetes.io/projected/68932bed-06eb-4c47-843c-a171ccbe23c3-kube-api-access-kb8w8\") pod \"aodh-db-sync-xmngf\" (UID: \"68932bed-06eb-4c47-843c-a171ccbe23c3\") " pod="openstack/aodh-db-sync-xmngf" Dec 06 06:19:30 crc kubenswrapper[4809]: I1206 06:19:30.496642 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-xmngf" Dec 06 06:19:30 crc kubenswrapper[4809]: I1206 06:19:30.568416 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"b6980646-cff8-45b6-a26b-53d6282f26a9","Type":"ContainerStarted","Data":"e19dfa7717885a4de9c30d0495f2a16b9bc5e74be5bd260ef8679bd31a822ec7"} Dec 06 06:19:30 crc kubenswrapper[4809]: I1206 06:19:30.568594 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="b6980646-cff8-45b6-a26b-53d6282f26a9" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://e19dfa7717885a4de9c30d0495f2a16b9bc5e74be5bd260ef8679bd31a822ec7" gracePeriod=30 Dec 06 06:19:30 crc kubenswrapper[4809]: I1206 06:19:30.574271 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1063337d-4069-4953-87af-504675fd2680","Type":"ContainerStarted","Data":"7ceec0ff566411a28456eb4f585c994762e59bf8f891b3f99703f04dead10b77"} Dec 06 06:19:30 crc kubenswrapper[4809]: I1206 06:19:30.575834 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2b09b87b-a70c-43d1-936d-0758596953aa","Type":"ContainerStarted","Data":"1aa255953561fd6511f3d4d017ca55dfb6b59f1450b964afb21e4629ac1f5eb0"} Dec 06 06:19:30 crc kubenswrapper[4809]: I1206 06:19:30.575864 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2b09b87b-a70c-43d1-936d-0758596953aa","Type":"ContainerStarted","Data":"e504c52d62558bba4a8d2c2eb04a40354fbc3c0b753fef04552ba9d053ba7a23"} Dec 06 06:19:30 crc kubenswrapper[4809]: I1206 06:19:30.576019 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2b09b87b-a70c-43d1-936d-0758596953aa" containerName="nova-metadata-log" containerID="cri-o://e504c52d62558bba4a8d2c2eb04a40354fbc3c0b753fef04552ba9d053ba7a23" gracePeriod=30 Dec 06 06:19:30 crc kubenswrapper[4809]: I1206 06:19:30.576272 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2b09b87b-a70c-43d1-936d-0758596953aa" containerName="nova-metadata-metadata" containerID="cri-o://1aa255953561fd6511f3d4d017ca55dfb6b59f1450b964afb21e4629ac1f5eb0" gracePeriod=30 Dec 06 06:19:30 crc kubenswrapper[4809]: I1206 06:19:30.584053 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d1eb5524-8750-4c35-8342-e93796d69d35","Type":"ContainerStarted","Data":"1065a6e6558600751f276837fa7df070df3ca6b43b3bf9d8cd8472342ecd50c8"} Dec 06 06:19:30 crc kubenswrapper[4809]: I1206 06:19:30.595363 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.276093456 podStartE2EDuration="7.595337667s" podCreationTimestamp="2025-12-06 06:19:23 +0000 UTC" firstStartedPulling="2025-12-06 06:19:25.347545804 +0000 UTC m=+1690.236528746" lastFinishedPulling="2025-12-06 06:19:29.666790015 +0000 UTC m=+1694.555772957" observedRunningTime="2025-12-06 06:19:30.591308928 +0000 UTC m=+1695.480291890" watchObservedRunningTime="2025-12-06 06:19:30.595337667 +0000 UTC m=+1695.484320609" Dec 06 06:19:30 crc kubenswrapper[4809]: I1206 06:19:30.631114 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.237963236 podStartE2EDuration="7.631035469s" podCreationTimestamp="2025-12-06 06:19:23 +0000 UTC" firstStartedPulling="2025-12-06 06:19:25.273345772 +0000 UTC m=+1690.162328714" lastFinishedPulling="2025-12-06 06:19:29.666418005 +0000 UTC m=+1694.555400947" observedRunningTime="2025-12-06 06:19:30.612091779 +0000 UTC m=+1695.501074721" watchObservedRunningTime="2025-12-06 06:19:30.631035469 +0000 UTC m=+1695.520018411" Dec 06 06:19:30 crc kubenswrapper[4809]: I1206 06:19:30.659440 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.267967027 podStartE2EDuration="7.659420046s" podCreationTimestamp="2025-12-06 06:19:23 +0000 UTC" firstStartedPulling="2025-12-06 06:19:25.272811878 +0000 UTC m=+1690.161794830" lastFinishedPulling="2025-12-06 06:19:29.664264907 +0000 UTC m=+1694.553247849" observedRunningTime="2025-12-06 06:19:30.6369495 +0000 UTC m=+1695.525932452" watchObservedRunningTime="2025-12-06 06:19:30.659420046 +0000 UTC m=+1695.548402988" Dec 06 06:19:31 crc kubenswrapper[4809]: I1206 06:19:31.077786 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-xmngf"] Dec 06 06:19:31 crc kubenswrapper[4809]: W1206 06:19:31.081271 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod68932bed_06eb_4c47_843c_a171ccbe23c3.slice/crio-999f689aa3d86c19e20a33f566a4db5dd186a8be03d6ae60c9975e167ae3ec31 WatchSource:0}: Error finding container 999f689aa3d86c19e20a33f566a4db5dd186a8be03d6ae60c9975e167ae3ec31: Status 404 returned error can't find the container with id 999f689aa3d86c19e20a33f566a4db5dd186a8be03d6ae60c9975e167ae3ec31 Dec 06 06:19:31 crc kubenswrapper[4809]: I1206 06:19:31.389318 4809 scope.go:117] "RemoveContainer" containerID="9bc2dc88b3b9f309ce0169aa510326dd77022dece4c69cab3141df00fe5c0106" Dec 06 06:19:31 crc kubenswrapper[4809]: E1206 06:19:31.389900 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:19:31 crc kubenswrapper[4809]: I1206 06:19:31.610705 4809 generic.go:334] "Generic (PLEG): container finished" podID="2b09b87b-a70c-43d1-936d-0758596953aa" containerID="e504c52d62558bba4a8d2c2eb04a40354fbc3c0b753fef04552ba9d053ba7a23" exitCode=143 Dec 06 06:19:31 crc kubenswrapper[4809]: I1206 06:19:31.610768 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2b09b87b-a70c-43d1-936d-0758596953aa","Type":"ContainerDied","Data":"e504c52d62558bba4a8d2c2eb04a40354fbc3c0b753fef04552ba9d053ba7a23"} Dec 06 06:19:31 crc kubenswrapper[4809]: I1206 06:19:31.612628 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-xmngf" event={"ID":"68932bed-06eb-4c47-843c-a171ccbe23c3","Type":"ContainerStarted","Data":"999f689aa3d86c19e20a33f566a4db5dd186a8be03d6ae60c9975e167ae3ec31"} Dec 06 06:19:31 crc kubenswrapper[4809]: I1206 06:19:31.615156 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1063337d-4069-4953-87af-504675fd2680","Type":"ContainerStarted","Data":"61cea5fc05e8fe6a1a0f811db70b555ea83d7f9182968979a307a5a971533ea6"} Dec 06 06:19:31 crc kubenswrapper[4809]: I1206 06:19:31.634683 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=4.242735085 podStartE2EDuration="8.634666077s" podCreationTimestamp="2025-12-06 06:19:23 +0000 UTC" firstStartedPulling="2025-12-06 06:19:25.273382113 +0000 UTC m=+1690.162365055" lastFinishedPulling="2025-12-06 06:19:29.665313105 +0000 UTC m=+1694.554296047" observedRunningTime="2025-12-06 06:19:31.630677839 +0000 UTC m=+1696.519660781" watchObservedRunningTime="2025-12-06 06:19:31.634666077 +0000 UTC m=+1696.523649019" Dec 06 06:19:33 crc kubenswrapper[4809]: I1206 06:19:33.544986 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 06:19:33 crc kubenswrapper[4809]: I1206 06:19:33.545338 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 06:19:33 crc kubenswrapper[4809]: I1206 06:19:33.565918 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 06 06:19:33 crc kubenswrapper[4809]: I1206 06:19:33.565974 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 06 06:19:33 crc kubenswrapper[4809]: I1206 06:19:33.623487 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 06 06:19:33 crc kubenswrapper[4809]: I1206 06:19:33.624236 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 06 06:19:33 crc kubenswrapper[4809]: I1206 06:19:33.655447 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 06 06:19:33 crc kubenswrapper[4809]: I1206 06:19:33.977579 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-9b86998b5-fg9jq" Dec 06 06:19:33 crc kubenswrapper[4809]: I1206 06:19:33.985443 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:19:34 crc kubenswrapper[4809]: I1206 06:19:34.057412 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-5drs7"] Dec 06 06:19:34 crc kubenswrapper[4809]: I1206 06:19:34.057636 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7756b9d78c-5drs7" podUID="0710c66a-2ed5-4bf6-8f12-e3754d122d08" containerName="dnsmasq-dns" containerID="cri-o://c7206cbf87c3ef1a5a62aa4aa4e23858e953abea02fc17523eee8ffb3a20dcd2" gracePeriod=10 Dec 06 06:19:34 crc kubenswrapper[4809]: I1206 06:19:34.631245 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="1063337d-4069-4953-87af-504675fd2680" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.236:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 06:19:34 crc kubenswrapper[4809]: I1206 06:19:34.632067 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="1063337d-4069-4953-87af-504675fd2680" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.236:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 06:19:34 crc kubenswrapper[4809]: I1206 06:19:34.648729 4809 generic.go:334] "Generic (PLEG): container finished" podID="0710c66a-2ed5-4bf6-8f12-e3754d122d08" containerID="c7206cbf87c3ef1a5a62aa4aa4e23858e953abea02fc17523eee8ffb3a20dcd2" exitCode=0 Dec 06 06:19:34 crc kubenswrapper[4809]: I1206 06:19:34.648812 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-5drs7" event={"ID":"0710c66a-2ed5-4bf6-8f12-e3754d122d08","Type":"ContainerDied","Data":"c7206cbf87c3ef1a5a62aa4aa4e23858e953abea02fc17523eee8ffb3a20dcd2"} Dec 06 06:19:34 crc kubenswrapper[4809]: I1206 06:19:34.684345 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 06 06:19:35 crc kubenswrapper[4809]: I1206 06:19:35.701377 4809 generic.go:334] "Generic (PLEG): container finished" podID="dad7df76-c6ae-4d0c-aef7-63ca9efb77a7" containerID="0c398009cd796b9757f33073f5ae46f28d2c1cb725130ab094f4dd9d64918527" exitCode=0 Dec 06 06:19:35 crc kubenswrapper[4809]: I1206 06:19:35.704151 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-x8zvs" event={"ID":"dad7df76-c6ae-4d0c-aef7-63ca9efb77a7","Type":"ContainerDied","Data":"0c398009cd796b9757f33073f5ae46f28d2c1cb725130ab094f4dd9d64918527"} Dec 06 06:19:37 crc kubenswrapper[4809]: I1206 06:19:37.236147 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7756b9d78c-5drs7" podUID="0710c66a-2ed5-4bf6-8f12-e3754d122d08" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.208:5353: connect: connection refused" Dec 06 06:19:38 crc kubenswrapper[4809]: I1206 06:19:38.583085 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-x8zvs" Dec 06 06:19:38 crc kubenswrapper[4809]: I1206 06:19:38.615477 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-5drs7" Dec 06 06:19:38 crc kubenswrapper[4809]: I1206 06:19:38.726678 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0710c66a-2ed5-4bf6-8f12-e3754d122d08-ovsdbserver-nb\") pod \"0710c66a-2ed5-4bf6-8f12-e3754d122d08\" (UID: \"0710c66a-2ed5-4bf6-8f12-e3754d122d08\") " Dec 06 06:19:38 crc kubenswrapper[4809]: I1206 06:19:38.726745 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dad7df76-c6ae-4d0c-aef7-63ca9efb77a7-combined-ca-bundle\") pod \"dad7df76-c6ae-4d0c-aef7-63ca9efb77a7\" (UID: \"dad7df76-c6ae-4d0c-aef7-63ca9efb77a7\") " Dec 06 06:19:38 crc kubenswrapper[4809]: I1206 06:19:38.726795 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dad7df76-c6ae-4d0c-aef7-63ca9efb77a7-config-data\") pod \"dad7df76-c6ae-4d0c-aef7-63ca9efb77a7\" (UID: \"dad7df76-c6ae-4d0c-aef7-63ca9efb77a7\") " Dec 06 06:19:38 crc kubenswrapper[4809]: I1206 06:19:38.726824 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0710c66a-2ed5-4bf6-8f12-e3754d122d08-ovsdbserver-sb\") pod \"0710c66a-2ed5-4bf6-8f12-e3754d122d08\" (UID: \"0710c66a-2ed5-4bf6-8f12-e3754d122d08\") " Dec 06 06:19:38 crc kubenswrapper[4809]: I1206 06:19:38.726912 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0710c66a-2ed5-4bf6-8f12-e3754d122d08-dns-swift-storage-0\") pod \"0710c66a-2ed5-4bf6-8f12-e3754d122d08\" (UID: \"0710c66a-2ed5-4bf6-8f12-e3754d122d08\") " Dec 06 06:19:38 crc kubenswrapper[4809]: I1206 06:19:38.727024 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0710c66a-2ed5-4bf6-8f12-e3754d122d08-config\") pod \"0710c66a-2ed5-4bf6-8f12-e3754d122d08\" (UID: \"0710c66a-2ed5-4bf6-8f12-e3754d122d08\") " Dec 06 06:19:38 crc kubenswrapper[4809]: I1206 06:19:38.727056 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5q64s\" (UniqueName: \"kubernetes.io/projected/dad7df76-c6ae-4d0c-aef7-63ca9efb77a7-kube-api-access-5q64s\") pod \"dad7df76-c6ae-4d0c-aef7-63ca9efb77a7\" (UID: \"dad7df76-c6ae-4d0c-aef7-63ca9efb77a7\") " Dec 06 06:19:38 crc kubenswrapper[4809]: I1206 06:19:38.727140 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dad7df76-c6ae-4d0c-aef7-63ca9efb77a7-scripts\") pod \"dad7df76-c6ae-4d0c-aef7-63ca9efb77a7\" (UID: \"dad7df76-c6ae-4d0c-aef7-63ca9efb77a7\") " Dec 06 06:19:38 crc kubenswrapper[4809]: I1206 06:19:38.727191 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4lsdv\" (UniqueName: \"kubernetes.io/projected/0710c66a-2ed5-4bf6-8f12-e3754d122d08-kube-api-access-4lsdv\") pod \"0710c66a-2ed5-4bf6-8f12-e3754d122d08\" (UID: \"0710c66a-2ed5-4bf6-8f12-e3754d122d08\") " Dec 06 06:19:38 crc kubenswrapper[4809]: I1206 06:19:38.727245 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0710c66a-2ed5-4bf6-8f12-e3754d122d08-dns-svc\") pod \"0710c66a-2ed5-4bf6-8f12-e3754d122d08\" (UID: \"0710c66a-2ed5-4bf6-8f12-e3754d122d08\") " Dec 06 06:19:38 crc kubenswrapper[4809]: I1206 06:19:38.732945 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dad7df76-c6ae-4d0c-aef7-63ca9efb77a7-scripts" (OuterVolumeSpecName: "scripts") pod "dad7df76-c6ae-4d0c-aef7-63ca9efb77a7" (UID: "dad7df76-c6ae-4d0c-aef7-63ca9efb77a7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:19:38 crc kubenswrapper[4809]: I1206 06:19:38.732920 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dad7df76-c6ae-4d0c-aef7-63ca9efb77a7-kube-api-access-5q64s" (OuterVolumeSpecName: "kube-api-access-5q64s") pod "dad7df76-c6ae-4d0c-aef7-63ca9efb77a7" (UID: "dad7df76-c6ae-4d0c-aef7-63ca9efb77a7"). InnerVolumeSpecName "kube-api-access-5q64s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:19:38 crc kubenswrapper[4809]: I1206 06:19:38.736516 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0710c66a-2ed5-4bf6-8f12-e3754d122d08-kube-api-access-4lsdv" (OuterVolumeSpecName: "kube-api-access-4lsdv") pod "0710c66a-2ed5-4bf6-8f12-e3754d122d08" (UID: "0710c66a-2ed5-4bf6-8f12-e3754d122d08"). InnerVolumeSpecName "kube-api-access-4lsdv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:19:38 crc kubenswrapper[4809]: I1206 06:19:38.736912 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-xmngf" event={"ID":"68932bed-06eb-4c47-843c-a171ccbe23c3","Type":"ContainerStarted","Data":"978ea797bd9b2c048c05d45d0e936251240a37ac0ab20db1bb2a1d2bb5c5493d"} Dec 06 06:19:38 crc kubenswrapper[4809]: I1206 06:19:38.738728 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-x8zvs" event={"ID":"dad7df76-c6ae-4d0c-aef7-63ca9efb77a7","Type":"ContainerDied","Data":"08844096fda39abf61f3e0470d8015cffc93a65cb1f4549dbde5b178cc3e1012"} Dec 06 06:19:38 crc kubenswrapper[4809]: I1206 06:19:38.738767 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="08844096fda39abf61f3e0470d8015cffc93a65cb1f4549dbde5b178cc3e1012" Dec 06 06:19:38 crc kubenswrapper[4809]: I1206 06:19:38.738820 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-x8zvs" Dec 06 06:19:38 crc kubenswrapper[4809]: I1206 06:19:38.741216 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-5drs7" event={"ID":"0710c66a-2ed5-4bf6-8f12-e3754d122d08","Type":"ContainerDied","Data":"ea9bfdc08ba590ca8519cf0f03d895fa18fc120fad32675a8d9fb1e12c6b72b1"} Dec 06 06:19:38 crc kubenswrapper[4809]: I1206 06:19:38.741264 4809 scope.go:117] "RemoveContainer" containerID="c7206cbf87c3ef1a5a62aa4aa4e23858e953abea02fc17523eee8ffb3a20dcd2" Dec 06 06:19:38 crc kubenswrapper[4809]: I1206 06:19:38.741308 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-5drs7" Dec 06 06:19:38 crc kubenswrapper[4809]: I1206 06:19:38.769576 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-xmngf" podStartSLOduration=1.431255369 podStartE2EDuration="8.769559132s" podCreationTimestamp="2025-12-06 06:19:30 +0000 UTC" firstStartedPulling="2025-12-06 06:19:31.085297526 +0000 UTC m=+1695.974280468" lastFinishedPulling="2025-12-06 06:19:38.423601289 +0000 UTC m=+1703.312584231" observedRunningTime="2025-12-06 06:19:38.756000906 +0000 UTC m=+1703.644983848" watchObservedRunningTime="2025-12-06 06:19:38.769559132 +0000 UTC m=+1703.658542074" Dec 06 06:19:38 crc kubenswrapper[4809]: I1206 06:19:38.783497 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dad7df76-c6ae-4d0c-aef7-63ca9efb77a7-config-data" (OuterVolumeSpecName: "config-data") pod "dad7df76-c6ae-4d0c-aef7-63ca9efb77a7" (UID: "dad7df76-c6ae-4d0c-aef7-63ca9efb77a7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:19:38 crc kubenswrapper[4809]: I1206 06:19:38.799077 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dad7df76-c6ae-4d0c-aef7-63ca9efb77a7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dad7df76-c6ae-4d0c-aef7-63ca9efb77a7" (UID: "dad7df76-c6ae-4d0c-aef7-63ca9efb77a7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:19:38 crc kubenswrapper[4809]: I1206 06:19:38.802256 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0710c66a-2ed5-4bf6-8f12-e3754d122d08-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "0710c66a-2ed5-4bf6-8f12-e3754d122d08" (UID: "0710c66a-2ed5-4bf6-8f12-e3754d122d08"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:19:38 crc kubenswrapper[4809]: I1206 06:19:38.804699 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0710c66a-2ed5-4bf6-8f12-e3754d122d08-config" (OuterVolumeSpecName: "config") pod "0710c66a-2ed5-4bf6-8f12-e3754d122d08" (UID: "0710c66a-2ed5-4bf6-8f12-e3754d122d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:19:38 crc kubenswrapper[4809]: I1206 06:19:38.814355 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0710c66a-2ed5-4bf6-8f12-e3754d122d08-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0710c66a-2ed5-4bf6-8f12-e3754d122d08" (UID: "0710c66a-2ed5-4bf6-8f12-e3754d122d08"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:19:38 crc kubenswrapper[4809]: I1206 06:19:38.827867 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0710c66a-2ed5-4bf6-8f12-e3754d122d08-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0710c66a-2ed5-4bf6-8f12-e3754d122d08" (UID: "0710c66a-2ed5-4bf6-8f12-e3754d122d08"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:19:38 crc kubenswrapper[4809]: I1206 06:19:38.830630 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0710c66a-2ed5-4bf6-8f12-e3754d122d08-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:38 crc kubenswrapper[4809]: I1206 06:19:38.830663 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dad7df76-c6ae-4d0c-aef7-63ca9efb77a7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:38 crc kubenswrapper[4809]: I1206 06:19:38.830673 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dad7df76-c6ae-4d0c-aef7-63ca9efb77a7-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:38 crc kubenswrapper[4809]: I1206 06:19:38.830682 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0710c66a-2ed5-4bf6-8f12-e3754d122d08-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:38 crc kubenswrapper[4809]: I1206 06:19:38.830691 4809 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0710c66a-2ed5-4bf6-8f12-e3754d122d08-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:38 crc kubenswrapper[4809]: I1206 06:19:38.830700 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0710c66a-2ed5-4bf6-8f12-e3754d122d08-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:38 crc kubenswrapper[4809]: I1206 06:19:38.830710 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5q64s\" (UniqueName: \"kubernetes.io/projected/dad7df76-c6ae-4d0c-aef7-63ca9efb77a7-kube-api-access-5q64s\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:38 crc kubenswrapper[4809]: I1206 06:19:38.830719 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dad7df76-c6ae-4d0c-aef7-63ca9efb77a7-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:38 crc kubenswrapper[4809]: I1206 06:19:38.830727 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4lsdv\" (UniqueName: \"kubernetes.io/projected/0710c66a-2ed5-4bf6-8f12-e3754d122d08-kube-api-access-4lsdv\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:38 crc kubenswrapper[4809]: I1206 06:19:38.842401 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0710c66a-2ed5-4bf6-8f12-e3754d122d08-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0710c66a-2ed5-4bf6-8f12-e3754d122d08" (UID: "0710c66a-2ed5-4bf6-8f12-e3754d122d08"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:19:38 crc kubenswrapper[4809]: I1206 06:19:38.943002 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0710c66a-2ed5-4bf6-8f12-e3754d122d08-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:38 crc kubenswrapper[4809]: I1206 06:19:38.952393 4809 scope.go:117] "RemoveContainer" containerID="c6e22fe4c2b6cc2192719a3b4daa8bcc285d59c1befbaf185eb3d743da6af958" Dec 06 06:19:39 crc kubenswrapper[4809]: I1206 06:19:39.089646 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-5drs7"] Dec 06 06:19:39 crc kubenswrapper[4809]: I1206 06:19:39.104942 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-5drs7"] Dec 06 06:19:39 crc kubenswrapper[4809]: I1206 06:19:39.420780 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0710c66a-2ed5-4bf6-8f12-e3754d122d08" path="/var/lib/kubelet/pods/0710c66a-2ed5-4bf6-8f12-e3754d122d08/volumes" Dec 06 06:19:39 crc kubenswrapper[4809]: I1206 06:19:39.753825 4809 generic.go:334] "Generic (PLEG): container finished" podID="d4af32e7-96e5-4f86-b271-6b4efcace0dc" containerID="e31814af765d0bec9a5bad864050f9b02bb215972eb7f9ffc2d5fba01a2f13d8" exitCode=0 Dec 06 06:19:39 crc kubenswrapper[4809]: I1206 06:19:39.753898 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8jvh2" event={"ID":"d4af32e7-96e5-4f86-b271-6b4efcace0dc","Type":"ContainerDied","Data":"e31814af765d0bec9a5bad864050f9b02bb215972eb7f9ffc2d5fba01a2f13d8"} Dec 06 06:19:39 crc kubenswrapper[4809]: I1206 06:19:39.805603 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 06 06:19:39 crc kubenswrapper[4809]: I1206 06:19:39.805921 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="1063337d-4069-4953-87af-504675fd2680" containerName="nova-api-log" containerID="cri-o://7ceec0ff566411a28456eb4f585c994762e59bf8f891b3f99703f04dead10b77" gracePeriod=30 Dec 06 06:19:39 crc kubenswrapper[4809]: I1206 06:19:39.806000 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="1063337d-4069-4953-87af-504675fd2680" containerName="nova-api-api" containerID="cri-o://61cea5fc05e8fe6a1a0f811db70b555ea83d7f9182968979a307a5a971533ea6" gracePeriod=30 Dec 06 06:19:39 crc kubenswrapper[4809]: I1206 06:19:39.825499 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 06:19:39 crc kubenswrapper[4809]: I1206 06:19:39.825745 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="d1eb5524-8750-4c35-8342-e93796d69d35" containerName="nova-scheduler-scheduler" containerID="cri-o://1065a6e6558600751f276837fa7df070df3ca6b43b3bf9d8cd8472342ecd50c8" gracePeriod=30 Dec 06 06:19:40 crc kubenswrapper[4809]: I1206 06:19:40.767793 4809 generic.go:334] "Generic (PLEG): container finished" podID="1063337d-4069-4953-87af-504675fd2680" containerID="7ceec0ff566411a28456eb4f585c994762e59bf8f891b3f99703f04dead10b77" exitCode=143 Dec 06 06:19:40 crc kubenswrapper[4809]: I1206 06:19:40.767861 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1063337d-4069-4953-87af-504675fd2680","Type":"ContainerDied","Data":"7ceec0ff566411a28456eb4f585c994762e59bf8f891b3f99703f04dead10b77"} Dec 06 06:19:41 crc kubenswrapper[4809]: I1206 06:19:41.255946 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8jvh2" Dec 06 06:19:41 crc kubenswrapper[4809]: I1206 06:19:41.401992 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4af32e7-96e5-4f86-b271-6b4efcace0dc-combined-ca-bundle\") pod \"d4af32e7-96e5-4f86-b271-6b4efcace0dc\" (UID: \"d4af32e7-96e5-4f86-b271-6b4efcace0dc\") " Dec 06 06:19:41 crc kubenswrapper[4809]: I1206 06:19:41.402374 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4af32e7-96e5-4f86-b271-6b4efcace0dc-config-data\") pod \"d4af32e7-96e5-4f86-b271-6b4efcace0dc\" (UID: \"d4af32e7-96e5-4f86-b271-6b4efcace0dc\") " Dec 06 06:19:41 crc kubenswrapper[4809]: I1206 06:19:41.402426 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbhtf\" (UniqueName: \"kubernetes.io/projected/d4af32e7-96e5-4f86-b271-6b4efcace0dc-kube-api-access-gbhtf\") pod \"d4af32e7-96e5-4f86-b271-6b4efcace0dc\" (UID: \"d4af32e7-96e5-4f86-b271-6b4efcace0dc\") " Dec 06 06:19:41 crc kubenswrapper[4809]: I1206 06:19:41.402534 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4af32e7-96e5-4f86-b271-6b4efcace0dc-scripts\") pod \"d4af32e7-96e5-4f86-b271-6b4efcace0dc\" (UID: \"d4af32e7-96e5-4f86-b271-6b4efcace0dc\") " Dec 06 06:19:41 crc kubenswrapper[4809]: I1206 06:19:41.408488 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4af32e7-96e5-4f86-b271-6b4efcace0dc-kube-api-access-gbhtf" (OuterVolumeSpecName: "kube-api-access-gbhtf") pod "d4af32e7-96e5-4f86-b271-6b4efcace0dc" (UID: "d4af32e7-96e5-4f86-b271-6b4efcace0dc"). InnerVolumeSpecName "kube-api-access-gbhtf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:19:41 crc kubenswrapper[4809]: I1206 06:19:41.413671 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4af32e7-96e5-4f86-b271-6b4efcace0dc-scripts" (OuterVolumeSpecName: "scripts") pod "d4af32e7-96e5-4f86-b271-6b4efcace0dc" (UID: "d4af32e7-96e5-4f86-b271-6b4efcace0dc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:19:41 crc kubenswrapper[4809]: I1206 06:19:41.437168 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4af32e7-96e5-4f86-b271-6b4efcace0dc-config-data" (OuterVolumeSpecName: "config-data") pod "d4af32e7-96e5-4f86-b271-6b4efcace0dc" (UID: "d4af32e7-96e5-4f86-b271-6b4efcace0dc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:19:41 crc kubenswrapper[4809]: I1206 06:19:41.438101 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4af32e7-96e5-4f86-b271-6b4efcace0dc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d4af32e7-96e5-4f86-b271-6b4efcace0dc" (UID: "d4af32e7-96e5-4f86-b271-6b4efcace0dc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:19:41 crc kubenswrapper[4809]: I1206 06:19:41.505665 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4af32e7-96e5-4f86-b271-6b4efcace0dc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:41 crc kubenswrapper[4809]: I1206 06:19:41.505713 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4af32e7-96e5-4f86-b271-6b4efcace0dc-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:41 crc kubenswrapper[4809]: I1206 06:19:41.505726 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbhtf\" (UniqueName: \"kubernetes.io/projected/d4af32e7-96e5-4f86-b271-6b4efcace0dc-kube-api-access-gbhtf\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:41 crc kubenswrapper[4809]: I1206 06:19:41.505737 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4af32e7-96e5-4f86-b271-6b4efcace0dc-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:41 crc kubenswrapper[4809]: I1206 06:19:41.780101 4809 generic.go:334] "Generic (PLEG): container finished" podID="68932bed-06eb-4c47-843c-a171ccbe23c3" containerID="978ea797bd9b2c048c05d45d0e936251240a37ac0ab20db1bb2a1d2bb5c5493d" exitCode=0 Dec 06 06:19:41 crc kubenswrapper[4809]: I1206 06:19:41.780227 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-xmngf" event={"ID":"68932bed-06eb-4c47-843c-a171ccbe23c3","Type":"ContainerDied","Data":"978ea797bd9b2c048c05d45d0e936251240a37ac0ab20db1bb2a1d2bb5c5493d"} Dec 06 06:19:41 crc kubenswrapper[4809]: I1206 06:19:41.782828 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8jvh2" event={"ID":"d4af32e7-96e5-4f86-b271-6b4efcace0dc","Type":"ContainerDied","Data":"553b7c1b21cbc2fabc21f34229e517151f16b7342e9d9b4e8fc713118ef833ed"} Dec 06 06:19:41 crc kubenswrapper[4809]: I1206 06:19:41.782872 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="553b7c1b21cbc2fabc21f34229e517151f16b7342e9d9b4e8fc713118ef833ed" Dec 06 06:19:41 crc kubenswrapper[4809]: I1206 06:19:41.782887 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8jvh2" Dec 06 06:19:41 crc kubenswrapper[4809]: I1206 06:19:41.864175 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 06 06:19:41 crc kubenswrapper[4809]: E1206 06:19:41.864807 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dad7df76-c6ae-4d0c-aef7-63ca9efb77a7" containerName="nova-manage" Dec 06 06:19:41 crc kubenswrapper[4809]: I1206 06:19:41.864835 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="dad7df76-c6ae-4d0c-aef7-63ca9efb77a7" containerName="nova-manage" Dec 06 06:19:41 crc kubenswrapper[4809]: E1206 06:19:41.864857 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0710c66a-2ed5-4bf6-8f12-e3754d122d08" containerName="init" Dec 06 06:19:41 crc kubenswrapper[4809]: I1206 06:19:41.864866 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="0710c66a-2ed5-4bf6-8f12-e3754d122d08" containerName="init" Dec 06 06:19:41 crc kubenswrapper[4809]: E1206 06:19:41.864883 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4af32e7-96e5-4f86-b271-6b4efcace0dc" containerName="nova-cell1-conductor-db-sync" Dec 06 06:19:41 crc kubenswrapper[4809]: I1206 06:19:41.864895 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4af32e7-96e5-4f86-b271-6b4efcace0dc" containerName="nova-cell1-conductor-db-sync" Dec 06 06:19:41 crc kubenswrapper[4809]: E1206 06:19:41.864920 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0710c66a-2ed5-4bf6-8f12-e3754d122d08" containerName="dnsmasq-dns" Dec 06 06:19:41 crc kubenswrapper[4809]: I1206 06:19:41.864927 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="0710c66a-2ed5-4bf6-8f12-e3754d122d08" containerName="dnsmasq-dns" Dec 06 06:19:41 crc kubenswrapper[4809]: I1206 06:19:41.865233 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="dad7df76-c6ae-4d0c-aef7-63ca9efb77a7" containerName="nova-manage" Dec 06 06:19:41 crc kubenswrapper[4809]: I1206 06:19:41.865277 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="0710c66a-2ed5-4bf6-8f12-e3754d122d08" containerName="dnsmasq-dns" Dec 06 06:19:41 crc kubenswrapper[4809]: I1206 06:19:41.865303 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4af32e7-96e5-4f86-b271-6b4efcace0dc" containerName="nova-cell1-conductor-db-sync" Dec 06 06:19:41 crc kubenswrapper[4809]: I1206 06:19:41.866729 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 06 06:19:41 crc kubenswrapper[4809]: I1206 06:19:41.869252 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 06 06:19:41 crc kubenswrapper[4809]: I1206 06:19:41.885329 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 06 06:19:42 crc kubenswrapper[4809]: I1206 06:19:42.017170 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bszgm\" (UniqueName: \"kubernetes.io/projected/ae53b50b-d35c-4035-ab90-0c0575a95200-kube-api-access-bszgm\") pod \"nova-cell1-conductor-0\" (UID: \"ae53b50b-d35c-4035-ab90-0c0575a95200\") " pod="openstack/nova-cell1-conductor-0" Dec 06 06:19:42 crc kubenswrapper[4809]: I1206 06:19:42.017264 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae53b50b-d35c-4035-ab90-0c0575a95200-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ae53b50b-d35c-4035-ab90-0c0575a95200\") " pod="openstack/nova-cell1-conductor-0" Dec 06 06:19:42 crc kubenswrapper[4809]: I1206 06:19:42.017548 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae53b50b-d35c-4035-ab90-0c0575a95200-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ae53b50b-d35c-4035-ab90-0c0575a95200\") " pod="openstack/nova-cell1-conductor-0" Dec 06 06:19:42 crc kubenswrapper[4809]: I1206 06:19:42.119249 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae53b50b-d35c-4035-ab90-0c0575a95200-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ae53b50b-d35c-4035-ab90-0c0575a95200\") " pod="openstack/nova-cell1-conductor-0" Dec 06 06:19:42 crc kubenswrapper[4809]: I1206 06:19:42.119318 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bszgm\" (UniqueName: \"kubernetes.io/projected/ae53b50b-d35c-4035-ab90-0c0575a95200-kube-api-access-bszgm\") pod \"nova-cell1-conductor-0\" (UID: \"ae53b50b-d35c-4035-ab90-0c0575a95200\") " pod="openstack/nova-cell1-conductor-0" Dec 06 06:19:42 crc kubenswrapper[4809]: I1206 06:19:42.119374 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae53b50b-d35c-4035-ab90-0c0575a95200-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ae53b50b-d35c-4035-ab90-0c0575a95200\") " pod="openstack/nova-cell1-conductor-0" Dec 06 06:19:42 crc kubenswrapper[4809]: I1206 06:19:42.123722 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae53b50b-d35c-4035-ab90-0c0575a95200-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ae53b50b-d35c-4035-ab90-0c0575a95200\") " pod="openstack/nova-cell1-conductor-0" Dec 06 06:19:42 crc kubenswrapper[4809]: I1206 06:19:42.123819 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae53b50b-d35c-4035-ab90-0c0575a95200-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ae53b50b-d35c-4035-ab90-0c0575a95200\") " pod="openstack/nova-cell1-conductor-0" Dec 06 06:19:42 crc kubenswrapper[4809]: I1206 06:19:42.153385 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bszgm\" (UniqueName: \"kubernetes.io/projected/ae53b50b-d35c-4035-ab90-0c0575a95200-kube-api-access-bszgm\") pod \"nova-cell1-conductor-0\" (UID: \"ae53b50b-d35c-4035-ab90-0c0575a95200\") " pod="openstack/nova-cell1-conductor-0" Dec 06 06:19:42 crc kubenswrapper[4809]: I1206 06:19:42.194017 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 06 06:19:42 crc kubenswrapper[4809]: I1206 06:19:42.392605 4809 scope.go:117] "RemoveContainer" containerID="9bc2dc88b3b9f309ce0169aa510326dd77022dece4c69cab3141df00fe5c0106" Dec 06 06:19:42 crc kubenswrapper[4809]: E1206 06:19:42.393108 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:19:42 crc kubenswrapper[4809]: W1206 06:19:42.554357 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae53b50b_d35c_4035_ab90_0c0575a95200.slice/crio-741615e17feafbc8253b2d9512fc408e3ee69e4006b3af3612c149686187d0cf WatchSource:0}: Error finding container 741615e17feafbc8253b2d9512fc408e3ee69e4006b3af3612c149686187d0cf: Status 404 returned error can't find the container with id 741615e17feafbc8253b2d9512fc408e3ee69e4006b3af3612c149686187d0cf Dec 06 06:19:42 crc kubenswrapper[4809]: I1206 06:19:42.565505 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 06 06:19:42 crc kubenswrapper[4809]: I1206 06:19:42.795783 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ae53b50b-d35c-4035-ab90-0c0575a95200","Type":"ContainerStarted","Data":"741615e17feafbc8253b2d9512fc408e3ee69e4006b3af3612c149686187d0cf"} Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.138080 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-xmngf" Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.295348 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68932bed-06eb-4c47-843c-a171ccbe23c3-scripts\") pod \"68932bed-06eb-4c47-843c-a171ccbe23c3\" (UID: \"68932bed-06eb-4c47-843c-a171ccbe23c3\") " Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.295668 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kb8w8\" (UniqueName: \"kubernetes.io/projected/68932bed-06eb-4c47-843c-a171ccbe23c3-kube-api-access-kb8w8\") pod \"68932bed-06eb-4c47-843c-a171ccbe23c3\" (UID: \"68932bed-06eb-4c47-843c-a171ccbe23c3\") " Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.295731 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68932bed-06eb-4c47-843c-a171ccbe23c3-combined-ca-bundle\") pod \"68932bed-06eb-4c47-843c-a171ccbe23c3\" (UID: \"68932bed-06eb-4c47-843c-a171ccbe23c3\") " Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.295770 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68932bed-06eb-4c47-843c-a171ccbe23c3-config-data\") pod \"68932bed-06eb-4c47-843c-a171ccbe23c3\" (UID: \"68932bed-06eb-4c47-843c-a171ccbe23c3\") " Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.301466 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68932bed-06eb-4c47-843c-a171ccbe23c3-scripts" (OuterVolumeSpecName: "scripts") pod "68932bed-06eb-4c47-843c-a171ccbe23c3" (UID: "68932bed-06eb-4c47-843c-a171ccbe23c3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.306351 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68932bed-06eb-4c47-843c-a171ccbe23c3-kube-api-access-kb8w8" (OuterVolumeSpecName: "kube-api-access-kb8w8") pod "68932bed-06eb-4c47-843c-a171ccbe23c3" (UID: "68932bed-06eb-4c47-843c-a171ccbe23c3"). InnerVolumeSpecName "kube-api-access-kb8w8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.348061 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68932bed-06eb-4c47-843c-a171ccbe23c3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "68932bed-06eb-4c47-843c-a171ccbe23c3" (UID: "68932bed-06eb-4c47-843c-a171ccbe23c3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.348215 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68932bed-06eb-4c47-843c-a171ccbe23c3-config-data" (OuterVolumeSpecName: "config-data") pod "68932bed-06eb-4c47-843c-a171ccbe23c3" (UID: "68932bed-06eb-4c47-843c-a171ccbe23c3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.398112 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68932bed-06eb-4c47-843c-a171ccbe23c3-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.398141 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kb8w8\" (UniqueName: \"kubernetes.io/projected/68932bed-06eb-4c47-843c-a171ccbe23c3-kube-api-access-kb8w8\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.398154 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68932bed-06eb-4c47-843c-a171ccbe23c3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.398163 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68932bed-06eb-4c47-843c-a171ccbe23c3-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.549148 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.703865 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8vcbp\" (UniqueName: \"kubernetes.io/projected/d1eb5524-8750-4c35-8342-e93796d69d35-kube-api-access-8vcbp\") pod \"d1eb5524-8750-4c35-8342-e93796d69d35\" (UID: \"d1eb5524-8750-4c35-8342-e93796d69d35\") " Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.704069 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1eb5524-8750-4c35-8342-e93796d69d35-config-data\") pod \"d1eb5524-8750-4c35-8342-e93796d69d35\" (UID: \"d1eb5524-8750-4c35-8342-e93796d69d35\") " Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.704098 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1eb5524-8750-4c35-8342-e93796d69d35-combined-ca-bundle\") pod \"d1eb5524-8750-4c35-8342-e93796d69d35\" (UID: \"d1eb5524-8750-4c35-8342-e93796d69d35\") " Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.713284 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1eb5524-8750-4c35-8342-e93796d69d35-kube-api-access-8vcbp" (OuterVolumeSpecName: "kube-api-access-8vcbp") pod "d1eb5524-8750-4c35-8342-e93796d69d35" (UID: "d1eb5524-8750-4c35-8342-e93796d69d35"). InnerVolumeSpecName "kube-api-access-8vcbp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.744668 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1eb5524-8750-4c35-8342-e93796d69d35-config-data" (OuterVolumeSpecName: "config-data") pod "d1eb5524-8750-4c35-8342-e93796d69d35" (UID: "d1eb5524-8750-4c35-8342-e93796d69d35"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.760995 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1eb5524-8750-4c35-8342-e93796d69d35-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d1eb5524-8750-4c35-8342-e93796d69d35" (UID: "d1eb5524-8750-4c35-8342-e93796d69d35"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.806838 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8vcbp\" (UniqueName: \"kubernetes.io/projected/d1eb5524-8750-4c35-8342-e93796d69d35-kube-api-access-8vcbp\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.806876 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1eb5524-8750-4c35-8342-e93796d69d35-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.806889 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1eb5524-8750-4c35-8342-e93796d69d35-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.814318 4809 generic.go:334] "Generic (PLEG): container finished" podID="1063337d-4069-4953-87af-504675fd2680" containerID="61cea5fc05e8fe6a1a0f811db70b555ea83d7f9182968979a307a5a971533ea6" exitCode=0 Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.814604 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1063337d-4069-4953-87af-504675fd2680","Type":"ContainerDied","Data":"61cea5fc05e8fe6a1a0f811db70b555ea83d7f9182968979a307a5a971533ea6"} Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.818263 4809 generic.go:334] "Generic (PLEG): container finished" podID="d1eb5524-8750-4c35-8342-e93796d69d35" containerID="1065a6e6558600751f276837fa7df070df3ca6b43b3bf9d8cd8472342ecd50c8" exitCode=0 Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.819208 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.819219 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d1eb5524-8750-4c35-8342-e93796d69d35","Type":"ContainerDied","Data":"1065a6e6558600751f276837fa7df070df3ca6b43b3bf9d8cd8472342ecd50c8"} Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.819307 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d1eb5524-8750-4c35-8342-e93796d69d35","Type":"ContainerDied","Data":"f41a50de85789e4ec57df7dd785e71ba62570ead8965f9259bf661263028b52a"} Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.819333 4809 scope.go:117] "RemoveContainer" containerID="1065a6e6558600751f276837fa7df070df3ca6b43b3bf9d8cd8472342ecd50c8" Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.831770 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ae53b50b-d35c-4035-ab90-0c0575a95200","Type":"ContainerStarted","Data":"bd82f8515b01c93d292eb97c4a9bd71b6d3da61a88c4534ec6976e9d272b30f9"} Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.841598 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.843572 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-xmngf" event={"ID":"68932bed-06eb-4c47-843c-a171ccbe23c3","Type":"ContainerDied","Data":"999f689aa3d86c19e20a33f566a4db5dd186a8be03d6ae60c9975e167ae3ec31"} Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.843603 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="999f689aa3d86c19e20a33f566a4db5dd186a8be03d6ae60c9975e167ae3ec31" Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.843657 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-xmngf" Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.868976 4809 scope.go:117] "RemoveContainer" containerID="1065a6e6558600751f276837fa7df070df3ca6b43b3bf9d8cd8472342ecd50c8" Dec 06 06:19:43 crc kubenswrapper[4809]: E1206 06:19:43.870403 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1065a6e6558600751f276837fa7df070df3ca6b43b3bf9d8cd8472342ecd50c8\": container with ID starting with 1065a6e6558600751f276837fa7df070df3ca6b43b3bf9d8cd8472342ecd50c8 not found: ID does not exist" containerID="1065a6e6558600751f276837fa7df070df3ca6b43b3bf9d8cd8472342ecd50c8" Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.870444 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1065a6e6558600751f276837fa7df070df3ca6b43b3bf9d8cd8472342ecd50c8"} err="failed to get container status \"1065a6e6558600751f276837fa7df070df3ca6b43b3bf9d8cd8472342ecd50c8\": rpc error: code = NotFound desc = could not find container \"1065a6e6558600751f276837fa7df070df3ca6b43b3bf9d8cd8472342ecd50c8\": container with ID starting with 1065a6e6558600751f276837fa7df070df3ca6b43b3bf9d8cd8472342ecd50c8 not found: ID does not exist" Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.870817 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.87079088 podStartE2EDuration="2.87079088s" podCreationTimestamp="2025-12-06 06:19:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:19:43.856981118 +0000 UTC m=+1708.745964060" watchObservedRunningTime="2025-12-06 06:19:43.87079088 +0000 UTC m=+1708.759773822" Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.902479 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.907838 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1063337d-4069-4953-87af-504675fd2680-combined-ca-bundle\") pod \"1063337d-4069-4953-87af-504675fd2680\" (UID: \"1063337d-4069-4953-87af-504675fd2680\") " Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.908046 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gdlxd\" (UniqueName: \"kubernetes.io/projected/1063337d-4069-4953-87af-504675fd2680-kube-api-access-gdlxd\") pod \"1063337d-4069-4953-87af-504675fd2680\" (UID: \"1063337d-4069-4953-87af-504675fd2680\") " Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.908276 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1063337d-4069-4953-87af-504675fd2680-config-data\") pod \"1063337d-4069-4953-87af-504675fd2680\" (UID: \"1063337d-4069-4953-87af-504675fd2680\") " Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.908315 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1063337d-4069-4953-87af-504675fd2680-logs\") pod \"1063337d-4069-4953-87af-504675fd2680\" (UID: \"1063337d-4069-4953-87af-504675fd2680\") " Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.909078 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1063337d-4069-4953-87af-504675fd2680-logs" (OuterVolumeSpecName: "logs") pod "1063337d-4069-4953-87af-504675fd2680" (UID: "1063337d-4069-4953-87af-504675fd2680"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.918118 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1063337d-4069-4953-87af-504675fd2680-kube-api-access-gdlxd" (OuterVolumeSpecName: "kube-api-access-gdlxd") pod "1063337d-4069-4953-87af-504675fd2680" (UID: "1063337d-4069-4953-87af-504675fd2680"). InnerVolumeSpecName "kube-api-access-gdlxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.919424 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.934231 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 06:19:43 crc kubenswrapper[4809]: E1206 06:19:43.934831 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1063337d-4069-4953-87af-504675fd2680" containerName="nova-api-api" Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.934856 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1063337d-4069-4953-87af-504675fd2680" containerName="nova-api-api" Dec 06 06:19:43 crc kubenswrapper[4809]: E1206 06:19:43.934916 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68932bed-06eb-4c47-843c-a171ccbe23c3" containerName="aodh-db-sync" Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.934927 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="68932bed-06eb-4c47-843c-a171ccbe23c3" containerName="aodh-db-sync" Dec 06 06:19:43 crc kubenswrapper[4809]: E1206 06:19:43.934970 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1eb5524-8750-4c35-8342-e93796d69d35" containerName="nova-scheduler-scheduler" Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.934981 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1eb5524-8750-4c35-8342-e93796d69d35" containerName="nova-scheduler-scheduler" Dec 06 06:19:43 crc kubenswrapper[4809]: E1206 06:19:43.935002 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1063337d-4069-4953-87af-504675fd2680" containerName="nova-api-log" Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.935010 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1063337d-4069-4953-87af-504675fd2680" containerName="nova-api-log" Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.935308 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1eb5524-8750-4c35-8342-e93796d69d35" containerName="nova-scheduler-scheduler" Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.935338 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="1063337d-4069-4953-87af-504675fd2680" containerName="nova-api-log" Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.935370 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="1063337d-4069-4953-87af-504675fd2680" containerName="nova-api-api" Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.935387 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="68932bed-06eb-4c47-843c-a171ccbe23c3" containerName="aodh-db-sync" Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.937784 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.944976 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.957126 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.963698 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1063337d-4069-4953-87af-504675fd2680-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1063337d-4069-4953-87af-504675fd2680" (UID: "1063337d-4069-4953-87af-504675fd2680"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:19:43 crc kubenswrapper[4809]: I1206 06:19:43.985792 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1063337d-4069-4953-87af-504675fd2680-config-data" (OuterVolumeSpecName: "config-data") pod "1063337d-4069-4953-87af-504675fd2680" (UID: "1063337d-4069-4953-87af-504675fd2680"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:19:44 crc kubenswrapper[4809]: I1206 06:19:44.011186 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65d1f54f-e7cd-467f-a385-e081568dc6f6-config-data\") pod \"nova-scheduler-0\" (UID: \"65d1f54f-e7cd-467f-a385-e081568dc6f6\") " pod="openstack/nova-scheduler-0" Dec 06 06:19:44 crc kubenswrapper[4809]: I1206 06:19:44.011243 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65d1f54f-e7cd-467f-a385-e081568dc6f6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"65d1f54f-e7cd-467f-a385-e081568dc6f6\") " pod="openstack/nova-scheduler-0" Dec 06 06:19:44 crc kubenswrapper[4809]: I1206 06:19:44.011268 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jk7qq\" (UniqueName: \"kubernetes.io/projected/65d1f54f-e7cd-467f-a385-e081568dc6f6-kube-api-access-jk7qq\") pod \"nova-scheduler-0\" (UID: \"65d1f54f-e7cd-467f-a385-e081568dc6f6\") " pod="openstack/nova-scheduler-0" Dec 06 06:19:44 crc kubenswrapper[4809]: I1206 06:19:44.011818 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gdlxd\" (UniqueName: \"kubernetes.io/projected/1063337d-4069-4953-87af-504675fd2680-kube-api-access-gdlxd\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:44 crc kubenswrapper[4809]: I1206 06:19:44.011855 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1063337d-4069-4953-87af-504675fd2680-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:44 crc kubenswrapper[4809]: I1206 06:19:44.011870 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1063337d-4069-4953-87af-504675fd2680-logs\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:44 crc kubenswrapper[4809]: I1206 06:19:44.011883 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1063337d-4069-4953-87af-504675fd2680-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:44 crc kubenswrapper[4809]: I1206 06:19:44.113707 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65d1f54f-e7cd-467f-a385-e081568dc6f6-config-data\") pod \"nova-scheduler-0\" (UID: \"65d1f54f-e7cd-467f-a385-e081568dc6f6\") " pod="openstack/nova-scheduler-0" Dec 06 06:19:44 crc kubenswrapper[4809]: I1206 06:19:44.113761 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65d1f54f-e7cd-467f-a385-e081568dc6f6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"65d1f54f-e7cd-467f-a385-e081568dc6f6\") " pod="openstack/nova-scheduler-0" Dec 06 06:19:44 crc kubenswrapper[4809]: I1206 06:19:44.113782 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jk7qq\" (UniqueName: \"kubernetes.io/projected/65d1f54f-e7cd-467f-a385-e081568dc6f6-kube-api-access-jk7qq\") pod \"nova-scheduler-0\" (UID: \"65d1f54f-e7cd-467f-a385-e081568dc6f6\") " pod="openstack/nova-scheduler-0" Dec 06 06:19:44 crc kubenswrapper[4809]: I1206 06:19:44.118342 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65d1f54f-e7cd-467f-a385-e081568dc6f6-config-data\") pod \"nova-scheduler-0\" (UID: \"65d1f54f-e7cd-467f-a385-e081568dc6f6\") " pod="openstack/nova-scheduler-0" Dec 06 06:19:44 crc kubenswrapper[4809]: I1206 06:19:44.119198 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65d1f54f-e7cd-467f-a385-e081568dc6f6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"65d1f54f-e7cd-467f-a385-e081568dc6f6\") " pod="openstack/nova-scheduler-0" Dec 06 06:19:44 crc kubenswrapper[4809]: I1206 06:19:44.137076 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jk7qq\" (UniqueName: \"kubernetes.io/projected/65d1f54f-e7cd-467f-a385-e081568dc6f6-kube-api-access-jk7qq\") pod \"nova-scheduler-0\" (UID: \"65d1f54f-e7cd-467f-a385-e081568dc6f6\") " pod="openstack/nova-scheduler-0" Dec 06 06:19:44 crc kubenswrapper[4809]: I1206 06:19:44.264961 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 06:19:44 crc kubenswrapper[4809]: I1206 06:19:44.774615 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 06:19:44 crc kubenswrapper[4809]: I1206 06:19:44.857018 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 06:19:44 crc kubenswrapper[4809]: I1206 06:19:44.857518 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1063337d-4069-4953-87af-504675fd2680","Type":"ContainerDied","Data":"5ba0e0184a3c76b27ac0fd21ddb91c91db328ddb1530ed2251e69c4a5c21a378"} Dec 06 06:19:44 crc kubenswrapper[4809]: I1206 06:19:44.858457 4809 scope.go:117] "RemoveContainer" containerID="61cea5fc05e8fe6a1a0f811db70b555ea83d7f9182968979a307a5a971533ea6" Dec 06 06:19:44 crc kubenswrapper[4809]: I1206 06:19:44.860808 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"65d1f54f-e7cd-467f-a385-e081568dc6f6","Type":"ContainerStarted","Data":"0f1fff03deeaa44564bdcaba4ee99ac1995319be786f9060dfd15b96289e4b21"} Dec 06 06:19:44 crc kubenswrapper[4809]: I1206 06:19:44.865414 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 06 06:19:44 crc kubenswrapper[4809]: I1206 06:19:44.941891 4809 scope.go:117] "RemoveContainer" containerID="7ceec0ff566411a28456eb4f585c994762e59bf8f891b3f99703f04dead10b77" Dec 06 06:19:44 crc kubenswrapper[4809]: I1206 06:19:44.942047 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 06 06:19:44 crc kubenswrapper[4809]: I1206 06:19:44.952277 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 06 06:19:44 crc kubenswrapper[4809]: I1206 06:19:44.955836 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-f84tp" Dec 06 06:19:44 crc kubenswrapper[4809]: I1206 06:19:44.956006 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 06 06:19:44 crc kubenswrapper[4809]: I1206 06:19:44.956637 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 06 06:19:44 crc kubenswrapper[4809]: I1206 06:19:44.958339 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 06 06:19:44 crc kubenswrapper[4809]: I1206 06:19:44.979665 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 06 06:19:44 crc kubenswrapper[4809]: I1206 06:19:44.993967 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 06 06:19:45 crc kubenswrapper[4809]: I1206 06:19:45.042639 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/413494e9-0401-4a6c-b85e-8e6be9997d20-combined-ca-bundle\") pod \"aodh-0\" (UID: \"413494e9-0401-4a6c-b85e-8e6be9997d20\") " pod="openstack/aodh-0" Dec 06 06:19:45 crc kubenswrapper[4809]: I1206 06:19:45.042682 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/413494e9-0401-4a6c-b85e-8e6be9997d20-scripts\") pod \"aodh-0\" (UID: \"413494e9-0401-4a6c-b85e-8e6be9997d20\") " pod="openstack/aodh-0" Dec 06 06:19:45 crc kubenswrapper[4809]: I1206 06:19:45.042862 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ss5v6\" (UniqueName: \"kubernetes.io/projected/413494e9-0401-4a6c-b85e-8e6be9997d20-kube-api-access-ss5v6\") pod \"aodh-0\" (UID: \"413494e9-0401-4a6c-b85e-8e6be9997d20\") " pod="openstack/aodh-0" Dec 06 06:19:45 crc kubenswrapper[4809]: I1206 06:19:45.044276 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/413494e9-0401-4a6c-b85e-8e6be9997d20-config-data\") pod \"aodh-0\" (UID: \"413494e9-0401-4a6c-b85e-8e6be9997d20\") " pod="openstack/aodh-0" Dec 06 06:19:45 crc kubenswrapper[4809]: I1206 06:19:45.047551 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 06 06:19:45 crc kubenswrapper[4809]: I1206 06:19:45.063782 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 06:19:45 crc kubenswrapper[4809]: I1206 06:19:45.065718 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 06 06:19:45 crc kubenswrapper[4809]: I1206 06:19:45.075910 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 06:19:45 crc kubenswrapper[4809]: I1206 06:19:45.146695 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdzsr\" (UniqueName: \"kubernetes.io/projected/215a53e4-4135-45b5-912c-e25d50e5c2d3-kube-api-access-sdzsr\") pod \"nova-api-0\" (UID: \"215a53e4-4135-45b5-912c-e25d50e5c2d3\") " pod="openstack/nova-api-0" Dec 06 06:19:45 crc kubenswrapper[4809]: I1206 06:19:45.146748 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/215a53e4-4135-45b5-912c-e25d50e5c2d3-config-data\") pod \"nova-api-0\" (UID: \"215a53e4-4135-45b5-912c-e25d50e5c2d3\") " pod="openstack/nova-api-0" Dec 06 06:19:45 crc kubenswrapper[4809]: I1206 06:19:45.146784 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ss5v6\" (UniqueName: \"kubernetes.io/projected/413494e9-0401-4a6c-b85e-8e6be9997d20-kube-api-access-ss5v6\") pod \"aodh-0\" (UID: \"413494e9-0401-4a6c-b85e-8e6be9997d20\") " pod="openstack/aodh-0" Dec 06 06:19:45 crc kubenswrapper[4809]: I1206 06:19:45.146805 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/413494e9-0401-4a6c-b85e-8e6be9997d20-config-data\") pod \"aodh-0\" (UID: \"413494e9-0401-4a6c-b85e-8e6be9997d20\") " pod="openstack/aodh-0" Dec 06 06:19:45 crc kubenswrapper[4809]: I1206 06:19:45.146854 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/215a53e4-4135-45b5-912c-e25d50e5c2d3-logs\") pod \"nova-api-0\" (UID: \"215a53e4-4135-45b5-912c-e25d50e5c2d3\") " pod="openstack/nova-api-0" Dec 06 06:19:45 crc kubenswrapper[4809]: I1206 06:19:45.146945 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/215a53e4-4135-45b5-912c-e25d50e5c2d3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"215a53e4-4135-45b5-912c-e25d50e5c2d3\") " pod="openstack/nova-api-0" Dec 06 06:19:45 crc kubenswrapper[4809]: I1206 06:19:45.147026 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/413494e9-0401-4a6c-b85e-8e6be9997d20-combined-ca-bundle\") pod \"aodh-0\" (UID: \"413494e9-0401-4a6c-b85e-8e6be9997d20\") " pod="openstack/aodh-0" Dec 06 06:19:45 crc kubenswrapper[4809]: I1206 06:19:45.147048 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/413494e9-0401-4a6c-b85e-8e6be9997d20-scripts\") pod \"aodh-0\" (UID: \"413494e9-0401-4a6c-b85e-8e6be9997d20\") " pod="openstack/aodh-0" Dec 06 06:19:45 crc kubenswrapper[4809]: I1206 06:19:45.151863 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/413494e9-0401-4a6c-b85e-8e6be9997d20-scripts\") pod \"aodh-0\" (UID: \"413494e9-0401-4a6c-b85e-8e6be9997d20\") " pod="openstack/aodh-0" Dec 06 06:19:45 crc kubenswrapper[4809]: I1206 06:19:45.152500 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/413494e9-0401-4a6c-b85e-8e6be9997d20-combined-ca-bundle\") pod \"aodh-0\" (UID: \"413494e9-0401-4a6c-b85e-8e6be9997d20\") " pod="openstack/aodh-0" Dec 06 06:19:45 crc kubenswrapper[4809]: I1206 06:19:45.152907 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/413494e9-0401-4a6c-b85e-8e6be9997d20-config-data\") pod \"aodh-0\" (UID: \"413494e9-0401-4a6c-b85e-8e6be9997d20\") " pod="openstack/aodh-0" Dec 06 06:19:45 crc kubenswrapper[4809]: I1206 06:19:45.169757 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ss5v6\" (UniqueName: \"kubernetes.io/projected/413494e9-0401-4a6c-b85e-8e6be9997d20-kube-api-access-ss5v6\") pod \"aodh-0\" (UID: \"413494e9-0401-4a6c-b85e-8e6be9997d20\") " pod="openstack/aodh-0" Dec 06 06:19:45 crc kubenswrapper[4809]: I1206 06:19:45.249254 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/215a53e4-4135-45b5-912c-e25d50e5c2d3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"215a53e4-4135-45b5-912c-e25d50e5c2d3\") " pod="openstack/nova-api-0" Dec 06 06:19:45 crc kubenswrapper[4809]: I1206 06:19:45.249608 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdzsr\" (UniqueName: \"kubernetes.io/projected/215a53e4-4135-45b5-912c-e25d50e5c2d3-kube-api-access-sdzsr\") pod \"nova-api-0\" (UID: \"215a53e4-4135-45b5-912c-e25d50e5c2d3\") " pod="openstack/nova-api-0" Dec 06 06:19:45 crc kubenswrapper[4809]: I1206 06:19:45.249718 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/215a53e4-4135-45b5-912c-e25d50e5c2d3-config-data\") pod \"nova-api-0\" (UID: \"215a53e4-4135-45b5-912c-e25d50e5c2d3\") " pod="openstack/nova-api-0" Dec 06 06:19:45 crc kubenswrapper[4809]: I1206 06:19:45.249826 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/215a53e4-4135-45b5-912c-e25d50e5c2d3-logs\") pod \"nova-api-0\" (UID: \"215a53e4-4135-45b5-912c-e25d50e5c2d3\") " pod="openstack/nova-api-0" Dec 06 06:19:45 crc kubenswrapper[4809]: I1206 06:19:45.250271 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/215a53e4-4135-45b5-912c-e25d50e5c2d3-logs\") pod \"nova-api-0\" (UID: \"215a53e4-4135-45b5-912c-e25d50e5c2d3\") " pod="openstack/nova-api-0" Dec 06 06:19:45 crc kubenswrapper[4809]: I1206 06:19:45.255499 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/215a53e4-4135-45b5-912c-e25d50e5c2d3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"215a53e4-4135-45b5-912c-e25d50e5c2d3\") " pod="openstack/nova-api-0" Dec 06 06:19:45 crc kubenswrapper[4809]: I1206 06:19:45.269809 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/215a53e4-4135-45b5-912c-e25d50e5c2d3-config-data\") pod \"nova-api-0\" (UID: \"215a53e4-4135-45b5-912c-e25d50e5c2d3\") " pod="openstack/nova-api-0" Dec 06 06:19:45 crc kubenswrapper[4809]: I1206 06:19:45.284619 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdzsr\" (UniqueName: \"kubernetes.io/projected/215a53e4-4135-45b5-912c-e25d50e5c2d3-kube-api-access-sdzsr\") pod \"nova-api-0\" (UID: \"215a53e4-4135-45b5-912c-e25d50e5c2d3\") " pod="openstack/nova-api-0" Dec 06 06:19:45 crc kubenswrapper[4809]: I1206 06:19:45.299890 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 06 06:19:45 crc kubenswrapper[4809]: I1206 06:19:45.405393 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 06:19:45 crc kubenswrapper[4809]: I1206 06:19:45.425629 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1063337d-4069-4953-87af-504675fd2680" path="/var/lib/kubelet/pods/1063337d-4069-4953-87af-504675fd2680/volumes" Dec 06 06:19:45 crc kubenswrapper[4809]: I1206 06:19:45.426415 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1eb5524-8750-4c35-8342-e93796d69d35" path="/var/lib/kubelet/pods/d1eb5524-8750-4c35-8342-e93796d69d35/volumes" Dec 06 06:19:45 crc kubenswrapper[4809]: I1206 06:19:45.898488 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"65d1f54f-e7cd-467f-a385-e081568dc6f6","Type":"ContainerStarted","Data":"651a13e2264e20f136248c2269f20b93f1368a8b7e29b9bd41af0c3c1ed835b8"} Dec 06 06:19:45 crc kubenswrapper[4809]: I1206 06:19:45.954858 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.954839497 podStartE2EDuration="2.954839497s" podCreationTimestamp="2025-12-06 06:19:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:19:45.932428352 +0000 UTC m=+1710.821411324" watchObservedRunningTime="2025-12-06 06:19:45.954839497 +0000 UTC m=+1710.843822429" Dec 06 06:19:46 crc kubenswrapper[4809]: I1206 06:19:46.001656 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 06 06:19:46 crc kubenswrapper[4809]: W1206 06:19:46.003486 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod413494e9_0401_4a6c_b85e_8e6be9997d20.slice/crio-f352c0ef812ecfcd4d21ad8a64c31e41640ca38385fb27da3ea01c9cfb3c88d0 WatchSource:0}: Error finding container f352c0ef812ecfcd4d21ad8a64c31e41640ca38385fb27da3ea01c9cfb3c88d0: Status 404 returned error can't find the container with id f352c0ef812ecfcd4d21ad8a64c31e41640ca38385fb27da3ea01c9cfb3c88d0 Dec 06 06:19:46 crc kubenswrapper[4809]: I1206 06:19:46.136781 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 06:19:46 crc kubenswrapper[4809]: I1206 06:19:46.921575 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"413494e9-0401-4a6c-b85e-8e6be9997d20","Type":"ContainerStarted","Data":"7e1832c98b7bbb803ca4af4d248b206551c35844e83b9caceeeff49ade9e0215"} Dec 06 06:19:46 crc kubenswrapper[4809]: I1206 06:19:46.922240 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"413494e9-0401-4a6c-b85e-8e6be9997d20","Type":"ContainerStarted","Data":"f352c0ef812ecfcd4d21ad8a64c31e41640ca38385fb27da3ea01c9cfb3c88d0"} Dec 06 06:19:46 crc kubenswrapper[4809]: I1206 06:19:46.924967 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"215a53e4-4135-45b5-912c-e25d50e5c2d3","Type":"ContainerStarted","Data":"87401fb28e1de011b3a112da08d30919cd13ce434f4e76261389e19eb091b1b8"} Dec 06 06:19:46 crc kubenswrapper[4809]: I1206 06:19:46.925042 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"215a53e4-4135-45b5-912c-e25d50e5c2d3","Type":"ContainerStarted","Data":"38942159c5dfe1eeb521ee0e2e56ac8d200f1be759cd9896b948b276d2094f2e"} Dec 06 06:19:46 crc kubenswrapper[4809]: I1206 06:19:46.925054 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"215a53e4-4135-45b5-912c-e25d50e5c2d3","Type":"ContainerStarted","Data":"4233a597f87bf500704b3c0e4974db08db0b180d93dd0e63c40543e7ecd4dbaf"} Dec 06 06:19:46 crc kubenswrapper[4809]: I1206 06:19:46.957637 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.957614381 podStartE2EDuration="2.957614381s" podCreationTimestamp="2025-12-06 06:19:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:19:46.947287243 +0000 UTC m=+1711.836270195" watchObservedRunningTime="2025-12-06 06:19:46.957614381 +0000 UTC m=+1711.846597323" Dec 06 06:19:47 crc kubenswrapper[4809]: I1206 06:19:47.231302 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 06 06:19:48 crc kubenswrapper[4809]: I1206 06:19:48.257821 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:19:48 crc kubenswrapper[4809]: I1206 06:19:48.259153 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a9170461-4a1f-43da-be11-12310d9b2574" containerName="ceilometer-central-agent" containerID="cri-o://e697b371434aa54fc302370afd097c1a6a1ab6c62b417790b09e40130cd2672e" gracePeriod=30 Dec 06 06:19:48 crc kubenswrapper[4809]: I1206 06:19:48.259734 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a9170461-4a1f-43da-be11-12310d9b2574" containerName="ceilometer-notification-agent" containerID="cri-o://db6f189fb52dcee256ce499fe10fafdebc05ca8f3f7a49453b7e2a39cf2ca71c" gracePeriod=30 Dec 06 06:19:48 crc kubenswrapper[4809]: I1206 06:19:48.259827 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a9170461-4a1f-43da-be11-12310d9b2574" containerName="sg-core" containerID="cri-o://c1b0d9118ab641766dd069635e54219a8078aba3d89dd6a49c1a4633398b7b99" gracePeriod=30 Dec 06 06:19:48 crc kubenswrapper[4809]: I1206 06:19:48.260164 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a9170461-4a1f-43da-be11-12310d9b2574" containerName="proxy-httpd" containerID="cri-o://d094101e80ac7e9b6b8cfe518288078b17e1d05fcf9784669c5e643c32b4b06e" gracePeriod=30 Dec 06 06:19:48 crc kubenswrapper[4809]: I1206 06:19:48.267284 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 06 06:19:48 crc kubenswrapper[4809]: I1206 06:19:48.379689 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="a9170461-4a1f-43da-be11-12310d9b2574" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.232:3000/\": dial tcp 10.217.0.232:3000: connect: connection refused" Dec 06 06:19:48 crc kubenswrapper[4809]: I1206 06:19:48.780500 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 06 06:19:48 crc kubenswrapper[4809]: I1206 06:19:48.951977 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"413494e9-0401-4a6c-b85e-8e6be9997d20","Type":"ContainerStarted","Data":"49a05cb8e9cda351bdd9cd29857e7bf2579eab07f1408ebc2efd1ce510d592c6"} Dec 06 06:19:48 crc kubenswrapper[4809]: I1206 06:19:48.956674 4809 generic.go:334] "Generic (PLEG): container finished" podID="a9170461-4a1f-43da-be11-12310d9b2574" containerID="d094101e80ac7e9b6b8cfe518288078b17e1d05fcf9784669c5e643c32b4b06e" exitCode=0 Dec 06 06:19:48 crc kubenswrapper[4809]: I1206 06:19:48.956713 4809 generic.go:334] "Generic (PLEG): container finished" podID="a9170461-4a1f-43da-be11-12310d9b2574" containerID="c1b0d9118ab641766dd069635e54219a8078aba3d89dd6a49c1a4633398b7b99" exitCode=2 Dec 06 06:19:48 crc kubenswrapper[4809]: I1206 06:19:48.956726 4809 generic.go:334] "Generic (PLEG): container finished" podID="a9170461-4a1f-43da-be11-12310d9b2574" containerID="e697b371434aa54fc302370afd097c1a6a1ab6c62b417790b09e40130cd2672e" exitCode=0 Dec 06 06:19:48 crc kubenswrapper[4809]: I1206 06:19:48.956748 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9170461-4a1f-43da-be11-12310d9b2574","Type":"ContainerDied","Data":"d094101e80ac7e9b6b8cfe518288078b17e1d05fcf9784669c5e643c32b4b06e"} Dec 06 06:19:48 crc kubenswrapper[4809]: I1206 06:19:48.956780 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9170461-4a1f-43da-be11-12310d9b2574","Type":"ContainerDied","Data":"c1b0d9118ab641766dd069635e54219a8078aba3d89dd6a49c1a4633398b7b99"} Dec 06 06:19:48 crc kubenswrapper[4809]: I1206 06:19:48.956794 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9170461-4a1f-43da-be11-12310d9b2574","Type":"ContainerDied","Data":"e697b371434aa54fc302370afd097c1a6a1ab6c62b417790b09e40130cd2672e"} Dec 06 06:19:49 crc kubenswrapper[4809]: I1206 06:19:49.265500 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 06 06:19:50 crc kubenswrapper[4809]: I1206 06:19:50.984090 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"413494e9-0401-4a6c-b85e-8e6be9997d20","Type":"ContainerStarted","Data":"c7a165ffefc33081d973a91c041184fc11186d1219545d5ae8e3ca366b218b2a"} Dec 06 06:19:52 crc kubenswrapper[4809]: I1206 06:19:51.998562 4809 generic.go:334] "Generic (PLEG): container finished" podID="a9170461-4a1f-43da-be11-12310d9b2574" containerID="db6f189fb52dcee256ce499fe10fafdebc05ca8f3f7a49453b7e2a39cf2ca71c" exitCode=0 Dec 06 06:19:52 crc kubenswrapper[4809]: I1206 06:19:51.998626 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9170461-4a1f-43da-be11-12310d9b2574","Type":"ContainerDied","Data":"db6f189fb52dcee256ce499fe10fafdebc05ca8f3f7a49453b7e2a39cf2ca71c"} Dec 06 06:19:52 crc kubenswrapper[4809]: I1206 06:19:52.437350 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:19:52 crc kubenswrapper[4809]: I1206 06:19:52.521843 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a9170461-4a1f-43da-be11-12310d9b2574-sg-core-conf-yaml\") pod \"a9170461-4a1f-43da-be11-12310d9b2574\" (UID: \"a9170461-4a1f-43da-be11-12310d9b2574\") " Dec 06 06:19:52 crc kubenswrapper[4809]: I1206 06:19:52.522191 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9170461-4a1f-43da-be11-12310d9b2574-run-httpd\") pod \"a9170461-4a1f-43da-be11-12310d9b2574\" (UID: \"a9170461-4a1f-43da-be11-12310d9b2574\") " Dec 06 06:19:52 crc kubenswrapper[4809]: I1206 06:19:52.522220 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hjhg\" (UniqueName: \"kubernetes.io/projected/a9170461-4a1f-43da-be11-12310d9b2574-kube-api-access-4hjhg\") pod \"a9170461-4a1f-43da-be11-12310d9b2574\" (UID: \"a9170461-4a1f-43da-be11-12310d9b2574\") " Dec 06 06:19:52 crc kubenswrapper[4809]: I1206 06:19:52.522252 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9170461-4a1f-43da-be11-12310d9b2574-log-httpd\") pod \"a9170461-4a1f-43da-be11-12310d9b2574\" (UID: \"a9170461-4a1f-43da-be11-12310d9b2574\") " Dec 06 06:19:52 crc kubenswrapper[4809]: I1206 06:19:52.522354 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9170461-4a1f-43da-be11-12310d9b2574-config-data\") pod \"a9170461-4a1f-43da-be11-12310d9b2574\" (UID: \"a9170461-4a1f-43da-be11-12310d9b2574\") " Dec 06 06:19:52 crc kubenswrapper[4809]: I1206 06:19:52.522396 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9170461-4a1f-43da-be11-12310d9b2574-combined-ca-bundle\") pod \"a9170461-4a1f-43da-be11-12310d9b2574\" (UID: \"a9170461-4a1f-43da-be11-12310d9b2574\") " Dec 06 06:19:52 crc kubenswrapper[4809]: I1206 06:19:52.522460 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9170461-4a1f-43da-be11-12310d9b2574-scripts\") pod \"a9170461-4a1f-43da-be11-12310d9b2574\" (UID: \"a9170461-4a1f-43da-be11-12310d9b2574\") " Dec 06 06:19:52 crc kubenswrapper[4809]: I1206 06:19:52.522693 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9170461-4a1f-43da-be11-12310d9b2574-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a9170461-4a1f-43da-be11-12310d9b2574" (UID: "a9170461-4a1f-43da-be11-12310d9b2574"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:19:52 crc kubenswrapper[4809]: I1206 06:19:52.522897 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9170461-4a1f-43da-be11-12310d9b2574-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a9170461-4a1f-43da-be11-12310d9b2574" (UID: "a9170461-4a1f-43da-be11-12310d9b2574"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:19:52 crc kubenswrapper[4809]: I1206 06:19:52.523238 4809 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9170461-4a1f-43da-be11-12310d9b2574-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:52 crc kubenswrapper[4809]: I1206 06:19:52.523252 4809 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9170461-4a1f-43da-be11-12310d9b2574-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:52 crc kubenswrapper[4809]: I1206 06:19:52.532131 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9170461-4a1f-43da-be11-12310d9b2574-kube-api-access-4hjhg" (OuterVolumeSpecName: "kube-api-access-4hjhg") pod "a9170461-4a1f-43da-be11-12310d9b2574" (UID: "a9170461-4a1f-43da-be11-12310d9b2574"). InnerVolumeSpecName "kube-api-access-4hjhg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:19:52 crc kubenswrapper[4809]: I1206 06:19:52.543233 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9170461-4a1f-43da-be11-12310d9b2574-scripts" (OuterVolumeSpecName: "scripts") pod "a9170461-4a1f-43da-be11-12310d9b2574" (UID: "a9170461-4a1f-43da-be11-12310d9b2574"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:19:52 crc kubenswrapper[4809]: I1206 06:19:52.623418 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9170461-4a1f-43da-be11-12310d9b2574-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a9170461-4a1f-43da-be11-12310d9b2574" (UID: "a9170461-4a1f-43da-be11-12310d9b2574"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:19:52 crc kubenswrapper[4809]: I1206 06:19:52.625255 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9170461-4a1f-43da-be11-12310d9b2574-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:52 crc kubenswrapper[4809]: I1206 06:19:52.625281 4809 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a9170461-4a1f-43da-be11-12310d9b2574-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:52 crc kubenswrapper[4809]: I1206 06:19:52.625292 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hjhg\" (UniqueName: \"kubernetes.io/projected/a9170461-4a1f-43da-be11-12310d9b2574-kube-api-access-4hjhg\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:52 crc kubenswrapper[4809]: I1206 06:19:52.706853 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9170461-4a1f-43da-be11-12310d9b2574-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a9170461-4a1f-43da-be11-12310d9b2574" (UID: "a9170461-4a1f-43da-be11-12310d9b2574"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:19:52 crc kubenswrapper[4809]: I1206 06:19:52.713632 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9170461-4a1f-43da-be11-12310d9b2574-config-data" (OuterVolumeSpecName: "config-data") pod "a9170461-4a1f-43da-be11-12310d9b2574" (UID: "a9170461-4a1f-43da-be11-12310d9b2574"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:19:52 crc kubenswrapper[4809]: I1206 06:19:52.727672 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9170461-4a1f-43da-be11-12310d9b2574-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:52 crc kubenswrapper[4809]: I1206 06:19:52.727713 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9170461-4a1f-43da-be11-12310d9b2574-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.016669 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"413494e9-0401-4a6c-b85e-8e6be9997d20","Type":"ContainerStarted","Data":"2e551eb383880763e22b099bbea7d3b3355007a1cc0163bc7551e4a9006d34a3"} Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.016964 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="413494e9-0401-4a6c-b85e-8e6be9997d20" containerName="aodh-api" containerID="cri-o://7e1832c98b7bbb803ca4af4d248b206551c35844e83b9caceeeff49ade9e0215" gracePeriod=30 Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.018077 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="413494e9-0401-4a6c-b85e-8e6be9997d20" containerName="aodh-listener" containerID="cri-o://2e551eb383880763e22b099bbea7d3b3355007a1cc0163bc7551e4a9006d34a3" gracePeriod=30 Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.018165 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="413494e9-0401-4a6c-b85e-8e6be9997d20" containerName="aodh-notifier" containerID="cri-o://c7a165ffefc33081d973a91c041184fc11186d1219545d5ae8e3ca366b218b2a" gracePeriod=30 Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.018260 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="413494e9-0401-4a6c-b85e-8e6be9997d20" containerName="aodh-evaluator" containerID="cri-o://49a05cb8e9cda351bdd9cd29857e7bf2579eab07f1408ebc2efd1ce510d592c6" gracePeriod=30 Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.022373 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9170461-4a1f-43da-be11-12310d9b2574","Type":"ContainerDied","Data":"64a4628b5a00fd07a1219a7d59cc7304f0a9c0db0073c3acdca7e0141f01e5b5"} Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.022438 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.022442 4809 scope.go:117] "RemoveContainer" containerID="d094101e80ac7e9b6b8cfe518288078b17e1d05fcf9784669c5e643c32b4b06e" Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.038004 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.818550119 podStartE2EDuration="9.037975836s" podCreationTimestamp="2025-12-06 06:19:44 +0000 UTC" firstStartedPulling="2025-12-06 06:19:46.014666871 +0000 UTC m=+1710.903649813" lastFinishedPulling="2025-12-06 06:19:52.234092588 +0000 UTC m=+1717.123075530" observedRunningTime="2025-12-06 06:19:53.035178591 +0000 UTC m=+1717.924161593" watchObservedRunningTime="2025-12-06 06:19:53.037975836 +0000 UTC m=+1717.926958778" Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.055304 4809 scope.go:117] "RemoveContainer" containerID="c1b0d9118ab641766dd069635e54219a8078aba3d89dd6a49c1a4633398b7b99" Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.092509 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.113568 4809 scope.go:117] "RemoveContainer" containerID="db6f189fb52dcee256ce499fe10fafdebc05ca8f3f7a49453b7e2a39cf2ca71c" Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.116900 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.144570 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:19:53 crc kubenswrapper[4809]: E1206 06:19:53.153197 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9170461-4a1f-43da-be11-12310d9b2574" containerName="proxy-httpd" Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.153239 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9170461-4a1f-43da-be11-12310d9b2574" containerName="proxy-httpd" Dec 06 06:19:53 crc kubenswrapper[4809]: E1206 06:19:53.153255 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9170461-4a1f-43da-be11-12310d9b2574" containerName="ceilometer-notification-agent" Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.153262 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9170461-4a1f-43da-be11-12310d9b2574" containerName="ceilometer-notification-agent" Dec 06 06:19:53 crc kubenswrapper[4809]: E1206 06:19:53.153360 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9170461-4a1f-43da-be11-12310d9b2574" containerName="sg-core" Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.153367 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9170461-4a1f-43da-be11-12310d9b2574" containerName="sg-core" Dec 06 06:19:53 crc kubenswrapper[4809]: E1206 06:19:53.153391 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9170461-4a1f-43da-be11-12310d9b2574" containerName="ceilometer-central-agent" Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.153398 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9170461-4a1f-43da-be11-12310d9b2574" containerName="ceilometer-central-agent" Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.153831 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9170461-4a1f-43da-be11-12310d9b2574" containerName="ceilometer-notification-agent" Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.153849 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9170461-4a1f-43da-be11-12310d9b2574" containerName="sg-core" Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.153862 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9170461-4a1f-43da-be11-12310d9b2574" containerName="proxy-httpd" Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.153880 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9170461-4a1f-43da-be11-12310d9b2574" containerName="ceilometer-central-agent" Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.158216 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.162751 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.164108 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.164288 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.206523 4809 scope.go:117] "RemoveContainer" containerID="e697b371434aa54fc302370afd097c1a6a1ab6c62b417790b09e40130cd2672e" Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.245757 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9xfw\" (UniqueName: \"kubernetes.io/projected/fd3f000e-9b7c-4f78-aadb-3a580c1441ff-kube-api-access-v9xfw\") pod \"ceilometer-0\" (UID: \"fd3f000e-9b7c-4f78-aadb-3a580c1441ff\") " pod="openstack/ceilometer-0" Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.245827 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fd3f000e-9b7c-4f78-aadb-3a580c1441ff-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fd3f000e-9b7c-4f78-aadb-3a580c1441ff\") " pod="openstack/ceilometer-0" Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.245895 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd3f000e-9b7c-4f78-aadb-3a580c1441ff-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fd3f000e-9b7c-4f78-aadb-3a580c1441ff\") " pod="openstack/ceilometer-0" Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.245950 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd3f000e-9b7c-4f78-aadb-3a580c1441ff-scripts\") pod \"ceilometer-0\" (UID: \"fd3f000e-9b7c-4f78-aadb-3a580c1441ff\") " pod="openstack/ceilometer-0" Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.246109 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd3f000e-9b7c-4f78-aadb-3a580c1441ff-log-httpd\") pod \"ceilometer-0\" (UID: \"fd3f000e-9b7c-4f78-aadb-3a580c1441ff\") " pod="openstack/ceilometer-0" Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.246176 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd3f000e-9b7c-4f78-aadb-3a580c1441ff-run-httpd\") pod \"ceilometer-0\" (UID: \"fd3f000e-9b7c-4f78-aadb-3a580c1441ff\") " pod="openstack/ceilometer-0" Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.246262 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd3f000e-9b7c-4f78-aadb-3a580c1441ff-config-data\") pod \"ceilometer-0\" (UID: \"fd3f000e-9b7c-4f78-aadb-3a580c1441ff\") " pod="openstack/ceilometer-0" Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.350502 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd3f000e-9b7c-4f78-aadb-3a580c1441ff-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fd3f000e-9b7c-4f78-aadb-3a580c1441ff\") " pod="openstack/ceilometer-0" Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.350917 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd3f000e-9b7c-4f78-aadb-3a580c1441ff-scripts\") pod \"ceilometer-0\" (UID: \"fd3f000e-9b7c-4f78-aadb-3a580c1441ff\") " pod="openstack/ceilometer-0" Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.351711 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd3f000e-9b7c-4f78-aadb-3a580c1441ff-log-httpd\") pod \"ceilometer-0\" (UID: \"fd3f000e-9b7c-4f78-aadb-3a580c1441ff\") " pod="openstack/ceilometer-0" Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.351812 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd3f000e-9b7c-4f78-aadb-3a580c1441ff-run-httpd\") pod \"ceilometer-0\" (UID: \"fd3f000e-9b7c-4f78-aadb-3a580c1441ff\") " pod="openstack/ceilometer-0" Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.351970 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd3f000e-9b7c-4f78-aadb-3a580c1441ff-config-data\") pod \"ceilometer-0\" (UID: \"fd3f000e-9b7c-4f78-aadb-3a580c1441ff\") " pod="openstack/ceilometer-0" Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.352031 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9xfw\" (UniqueName: \"kubernetes.io/projected/fd3f000e-9b7c-4f78-aadb-3a580c1441ff-kube-api-access-v9xfw\") pod \"ceilometer-0\" (UID: \"fd3f000e-9b7c-4f78-aadb-3a580c1441ff\") " pod="openstack/ceilometer-0" Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.352078 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fd3f000e-9b7c-4f78-aadb-3a580c1441ff-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fd3f000e-9b7c-4f78-aadb-3a580c1441ff\") " pod="openstack/ceilometer-0" Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.352225 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd3f000e-9b7c-4f78-aadb-3a580c1441ff-log-httpd\") pod \"ceilometer-0\" (UID: \"fd3f000e-9b7c-4f78-aadb-3a580c1441ff\") " pod="openstack/ceilometer-0" Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.352966 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd3f000e-9b7c-4f78-aadb-3a580c1441ff-run-httpd\") pod \"ceilometer-0\" (UID: \"fd3f000e-9b7c-4f78-aadb-3a580c1441ff\") " pod="openstack/ceilometer-0" Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.364729 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd3f000e-9b7c-4f78-aadb-3a580c1441ff-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fd3f000e-9b7c-4f78-aadb-3a580c1441ff\") " pod="openstack/ceilometer-0" Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.371985 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fd3f000e-9b7c-4f78-aadb-3a580c1441ff-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fd3f000e-9b7c-4f78-aadb-3a580c1441ff\") " pod="openstack/ceilometer-0" Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.374588 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd3f000e-9b7c-4f78-aadb-3a580c1441ff-scripts\") pod \"ceilometer-0\" (UID: \"fd3f000e-9b7c-4f78-aadb-3a580c1441ff\") " pod="openstack/ceilometer-0" Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.386083 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd3f000e-9b7c-4f78-aadb-3a580c1441ff-config-data\") pod \"ceilometer-0\" (UID: \"fd3f000e-9b7c-4f78-aadb-3a580c1441ff\") " pod="openstack/ceilometer-0" Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.403212 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9xfw\" (UniqueName: \"kubernetes.io/projected/fd3f000e-9b7c-4f78-aadb-3a580c1441ff-kube-api-access-v9xfw\") pod \"ceilometer-0\" (UID: \"fd3f000e-9b7c-4f78-aadb-3a580c1441ff\") " pod="openstack/ceilometer-0" Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.414618 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9170461-4a1f-43da-be11-12310d9b2574" path="/var/lib/kubelet/pods/a9170461-4a1f-43da-be11-12310d9b2574/volumes" Dec 06 06:19:53 crc kubenswrapper[4809]: I1206 06:19:53.559523 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:19:54 crc kubenswrapper[4809]: I1206 06:19:54.037210 4809 generic.go:334] "Generic (PLEG): container finished" podID="413494e9-0401-4a6c-b85e-8e6be9997d20" containerID="c7a165ffefc33081d973a91c041184fc11186d1219545d5ae8e3ca366b218b2a" exitCode=0 Dec 06 06:19:54 crc kubenswrapper[4809]: I1206 06:19:54.037475 4809 generic.go:334] "Generic (PLEG): container finished" podID="413494e9-0401-4a6c-b85e-8e6be9997d20" containerID="49a05cb8e9cda351bdd9cd29857e7bf2579eab07f1408ebc2efd1ce510d592c6" exitCode=0 Dec 06 06:19:54 crc kubenswrapper[4809]: I1206 06:19:54.037486 4809 generic.go:334] "Generic (PLEG): container finished" podID="413494e9-0401-4a6c-b85e-8e6be9997d20" containerID="7e1832c98b7bbb803ca4af4d248b206551c35844e83b9caceeeff49ade9e0215" exitCode=0 Dec 06 06:19:54 crc kubenswrapper[4809]: I1206 06:19:54.037285 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"413494e9-0401-4a6c-b85e-8e6be9997d20","Type":"ContainerDied","Data":"c7a165ffefc33081d973a91c041184fc11186d1219545d5ae8e3ca366b218b2a"} Dec 06 06:19:54 crc kubenswrapper[4809]: I1206 06:19:54.037514 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"413494e9-0401-4a6c-b85e-8e6be9997d20","Type":"ContainerDied","Data":"49a05cb8e9cda351bdd9cd29857e7bf2579eab07f1408ebc2efd1ce510d592c6"} Dec 06 06:19:54 crc kubenswrapper[4809]: I1206 06:19:54.037524 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"413494e9-0401-4a6c-b85e-8e6be9997d20","Type":"ContainerDied","Data":"7e1832c98b7bbb803ca4af4d248b206551c35844e83b9caceeeff49ade9e0215"} Dec 06 06:19:54 crc kubenswrapper[4809]: I1206 06:19:54.084805 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:19:54 crc kubenswrapper[4809]: W1206 06:19:54.092517 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfd3f000e_9b7c_4f78_aadb_3a580c1441ff.slice/crio-e161a077d126bd21c3150855d83b4df86db5220c69d202f4cf5983565e9063de WatchSource:0}: Error finding container e161a077d126bd21c3150855d83b4df86db5220c69d202f4cf5983565e9063de: Status 404 returned error can't find the container with id e161a077d126bd21c3150855d83b4df86db5220c69d202f4cf5983565e9063de Dec 06 06:19:54 crc kubenswrapper[4809]: I1206 06:19:54.266089 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 06 06:19:54 crc kubenswrapper[4809]: I1206 06:19:54.304046 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 06 06:19:55 crc kubenswrapper[4809]: I1206 06:19:55.051306 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd3f000e-9b7c-4f78-aadb-3a580c1441ff","Type":"ContainerStarted","Data":"0c0dd3b18a329afa92b650647e32f202a2912c7cf59e2eb5b336f7721fdbca2f"} Dec 06 06:19:55 crc kubenswrapper[4809]: I1206 06:19:55.051743 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd3f000e-9b7c-4f78-aadb-3a580c1441ff","Type":"ContainerStarted","Data":"e161a077d126bd21c3150855d83b4df86db5220c69d202f4cf5983565e9063de"} Dec 06 06:19:55 crc kubenswrapper[4809]: I1206 06:19:55.095767 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 06 06:19:55 crc kubenswrapper[4809]: I1206 06:19:55.410235 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 06:19:55 crc kubenswrapper[4809]: I1206 06:19:55.410651 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 06:19:55 crc kubenswrapper[4809]: I1206 06:19:55.410480 4809 scope.go:117] "RemoveContainer" containerID="9bc2dc88b3b9f309ce0169aa510326dd77022dece4c69cab3141df00fe5c0106" Dec 06 06:19:55 crc kubenswrapper[4809]: E1206 06:19:55.411126 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:19:56 crc kubenswrapper[4809]: I1206 06:19:56.063603 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd3f000e-9b7c-4f78-aadb-3a580c1441ff","Type":"ContainerStarted","Data":"809d2993c84e05dd2c952a29844b069e574e6bcc602d224076d7784b43533448"} Dec 06 06:19:56 crc kubenswrapper[4809]: I1206 06:19:56.493140 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="215a53e4-4135-45b5-912c-e25d50e5c2d3" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.246:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 06:19:56 crc kubenswrapper[4809]: I1206 06:19:56.493170 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="215a53e4-4135-45b5-912c-e25d50e5c2d3" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.246:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 06:19:57 crc kubenswrapper[4809]: I1206 06:19:57.084147 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd3f000e-9b7c-4f78-aadb-3a580c1441ff","Type":"ContainerStarted","Data":"60364c8e42fdae8f7e3a7843c59bb73a14b32f72db889110f9b2332a04514907"} Dec 06 06:19:59 crc kubenswrapper[4809]: I1206 06:19:59.117357 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd3f000e-9b7c-4f78-aadb-3a580c1441ff","Type":"ContainerStarted","Data":"ba7aa49c28192d51601f9fd7313c4a2f8901b297738f9b193be5e98c78dd7eff"} Dec 06 06:19:59 crc kubenswrapper[4809]: I1206 06:19:59.119237 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 06:19:59 crc kubenswrapper[4809]: I1206 06:19:59.146625 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.475755896 podStartE2EDuration="6.146598405s" podCreationTimestamp="2025-12-06 06:19:53 +0000 UTC" firstStartedPulling="2025-12-06 06:19:54.09633257 +0000 UTC m=+1718.985315512" lastFinishedPulling="2025-12-06 06:19:58.767175079 +0000 UTC m=+1723.656158021" observedRunningTime="2025-12-06 06:19:59.139568815 +0000 UTC m=+1724.028551807" watchObservedRunningTime="2025-12-06 06:19:59.146598405 +0000 UTC m=+1724.035581357" Dec 06 06:20:00 crc kubenswrapper[4809]: E1206 06:20:00.817505 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb6980646_cff8_45b6_a26b_53d6282f26a9.slice/crio-conmon-e19dfa7717885a4de9c30d0495f2a16b9bc5e74be5bd260ef8679bd31a822ec7.scope\": RecentStats: unable to find data in memory cache]" Dec 06 06:20:00 crc kubenswrapper[4809]: E1206 06:20:00.817829 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb6980646_cff8_45b6_a26b_53d6282f26a9.slice/crio-conmon-e19dfa7717885a4de9c30d0495f2a16b9bc5e74be5bd260ef8679bd31a822ec7.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb6980646_cff8_45b6_a26b_53d6282f26a9.slice/crio-e19dfa7717885a4de9c30d0495f2a16b9bc5e74be5bd260ef8679bd31a822ec7.scope\": RecentStats: unable to find data in memory cache]" Dec 06 06:20:01 crc kubenswrapper[4809]: I1206 06:20:01.143334 4809 generic.go:334] "Generic (PLEG): container finished" podID="2b09b87b-a70c-43d1-936d-0758596953aa" containerID="1aa255953561fd6511f3d4d017ca55dfb6b59f1450b964afb21e4629ac1f5eb0" exitCode=137 Dec 06 06:20:01 crc kubenswrapper[4809]: I1206 06:20:01.143405 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2b09b87b-a70c-43d1-936d-0758596953aa","Type":"ContainerDied","Data":"1aa255953561fd6511f3d4d017ca55dfb6b59f1450b964afb21e4629ac1f5eb0"} Dec 06 06:20:01 crc kubenswrapper[4809]: I1206 06:20:01.144891 4809 generic.go:334] "Generic (PLEG): container finished" podID="b6980646-cff8-45b6-a26b-53d6282f26a9" containerID="e19dfa7717885a4de9c30d0495f2a16b9bc5e74be5bd260ef8679bd31a822ec7" exitCode=137 Dec 06 06:20:01 crc kubenswrapper[4809]: I1206 06:20:01.144922 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"b6980646-cff8-45b6-a26b-53d6282f26a9","Type":"ContainerDied","Data":"e19dfa7717885a4de9c30d0495f2a16b9bc5e74be5bd260ef8679bd31a822ec7"} Dec 06 06:20:01 crc kubenswrapper[4809]: I1206 06:20:01.286629 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 06:20:01 crc kubenswrapper[4809]: I1206 06:20:01.294576 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:20:01 crc kubenswrapper[4809]: I1206 06:20:01.442677 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6980646-cff8-45b6-a26b-53d6282f26a9-combined-ca-bundle\") pod \"b6980646-cff8-45b6-a26b-53d6282f26a9\" (UID: \"b6980646-cff8-45b6-a26b-53d6282f26a9\") " Dec 06 06:20:01 crc kubenswrapper[4809]: I1206 06:20:01.444560 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6980646-cff8-45b6-a26b-53d6282f26a9-config-data\") pod \"b6980646-cff8-45b6-a26b-53d6282f26a9\" (UID: \"b6980646-cff8-45b6-a26b-53d6282f26a9\") " Dec 06 06:20:01 crc kubenswrapper[4809]: I1206 06:20:01.445092 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w2kdb\" (UniqueName: \"kubernetes.io/projected/b6980646-cff8-45b6-a26b-53d6282f26a9-kube-api-access-w2kdb\") pod \"b6980646-cff8-45b6-a26b-53d6282f26a9\" (UID: \"b6980646-cff8-45b6-a26b-53d6282f26a9\") " Dec 06 06:20:01 crc kubenswrapper[4809]: I1206 06:20:01.445240 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b09b87b-a70c-43d1-936d-0758596953aa-logs\") pod \"2b09b87b-a70c-43d1-936d-0758596953aa\" (UID: \"2b09b87b-a70c-43d1-936d-0758596953aa\") " Dec 06 06:20:01 crc kubenswrapper[4809]: I1206 06:20:01.445510 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b09b87b-a70c-43d1-936d-0758596953aa-logs" (OuterVolumeSpecName: "logs") pod "2b09b87b-a70c-43d1-936d-0758596953aa" (UID: "2b09b87b-a70c-43d1-936d-0758596953aa"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:20:01 crc kubenswrapper[4809]: I1206 06:20:01.445557 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b09b87b-a70c-43d1-936d-0758596953aa-combined-ca-bundle\") pod \"2b09b87b-a70c-43d1-936d-0758596953aa\" (UID: \"2b09b87b-a70c-43d1-936d-0758596953aa\") " Dec 06 06:20:01 crc kubenswrapper[4809]: I1206 06:20:01.445595 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g62t4\" (UniqueName: \"kubernetes.io/projected/2b09b87b-a70c-43d1-936d-0758596953aa-kube-api-access-g62t4\") pod \"2b09b87b-a70c-43d1-936d-0758596953aa\" (UID: \"2b09b87b-a70c-43d1-936d-0758596953aa\") " Dec 06 06:20:01 crc kubenswrapper[4809]: I1206 06:20:01.445995 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b09b87b-a70c-43d1-936d-0758596953aa-config-data\") pod \"2b09b87b-a70c-43d1-936d-0758596953aa\" (UID: \"2b09b87b-a70c-43d1-936d-0758596953aa\") " Dec 06 06:20:01 crc kubenswrapper[4809]: I1206 06:20:01.446809 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b09b87b-a70c-43d1-936d-0758596953aa-logs\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:01 crc kubenswrapper[4809]: I1206 06:20:01.449219 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6980646-cff8-45b6-a26b-53d6282f26a9-kube-api-access-w2kdb" (OuterVolumeSpecName: "kube-api-access-w2kdb") pod "b6980646-cff8-45b6-a26b-53d6282f26a9" (UID: "b6980646-cff8-45b6-a26b-53d6282f26a9"). InnerVolumeSpecName "kube-api-access-w2kdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:20:01 crc kubenswrapper[4809]: I1206 06:20:01.449736 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b09b87b-a70c-43d1-936d-0758596953aa-kube-api-access-g62t4" (OuterVolumeSpecName: "kube-api-access-g62t4") pod "2b09b87b-a70c-43d1-936d-0758596953aa" (UID: "2b09b87b-a70c-43d1-936d-0758596953aa"). InnerVolumeSpecName "kube-api-access-g62t4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:20:01 crc kubenswrapper[4809]: I1206 06:20:01.482539 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6980646-cff8-45b6-a26b-53d6282f26a9-config-data" (OuterVolumeSpecName: "config-data") pod "b6980646-cff8-45b6-a26b-53d6282f26a9" (UID: "b6980646-cff8-45b6-a26b-53d6282f26a9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:20:01 crc kubenswrapper[4809]: I1206 06:20:01.486302 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b09b87b-a70c-43d1-936d-0758596953aa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2b09b87b-a70c-43d1-936d-0758596953aa" (UID: "2b09b87b-a70c-43d1-936d-0758596953aa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:20:01 crc kubenswrapper[4809]: I1206 06:20:01.491028 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b09b87b-a70c-43d1-936d-0758596953aa-config-data" (OuterVolumeSpecName: "config-data") pod "2b09b87b-a70c-43d1-936d-0758596953aa" (UID: "2b09b87b-a70c-43d1-936d-0758596953aa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:20:01 crc kubenswrapper[4809]: I1206 06:20:01.492214 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6980646-cff8-45b6-a26b-53d6282f26a9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b6980646-cff8-45b6-a26b-53d6282f26a9" (UID: "b6980646-cff8-45b6-a26b-53d6282f26a9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:20:01 crc kubenswrapper[4809]: I1206 06:20:01.549559 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6980646-cff8-45b6-a26b-53d6282f26a9-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:01 crc kubenswrapper[4809]: I1206 06:20:01.549597 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w2kdb\" (UniqueName: \"kubernetes.io/projected/b6980646-cff8-45b6-a26b-53d6282f26a9-kube-api-access-w2kdb\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:01 crc kubenswrapper[4809]: I1206 06:20:01.549613 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b09b87b-a70c-43d1-936d-0758596953aa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:01 crc kubenswrapper[4809]: I1206 06:20:01.549624 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g62t4\" (UniqueName: \"kubernetes.io/projected/2b09b87b-a70c-43d1-936d-0758596953aa-kube-api-access-g62t4\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:01 crc kubenswrapper[4809]: I1206 06:20:01.549636 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b09b87b-a70c-43d1-936d-0758596953aa-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:01 crc kubenswrapper[4809]: I1206 06:20:01.549646 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6980646-cff8-45b6-a26b-53d6282f26a9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.163600 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2b09b87b-a70c-43d1-936d-0758596953aa","Type":"ContainerDied","Data":"ed09ea49ce462f1c0c5f1b4e23e4e276708a81fa9aa242f26d82305d8cea4ba7"} Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.163654 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.163671 4809 scope.go:117] "RemoveContainer" containerID="1aa255953561fd6511f3d4d017ca55dfb6b59f1450b964afb21e4629ac1f5eb0" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.167620 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"b6980646-cff8-45b6-a26b-53d6282f26a9","Type":"ContainerDied","Data":"c30e1938b5b8b25ef56f2066873a9ad7a9eff7cd7c024e5d375f8720edf838c5"} Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.167722 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.214102 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.229492 4809 scope.go:117] "RemoveContainer" containerID="e504c52d62558bba4a8d2c2eb04a40354fbc3c0b753fef04552ba9d053ba7a23" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.259248 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.282585 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 06 06:20:02 crc kubenswrapper[4809]: E1206 06:20:02.283438 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b09b87b-a70c-43d1-936d-0758596953aa" containerName="nova-metadata-log" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.283453 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b09b87b-a70c-43d1-936d-0758596953aa" containerName="nova-metadata-log" Dec 06 06:20:02 crc kubenswrapper[4809]: E1206 06:20:02.283498 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6980646-cff8-45b6-a26b-53d6282f26a9" containerName="nova-cell1-novncproxy-novncproxy" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.283506 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6980646-cff8-45b6-a26b-53d6282f26a9" containerName="nova-cell1-novncproxy-novncproxy" Dec 06 06:20:02 crc kubenswrapper[4809]: E1206 06:20:02.283536 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b09b87b-a70c-43d1-936d-0758596953aa" containerName="nova-metadata-metadata" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.283544 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b09b87b-a70c-43d1-936d-0758596953aa" containerName="nova-metadata-metadata" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.283818 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b09b87b-a70c-43d1-936d-0758596953aa" containerName="nova-metadata-metadata" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.283839 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6980646-cff8-45b6-a26b-53d6282f26a9" containerName="nova-cell1-novncproxy-novncproxy" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.283870 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b09b87b-a70c-43d1-936d-0758596953aa" containerName="nova-metadata-log" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.285443 4809 scope.go:117] "RemoveContainer" containerID="e19dfa7717885a4de9c30d0495f2a16b9bc5e74be5bd260ef8679bd31a822ec7" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.285641 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.288550 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.288656 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.298241 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.313288 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.325709 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.337038 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.339708 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.343430 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.343496 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.343616 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.351459 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.369160 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f17fe211-3570-4b69-866c-9607c6489341-logs\") pod \"nova-metadata-0\" (UID: \"f17fe211-3570-4b69-866c-9607c6489341\") " pod="openstack/nova-metadata-0" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.369350 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f17fe211-3570-4b69-866c-9607c6489341-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f17fe211-3570-4b69-866c-9607c6489341\") " pod="openstack/nova-metadata-0" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.369439 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f17fe211-3570-4b69-866c-9607c6489341-config-data\") pod \"nova-metadata-0\" (UID: \"f17fe211-3570-4b69-866c-9607c6489341\") " pod="openstack/nova-metadata-0" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.369468 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f17fe211-3570-4b69-866c-9607c6489341-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f17fe211-3570-4b69-866c-9607c6489341\") " pod="openstack/nova-metadata-0" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.369507 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpk4p\" (UniqueName: \"kubernetes.io/projected/f17fe211-3570-4b69-866c-9607c6489341-kube-api-access-xpk4p\") pod \"nova-metadata-0\" (UID: \"f17fe211-3570-4b69-866c-9607c6489341\") " pod="openstack/nova-metadata-0" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.470999 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/663c54d8-faa9-466a-9126-678a92bc92f9-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"663c54d8-faa9-466a-9126-678a92bc92f9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.471056 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwwrn\" (UniqueName: \"kubernetes.io/projected/663c54d8-faa9-466a-9126-678a92bc92f9-kube-api-access-kwwrn\") pod \"nova-cell1-novncproxy-0\" (UID: \"663c54d8-faa9-466a-9126-678a92bc92f9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.471116 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f17fe211-3570-4b69-866c-9607c6489341-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f17fe211-3570-4b69-866c-9607c6489341\") " pod="openstack/nova-metadata-0" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.471141 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/663c54d8-faa9-466a-9126-678a92bc92f9-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"663c54d8-faa9-466a-9126-678a92bc92f9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.471199 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/663c54d8-faa9-466a-9126-678a92bc92f9-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"663c54d8-faa9-466a-9126-678a92bc92f9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.471260 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f17fe211-3570-4b69-866c-9607c6489341-config-data\") pod \"nova-metadata-0\" (UID: \"f17fe211-3570-4b69-866c-9607c6489341\") " pod="openstack/nova-metadata-0" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.471289 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f17fe211-3570-4b69-866c-9607c6489341-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f17fe211-3570-4b69-866c-9607c6489341\") " pod="openstack/nova-metadata-0" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.471319 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpk4p\" (UniqueName: \"kubernetes.io/projected/f17fe211-3570-4b69-866c-9607c6489341-kube-api-access-xpk4p\") pod \"nova-metadata-0\" (UID: \"f17fe211-3570-4b69-866c-9607c6489341\") " pod="openstack/nova-metadata-0" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.471337 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/663c54d8-faa9-466a-9126-678a92bc92f9-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"663c54d8-faa9-466a-9126-678a92bc92f9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.471397 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f17fe211-3570-4b69-866c-9607c6489341-logs\") pod \"nova-metadata-0\" (UID: \"f17fe211-3570-4b69-866c-9607c6489341\") " pod="openstack/nova-metadata-0" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.471880 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f17fe211-3570-4b69-866c-9607c6489341-logs\") pod \"nova-metadata-0\" (UID: \"f17fe211-3570-4b69-866c-9607c6489341\") " pod="openstack/nova-metadata-0" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.477668 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f17fe211-3570-4b69-866c-9607c6489341-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f17fe211-3570-4b69-866c-9607c6489341\") " pod="openstack/nova-metadata-0" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.478278 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f17fe211-3570-4b69-866c-9607c6489341-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f17fe211-3570-4b69-866c-9607c6489341\") " pod="openstack/nova-metadata-0" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.491312 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f17fe211-3570-4b69-866c-9607c6489341-config-data\") pod \"nova-metadata-0\" (UID: \"f17fe211-3570-4b69-866c-9607c6489341\") " pod="openstack/nova-metadata-0" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.494568 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpk4p\" (UniqueName: \"kubernetes.io/projected/f17fe211-3570-4b69-866c-9607c6489341-kube-api-access-xpk4p\") pod \"nova-metadata-0\" (UID: \"f17fe211-3570-4b69-866c-9607c6489341\") " pod="openstack/nova-metadata-0" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.573007 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/663c54d8-faa9-466a-9126-678a92bc92f9-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"663c54d8-faa9-466a-9126-678a92bc92f9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.573061 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwwrn\" (UniqueName: \"kubernetes.io/projected/663c54d8-faa9-466a-9126-678a92bc92f9-kube-api-access-kwwrn\") pod \"nova-cell1-novncproxy-0\" (UID: \"663c54d8-faa9-466a-9126-678a92bc92f9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.573127 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/663c54d8-faa9-466a-9126-678a92bc92f9-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"663c54d8-faa9-466a-9126-678a92bc92f9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.573191 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/663c54d8-faa9-466a-9126-678a92bc92f9-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"663c54d8-faa9-466a-9126-678a92bc92f9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.573254 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/663c54d8-faa9-466a-9126-678a92bc92f9-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"663c54d8-faa9-466a-9126-678a92bc92f9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.576250 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/663c54d8-faa9-466a-9126-678a92bc92f9-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"663c54d8-faa9-466a-9126-678a92bc92f9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.576886 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/663c54d8-faa9-466a-9126-678a92bc92f9-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"663c54d8-faa9-466a-9126-678a92bc92f9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.577093 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/663c54d8-faa9-466a-9126-678a92bc92f9-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"663c54d8-faa9-466a-9126-678a92bc92f9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.577343 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/663c54d8-faa9-466a-9126-678a92bc92f9-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"663c54d8-faa9-466a-9126-678a92bc92f9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.589982 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwwrn\" (UniqueName: \"kubernetes.io/projected/663c54d8-faa9-466a-9126-678a92bc92f9-kube-api-access-kwwrn\") pod \"nova-cell1-novncproxy-0\" (UID: \"663c54d8-faa9-466a-9126-678a92bc92f9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.610982 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 06:20:02 crc kubenswrapper[4809]: I1206 06:20:02.660557 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:20:03 crc kubenswrapper[4809]: I1206 06:20:03.089543 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 06:20:03 crc kubenswrapper[4809]: W1206 06:20:03.091162 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf17fe211_3570_4b69_866c_9607c6489341.slice/crio-87bc3a89585fba86c856829307710e58092b7643b3cecb821d9055b4a507c3a6 WatchSource:0}: Error finding container 87bc3a89585fba86c856829307710e58092b7643b3cecb821d9055b4a507c3a6: Status 404 returned error can't find the container with id 87bc3a89585fba86c856829307710e58092b7643b3cecb821d9055b4a507c3a6 Dec 06 06:20:03 crc kubenswrapper[4809]: I1206 06:20:03.187278 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f17fe211-3570-4b69-866c-9607c6489341","Type":"ContainerStarted","Data":"87bc3a89585fba86c856829307710e58092b7643b3cecb821d9055b4a507c3a6"} Dec 06 06:20:03 crc kubenswrapper[4809]: I1206 06:20:03.204873 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 06:20:03 crc kubenswrapper[4809]: I1206 06:20:03.404179 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b09b87b-a70c-43d1-936d-0758596953aa" path="/var/lib/kubelet/pods/2b09b87b-a70c-43d1-936d-0758596953aa/volumes" Dec 06 06:20:03 crc kubenswrapper[4809]: I1206 06:20:03.405373 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6980646-cff8-45b6-a26b-53d6282f26a9" path="/var/lib/kubelet/pods/b6980646-cff8-45b6-a26b-53d6282f26a9/volumes" Dec 06 06:20:04 crc kubenswrapper[4809]: I1206 06:20:04.204397 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f17fe211-3570-4b69-866c-9607c6489341","Type":"ContainerStarted","Data":"5375b409ca2dbb8cc53e5c42961b8fbe2e4b34d2ddaf260aae13f98079f6fffd"} Dec 06 06:20:04 crc kubenswrapper[4809]: I1206 06:20:04.206804 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"663c54d8-faa9-466a-9126-678a92bc92f9","Type":"ContainerStarted","Data":"5cd597902d3b2e2e6cebaa0df7cc9f72c4fb72e5b0e75f1b42f210aeb3c75363"} Dec 06 06:20:04 crc kubenswrapper[4809]: I1206 06:20:04.206840 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"663c54d8-faa9-466a-9126-678a92bc92f9","Type":"ContainerStarted","Data":"346fa498981277e47cf316d4d9e17311a6711b11903c37a707af206de5b4d5a5"} Dec 06 06:20:04 crc kubenswrapper[4809]: I1206 06:20:04.236758 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.236735924 podStartE2EDuration="2.236735924s" podCreationTimestamp="2025-12-06 06:20:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:20:04.222675295 +0000 UTC m=+1729.111658237" watchObservedRunningTime="2025-12-06 06:20:04.236735924 +0000 UTC m=+1729.125718886" Dec 06 06:20:05 crc kubenswrapper[4809]: I1206 06:20:05.223013 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f17fe211-3570-4b69-866c-9607c6489341","Type":"ContainerStarted","Data":"9399ca5d70e9db739895e6db6766cd6ad4c43d81e9ba06135277d222b541cc75"} Dec 06 06:20:05 crc kubenswrapper[4809]: I1206 06:20:05.247124 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.247092613 podStartE2EDuration="3.247092613s" podCreationTimestamp="2025-12-06 06:20:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:20:05.245586272 +0000 UTC m=+1730.134569224" watchObservedRunningTime="2025-12-06 06:20:05.247092613 +0000 UTC m=+1730.136075575" Dec 06 06:20:05 crc kubenswrapper[4809]: I1206 06:20:05.409088 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 06 06:20:05 crc kubenswrapper[4809]: I1206 06:20:05.409797 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 06 06:20:05 crc kubenswrapper[4809]: I1206 06:20:05.412973 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 06 06:20:05 crc kubenswrapper[4809]: I1206 06:20:05.416021 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 06 06:20:06 crc kubenswrapper[4809]: I1206 06:20:06.235519 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 06 06:20:06 crc kubenswrapper[4809]: I1206 06:20:06.239109 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 06 06:20:06 crc kubenswrapper[4809]: I1206 06:20:06.391761 4809 scope.go:117] "RemoveContainer" containerID="9bc2dc88b3b9f309ce0169aa510326dd77022dece4c69cab3141df00fe5c0106" Dec 06 06:20:06 crc kubenswrapper[4809]: E1206 06:20:06.392106 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:20:06 crc kubenswrapper[4809]: I1206 06:20:06.476008 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-2kbmw"] Dec 06 06:20:06 crc kubenswrapper[4809]: I1206 06:20:06.478294 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-2kbmw" Dec 06 06:20:06 crc kubenswrapper[4809]: I1206 06:20:06.504679 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-2kbmw"] Dec 06 06:20:06 crc kubenswrapper[4809]: I1206 06:20:06.581910 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/624776a0-1d73-4785-9c78-45a4a30e2f80-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7bbf7cf9-2kbmw\" (UID: \"624776a0-1d73-4785-9c78-45a4a30e2f80\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-2kbmw" Dec 06 06:20:06 crc kubenswrapper[4809]: I1206 06:20:06.582043 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/624776a0-1d73-4785-9c78-45a4a30e2f80-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7bbf7cf9-2kbmw\" (UID: \"624776a0-1d73-4785-9c78-45a4a30e2f80\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-2kbmw" Dec 06 06:20:06 crc kubenswrapper[4809]: I1206 06:20:06.582070 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/624776a0-1d73-4785-9c78-45a4a30e2f80-config\") pod \"dnsmasq-dns-6b7bbf7cf9-2kbmw\" (UID: \"624776a0-1d73-4785-9c78-45a4a30e2f80\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-2kbmw" Dec 06 06:20:06 crc kubenswrapper[4809]: I1206 06:20:06.582092 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6br6\" (UniqueName: \"kubernetes.io/projected/624776a0-1d73-4785-9c78-45a4a30e2f80-kube-api-access-z6br6\") pod \"dnsmasq-dns-6b7bbf7cf9-2kbmw\" (UID: \"624776a0-1d73-4785-9c78-45a4a30e2f80\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-2kbmw" Dec 06 06:20:06 crc kubenswrapper[4809]: I1206 06:20:06.582115 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/624776a0-1d73-4785-9c78-45a4a30e2f80-dns-svc\") pod \"dnsmasq-dns-6b7bbf7cf9-2kbmw\" (UID: \"624776a0-1d73-4785-9c78-45a4a30e2f80\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-2kbmw" Dec 06 06:20:06 crc kubenswrapper[4809]: I1206 06:20:06.582184 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/624776a0-1d73-4785-9c78-45a4a30e2f80-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7bbf7cf9-2kbmw\" (UID: \"624776a0-1d73-4785-9c78-45a4a30e2f80\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-2kbmw" Dec 06 06:20:06 crc kubenswrapper[4809]: I1206 06:20:06.684546 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/624776a0-1d73-4785-9c78-45a4a30e2f80-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7bbf7cf9-2kbmw\" (UID: \"624776a0-1d73-4785-9c78-45a4a30e2f80\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-2kbmw" Dec 06 06:20:06 crc kubenswrapper[4809]: I1206 06:20:06.684676 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/624776a0-1d73-4785-9c78-45a4a30e2f80-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7bbf7cf9-2kbmw\" (UID: \"624776a0-1d73-4785-9c78-45a4a30e2f80\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-2kbmw" Dec 06 06:20:06 crc kubenswrapper[4809]: I1206 06:20:06.684704 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/624776a0-1d73-4785-9c78-45a4a30e2f80-config\") pod \"dnsmasq-dns-6b7bbf7cf9-2kbmw\" (UID: \"624776a0-1d73-4785-9c78-45a4a30e2f80\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-2kbmw" Dec 06 06:20:06 crc kubenswrapper[4809]: I1206 06:20:06.684725 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6br6\" (UniqueName: \"kubernetes.io/projected/624776a0-1d73-4785-9c78-45a4a30e2f80-kube-api-access-z6br6\") pod \"dnsmasq-dns-6b7bbf7cf9-2kbmw\" (UID: \"624776a0-1d73-4785-9c78-45a4a30e2f80\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-2kbmw" Dec 06 06:20:06 crc kubenswrapper[4809]: I1206 06:20:06.684751 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/624776a0-1d73-4785-9c78-45a4a30e2f80-dns-svc\") pod \"dnsmasq-dns-6b7bbf7cf9-2kbmw\" (UID: \"624776a0-1d73-4785-9c78-45a4a30e2f80\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-2kbmw" Dec 06 06:20:06 crc kubenswrapper[4809]: I1206 06:20:06.684816 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/624776a0-1d73-4785-9c78-45a4a30e2f80-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7bbf7cf9-2kbmw\" (UID: \"624776a0-1d73-4785-9c78-45a4a30e2f80\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-2kbmw" Dec 06 06:20:06 crc kubenswrapper[4809]: I1206 06:20:06.686189 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/624776a0-1d73-4785-9c78-45a4a30e2f80-dns-svc\") pod \"dnsmasq-dns-6b7bbf7cf9-2kbmw\" (UID: \"624776a0-1d73-4785-9c78-45a4a30e2f80\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-2kbmw" Dec 06 06:20:06 crc kubenswrapper[4809]: I1206 06:20:06.686241 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/624776a0-1d73-4785-9c78-45a4a30e2f80-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7bbf7cf9-2kbmw\" (UID: \"624776a0-1d73-4785-9c78-45a4a30e2f80\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-2kbmw" Dec 06 06:20:06 crc kubenswrapper[4809]: I1206 06:20:06.686817 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/624776a0-1d73-4785-9c78-45a4a30e2f80-config\") pod \"dnsmasq-dns-6b7bbf7cf9-2kbmw\" (UID: \"624776a0-1d73-4785-9c78-45a4a30e2f80\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-2kbmw" Dec 06 06:20:06 crc kubenswrapper[4809]: I1206 06:20:06.686845 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/624776a0-1d73-4785-9c78-45a4a30e2f80-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7bbf7cf9-2kbmw\" (UID: \"624776a0-1d73-4785-9c78-45a4a30e2f80\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-2kbmw" Dec 06 06:20:06 crc kubenswrapper[4809]: I1206 06:20:06.687379 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/624776a0-1d73-4785-9c78-45a4a30e2f80-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7bbf7cf9-2kbmw\" (UID: \"624776a0-1d73-4785-9c78-45a4a30e2f80\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-2kbmw" Dec 06 06:20:06 crc kubenswrapper[4809]: I1206 06:20:06.709495 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6br6\" (UniqueName: \"kubernetes.io/projected/624776a0-1d73-4785-9c78-45a4a30e2f80-kube-api-access-z6br6\") pod \"dnsmasq-dns-6b7bbf7cf9-2kbmw\" (UID: \"624776a0-1d73-4785-9c78-45a4a30e2f80\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-2kbmw" Dec 06 06:20:06 crc kubenswrapper[4809]: I1206 06:20:06.821802 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-2kbmw" Dec 06 06:20:07 crc kubenswrapper[4809]: I1206 06:20:07.494132 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-2kbmw"] Dec 06 06:20:07 crc kubenswrapper[4809]: I1206 06:20:07.611044 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 06 06:20:07 crc kubenswrapper[4809]: I1206 06:20:07.611440 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 06 06:20:07 crc kubenswrapper[4809]: I1206 06:20:07.667518 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:20:08 crc kubenswrapper[4809]: I1206 06:20:08.273281 4809 generic.go:334] "Generic (PLEG): container finished" podID="624776a0-1d73-4785-9c78-45a4a30e2f80" containerID="d29627bf8956afc444882ed97f05eb1f9679336e8432e00c64d3c772a8acd554" exitCode=0 Dec 06 06:20:08 crc kubenswrapper[4809]: I1206 06:20:08.273332 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-2kbmw" event={"ID":"624776a0-1d73-4785-9c78-45a4a30e2f80","Type":"ContainerDied","Data":"d29627bf8956afc444882ed97f05eb1f9679336e8432e00c64d3c772a8acd554"} Dec 06 06:20:08 crc kubenswrapper[4809]: I1206 06:20:08.273668 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-2kbmw" event={"ID":"624776a0-1d73-4785-9c78-45a4a30e2f80","Type":"ContainerStarted","Data":"1859763d34b18ac613b6a7721b806521d6b0f3c3482f7c9318562059bb68864f"} Dec 06 06:20:09 crc kubenswrapper[4809]: I1206 06:20:09.046907 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 06 06:20:09 crc kubenswrapper[4809]: I1206 06:20:09.141953 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:20:09 crc kubenswrapper[4809]: I1206 06:20:09.142578 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fd3f000e-9b7c-4f78-aadb-3a580c1441ff" containerName="ceilometer-central-agent" containerID="cri-o://0c0dd3b18a329afa92b650647e32f202a2912c7cf59e2eb5b336f7721fdbca2f" gracePeriod=30 Dec 06 06:20:09 crc kubenswrapper[4809]: I1206 06:20:09.142642 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fd3f000e-9b7c-4f78-aadb-3a580c1441ff" containerName="proxy-httpd" containerID="cri-o://ba7aa49c28192d51601f9fd7313c4a2f8901b297738f9b193be5e98c78dd7eff" gracePeriod=30 Dec 06 06:20:09 crc kubenswrapper[4809]: I1206 06:20:09.142659 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fd3f000e-9b7c-4f78-aadb-3a580c1441ff" containerName="sg-core" containerID="cri-o://60364c8e42fdae8f7e3a7843c59bb73a14b32f72db889110f9b2332a04514907" gracePeriod=30 Dec 06 06:20:09 crc kubenswrapper[4809]: I1206 06:20:09.142722 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fd3f000e-9b7c-4f78-aadb-3a580c1441ff" containerName="ceilometer-notification-agent" containerID="cri-o://809d2993c84e05dd2c952a29844b069e574e6bcc602d224076d7784b43533448" gracePeriod=30 Dec 06 06:20:09 crc kubenswrapper[4809]: I1206 06:20:09.156701 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="fd3f000e-9b7c-4f78-aadb-3a580c1441ff" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.247:3000/\": EOF" Dec 06 06:20:09 crc kubenswrapper[4809]: I1206 06:20:09.292881 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-2kbmw" event={"ID":"624776a0-1d73-4785-9c78-45a4a30e2f80","Type":"ContainerStarted","Data":"f03d7617420960dcd030c464e2ad37d2d398434079e59ba703a6bc97f5a36856"} Dec 06 06:20:09 crc kubenswrapper[4809]: I1206 06:20:09.294214 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b7bbf7cf9-2kbmw" Dec 06 06:20:09 crc kubenswrapper[4809]: I1206 06:20:09.300710 4809 generic.go:334] "Generic (PLEG): container finished" podID="fd3f000e-9b7c-4f78-aadb-3a580c1441ff" containerID="60364c8e42fdae8f7e3a7843c59bb73a14b32f72db889110f9b2332a04514907" exitCode=2 Dec 06 06:20:09 crc kubenswrapper[4809]: I1206 06:20:09.301050 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="215a53e4-4135-45b5-912c-e25d50e5c2d3" containerName="nova-api-log" containerID="cri-o://38942159c5dfe1eeb521ee0e2e56ac8d200f1be759cd9896b948b276d2094f2e" gracePeriod=30 Dec 06 06:20:09 crc kubenswrapper[4809]: I1206 06:20:09.301329 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd3f000e-9b7c-4f78-aadb-3a580c1441ff","Type":"ContainerDied","Data":"60364c8e42fdae8f7e3a7843c59bb73a14b32f72db889110f9b2332a04514907"} Dec 06 06:20:09 crc kubenswrapper[4809]: I1206 06:20:09.301450 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="215a53e4-4135-45b5-912c-e25d50e5c2d3" containerName="nova-api-api" containerID="cri-o://87401fb28e1de011b3a112da08d30919cd13ce434f4e76261389e19eb091b1b8" gracePeriod=30 Dec 06 06:20:09 crc kubenswrapper[4809]: I1206 06:20:09.323004 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b7bbf7cf9-2kbmw" podStartSLOduration=3.3229825379999998 podStartE2EDuration="3.322982538s" podCreationTimestamp="2025-12-06 06:20:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:20:09.315713852 +0000 UTC m=+1734.204696814" watchObservedRunningTime="2025-12-06 06:20:09.322982538 +0000 UTC m=+1734.211965480" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.106899 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.194330 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd3f000e-9b7c-4f78-aadb-3a580c1441ff-config-data\") pod \"fd3f000e-9b7c-4f78-aadb-3a580c1441ff\" (UID: \"fd3f000e-9b7c-4f78-aadb-3a580c1441ff\") " Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.194404 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fd3f000e-9b7c-4f78-aadb-3a580c1441ff-sg-core-conf-yaml\") pod \"fd3f000e-9b7c-4f78-aadb-3a580c1441ff\" (UID: \"fd3f000e-9b7c-4f78-aadb-3a580c1441ff\") " Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.194426 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd3f000e-9b7c-4f78-aadb-3a580c1441ff-scripts\") pod \"fd3f000e-9b7c-4f78-aadb-3a580c1441ff\" (UID: \"fd3f000e-9b7c-4f78-aadb-3a580c1441ff\") " Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.194459 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd3f000e-9b7c-4f78-aadb-3a580c1441ff-log-httpd\") pod \"fd3f000e-9b7c-4f78-aadb-3a580c1441ff\" (UID: \"fd3f000e-9b7c-4f78-aadb-3a580c1441ff\") " Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.194534 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd3f000e-9b7c-4f78-aadb-3a580c1441ff-combined-ca-bundle\") pod \"fd3f000e-9b7c-4f78-aadb-3a580c1441ff\" (UID: \"fd3f000e-9b7c-4f78-aadb-3a580c1441ff\") " Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.194608 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9xfw\" (UniqueName: \"kubernetes.io/projected/fd3f000e-9b7c-4f78-aadb-3a580c1441ff-kube-api-access-v9xfw\") pod \"fd3f000e-9b7c-4f78-aadb-3a580c1441ff\" (UID: \"fd3f000e-9b7c-4f78-aadb-3a580c1441ff\") " Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.194641 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd3f000e-9b7c-4f78-aadb-3a580c1441ff-run-httpd\") pod \"fd3f000e-9b7c-4f78-aadb-3a580c1441ff\" (UID: \"fd3f000e-9b7c-4f78-aadb-3a580c1441ff\") " Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.195075 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd3f000e-9b7c-4f78-aadb-3a580c1441ff-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "fd3f000e-9b7c-4f78-aadb-3a580c1441ff" (UID: "fd3f000e-9b7c-4f78-aadb-3a580c1441ff"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.195097 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd3f000e-9b7c-4f78-aadb-3a580c1441ff-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "fd3f000e-9b7c-4f78-aadb-3a580c1441ff" (UID: "fd3f000e-9b7c-4f78-aadb-3a580c1441ff"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.201539 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd3f000e-9b7c-4f78-aadb-3a580c1441ff-kube-api-access-v9xfw" (OuterVolumeSpecName: "kube-api-access-v9xfw") pod "fd3f000e-9b7c-4f78-aadb-3a580c1441ff" (UID: "fd3f000e-9b7c-4f78-aadb-3a580c1441ff"). InnerVolumeSpecName "kube-api-access-v9xfw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.202011 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd3f000e-9b7c-4f78-aadb-3a580c1441ff-scripts" (OuterVolumeSpecName: "scripts") pod "fd3f000e-9b7c-4f78-aadb-3a580c1441ff" (UID: "fd3f000e-9b7c-4f78-aadb-3a580c1441ff"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.242057 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd3f000e-9b7c-4f78-aadb-3a580c1441ff-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "fd3f000e-9b7c-4f78-aadb-3a580c1441ff" (UID: "fd3f000e-9b7c-4f78-aadb-3a580c1441ff"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.297557 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9xfw\" (UniqueName: \"kubernetes.io/projected/fd3f000e-9b7c-4f78-aadb-3a580c1441ff-kube-api-access-v9xfw\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.297593 4809 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd3f000e-9b7c-4f78-aadb-3a580c1441ff-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.297605 4809 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fd3f000e-9b7c-4f78-aadb-3a580c1441ff-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.297615 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd3f000e-9b7c-4f78-aadb-3a580c1441ff-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.297623 4809 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd3f000e-9b7c-4f78-aadb-3a580c1441ff-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.315268 4809 generic.go:334] "Generic (PLEG): container finished" podID="fd3f000e-9b7c-4f78-aadb-3a580c1441ff" containerID="ba7aa49c28192d51601f9fd7313c4a2f8901b297738f9b193be5e98c78dd7eff" exitCode=0 Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.315335 4809 generic.go:334] "Generic (PLEG): container finished" podID="fd3f000e-9b7c-4f78-aadb-3a580c1441ff" containerID="809d2993c84e05dd2c952a29844b069e574e6bcc602d224076d7784b43533448" exitCode=0 Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.315348 4809 generic.go:334] "Generic (PLEG): container finished" podID="fd3f000e-9b7c-4f78-aadb-3a580c1441ff" containerID="0c0dd3b18a329afa92b650647e32f202a2912c7cf59e2eb5b336f7721fdbca2f" exitCode=0 Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.315387 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd3f000e-9b7c-4f78-aadb-3a580c1441ff","Type":"ContainerDied","Data":"ba7aa49c28192d51601f9fd7313c4a2f8901b297738f9b193be5e98c78dd7eff"} Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.315425 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd3f000e-9b7c-4f78-aadb-3a580c1441ff","Type":"ContainerDied","Data":"809d2993c84e05dd2c952a29844b069e574e6bcc602d224076d7784b43533448"} Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.315434 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd3f000e-9b7c-4f78-aadb-3a580c1441ff","Type":"ContainerDied","Data":"0c0dd3b18a329afa92b650647e32f202a2912c7cf59e2eb5b336f7721fdbca2f"} Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.315444 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd3f000e-9b7c-4f78-aadb-3a580c1441ff","Type":"ContainerDied","Data":"e161a077d126bd21c3150855d83b4df86db5220c69d202f4cf5983565e9063de"} Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.315458 4809 scope.go:117] "RemoveContainer" containerID="ba7aa49c28192d51601f9fd7313c4a2f8901b297738f9b193be5e98c78dd7eff" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.315592 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.320354 4809 generic.go:334] "Generic (PLEG): container finished" podID="215a53e4-4135-45b5-912c-e25d50e5c2d3" containerID="38942159c5dfe1eeb521ee0e2e56ac8d200f1be759cd9896b948b276d2094f2e" exitCode=143 Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.321357 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"215a53e4-4135-45b5-912c-e25d50e5c2d3","Type":"ContainerDied","Data":"38942159c5dfe1eeb521ee0e2e56ac8d200f1be759cd9896b948b276d2094f2e"} Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.335957 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd3f000e-9b7c-4f78-aadb-3a580c1441ff-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fd3f000e-9b7c-4f78-aadb-3a580c1441ff" (UID: "fd3f000e-9b7c-4f78-aadb-3a580c1441ff"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.341324 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd3f000e-9b7c-4f78-aadb-3a580c1441ff-config-data" (OuterVolumeSpecName: "config-data") pod "fd3f000e-9b7c-4f78-aadb-3a580c1441ff" (UID: "fd3f000e-9b7c-4f78-aadb-3a580c1441ff"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.398722 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd3f000e-9b7c-4f78-aadb-3a580c1441ff-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.398764 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd3f000e-9b7c-4f78-aadb-3a580c1441ff-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.444213 4809 scope.go:117] "RemoveContainer" containerID="60364c8e42fdae8f7e3a7843c59bb73a14b32f72db889110f9b2332a04514907" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.468667 4809 scope.go:117] "RemoveContainer" containerID="809d2993c84e05dd2c952a29844b069e574e6bcc602d224076d7784b43533448" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.493399 4809 scope.go:117] "RemoveContainer" containerID="0c0dd3b18a329afa92b650647e32f202a2912c7cf59e2eb5b336f7721fdbca2f" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.513633 4809 scope.go:117] "RemoveContainer" containerID="ba7aa49c28192d51601f9fd7313c4a2f8901b297738f9b193be5e98c78dd7eff" Dec 06 06:20:10 crc kubenswrapper[4809]: E1206 06:20:10.514200 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba7aa49c28192d51601f9fd7313c4a2f8901b297738f9b193be5e98c78dd7eff\": container with ID starting with ba7aa49c28192d51601f9fd7313c4a2f8901b297738f9b193be5e98c78dd7eff not found: ID does not exist" containerID="ba7aa49c28192d51601f9fd7313c4a2f8901b297738f9b193be5e98c78dd7eff" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.514248 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba7aa49c28192d51601f9fd7313c4a2f8901b297738f9b193be5e98c78dd7eff"} err="failed to get container status \"ba7aa49c28192d51601f9fd7313c4a2f8901b297738f9b193be5e98c78dd7eff\": rpc error: code = NotFound desc = could not find container \"ba7aa49c28192d51601f9fd7313c4a2f8901b297738f9b193be5e98c78dd7eff\": container with ID starting with ba7aa49c28192d51601f9fd7313c4a2f8901b297738f9b193be5e98c78dd7eff not found: ID does not exist" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.514274 4809 scope.go:117] "RemoveContainer" containerID="60364c8e42fdae8f7e3a7843c59bb73a14b32f72db889110f9b2332a04514907" Dec 06 06:20:10 crc kubenswrapper[4809]: E1206 06:20:10.514595 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60364c8e42fdae8f7e3a7843c59bb73a14b32f72db889110f9b2332a04514907\": container with ID starting with 60364c8e42fdae8f7e3a7843c59bb73a14b32f72db889110f9b2332a04514907 not found: ID does not exist" containerID="60364c8e42fdae8f7e3a7843c59bb73a14b32f72db889110f9b2332a04514907" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.514634 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60364c8e42fdae8f7e3a7843c59bb73a14b32f72db889110f9b2332a04514907"} err="failed to get container status \"60364c8e42fdae8f7e3a7843c59bb73a14b32f72db889110f9b2332a04514907\": rpc error: code = NotFound desc = could not find container \"60364c8e42fdae8f7e3a7843c59bb73a14b32f72db889110f9b2332a04514907\": container with ID starting with 60364c8e42fdae8f7e3a7843c59bb73a14b32f72db889110f9b2332a04514907 not found: ID does not exist" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.514658 4809 scope.go:117] "RemoveContainer" containerID="809d2993c84e05dd2c952a29844b069e574e6bcc602d224076d7784b43533448" Dec 06 06:20:10 crc kubenswrapper[4809]: E1206 06:20:10.515079 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"809d2993c84e05dd2c952a29844b069e574e6bcc602d224076d7784b43533448\": container with ID starting with 809d2993c84e05dd2c952a29844b069e574e6bcc602d224076d7784b43533448 not found: ID does not exist" containerID="809d2993c84e05dd2c952a29844b069e574e6bcc602d224076d7784b43533448" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.515111 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"809d2993c84e05dd2c952a29844b069e574e6bcc602d224076d7784b43533448"} err="failed to get container status \"809d2993c84e05dd2c952a29844b069e574e6bcc602d224076d7784b43533448\": rpc error: code = NotFound desc = could not find container \"809d2993c84e05dd2c952a29844b069e574e6bcc602d224076d7784b43533448\": container with ID starting with 809d2993c84e05dd2c952a29844b069e574e6bcc602d224076d7784b43533448 not found: ID does not exist" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.515127 4809 scope.go:117] "RemoveContainer" containerID="0c0dd3b18a329afa92b650647e32f202a2912c7cf59e2eb5b336f7721fdbca2f" Dec 06 06:20:10 crc kubenswrapper[4809]: E1206 06:20:10.515920 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c0dd3b18a329afa92b650647e32f202a2912c7cf59e2eb5b336f7721fdbca2f\": container with ID starting with 0c0dd3b18a329afa92b650647e32f202a2912c7cf59e2eb5b336f7721fdbca2f not found: ID does not exist" containerID="0c0dd3b18a329afa92b650647e32f202a2912c7cf59e2eb5b336f7721fdbca2f" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.515989 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c0dd3b18a329afa92b650647e32f202a2912c7cf59e2eb5b336f7721fdbca2f"} err="failed to get container status \"0c0dd3b18a329afa92b650647e32f202a2912c7cf59e2eb5b336f7721fdbca2f\": rpc error: code = NotFound desc = could not find container \"0c0dd3b18a329afa92b650647e32f202a2912c7cf59e2eb5b336f7721fdbca2f\": container with ID starting with 0c0dd3b18a329afa92b650647e32f202a2912c7cf59e2eb5b336f7721fdbca2f not found: ID does not exist" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.516005 4809 scope.go:117] "RemoveContainer" containerID="ba7aa49c28192d51601f9fd7313c4a2f8901b297738f9b193be5e98c78dd7eff" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.516264 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba7aa49c28192d51601f9fd7313c4a2f8901b297738f9b193be5e98c78dd7eff"} err="failed to get container status \"ba7aa49c28192d51601f9fd7313c4a2f8901b297738f9b193be5e98c78dd7eff\": rpc error: code = NotFound desc = could not find container \"ba7aa49c28192d51601f9fd7313c4a2f8901b297738f9b193be5e98c78dd7eff\": container with ID starting with ba7aa49c28192d51601f9fd7313c4a2f8901b297738f9b193be5e98c78dd7eff not found: ID does not exist" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.516299 4809 scope.go:117] "RemoveContainer" containerID="60364c8e42fdae8f7e3a7843c59bb73a14b32f72db889110f9b2332a04514907" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.516552 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60364c8e42fdae8f7e3a7843c59bb73a14b32f72db889110f9b2332a04514907"} err="failed to get container status \"60364c8e42fdae8f7e3a7843c59bb73a14b32f72db889110f9b2332a04514907\": rpc error: code = NotFound desc = could not find container \"60364c8e42fdae8f7e3a7843c59bb73a14b32f72db889110f9b2332a04514907\": container with ID starting with 60364c8e42fdae8f7e3a7843c59bb73a14b32f72db889110f9b2332a04514907 not found: ID does not exist" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.516574 4809 scope.go:117] "RemoveContainer" containerID="809d2993c84e05dd2c952a29844b069e574e6bcc602d224076d7784b43533448" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.517019 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"809d2993c84e05dd2c952a29844b069e574e6bcc602d224076d7784b43533448"} err="failed to get container status \"809d2993c84e05dd2c952a29844b069e574e6bcc602d224076d7784b43533448\": rpc error: code = NotFound desc = could not find container \"809d2993c84e05dd2c952a29844b069e574e6bcc602d224076d7784b43533448\": container with ID starting with 809d2993c84e05dd2c952a29844b069e574e6bcc602d224076d7784b43533448 not found: ID does not exist" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.517049 4809 scope.go:117] "RemoveContainer" containerID="0c0dd3b18a329afa92b650647e32f202a2912c7cf59e2eb5b336f7721fdbca2f" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.517400 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c0dd3b18a329afa92b650647e32f202a2912c7cf59e2eb5b336f7721fdbca2f"} err="failed to get container status \"0c0dd3b18a329afa92b650647e32f202a2912c7cf59e2eb5b336f7721fdbca2f\": rpc error: code = NotFound desc = could not find container \"0c0dd3b18a329afa92b650647e32f202a2912c7cf59e2eb5b336f7721fdbca2f\": container with ID starting with 0c0dd3b18a329afa92b650647e32f202a2912c7cf59e2eb5b336f7721fdbca2f not found: ID does not exist" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.517422 4809 scope.go:117] "RemoveContainer" containerID="ba7aa49c28192d51601f9fd7313c4a2f8901b297738f9b193be5e98c78dd7eff" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.517677 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba7aa49c28192d51601f9fd7313c4a2f8901b297738f9b193be5e98c78dd7eff"} err="failed to get container status \"ba7aa49c28192d51601f9fd7313c4a2f8901b297738f9b193be5e98c78dd7eff\": rpc error: code = NotFound desc = could not find container \"ba7aa49c28192d51601f9fd7313c4a2f8901b297738f9b193be5e98c78dd7eff\": container with ID starting with ba7aa49c28192d51601f9fd7313c4a2f8901b297738f9b193be5e98c78dd7eff not found: ID does not exist" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.517701 4809 scope.go:117] "RemoveContainer" containerID="60364c8e42fdae8f7e3a7843c59bb73a14b32f72db889110f9b2332a04514907" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.518056 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60364c8e42fdae8f7e3a7843c59bb73a14b32f72db889110f9b2332a04514907"} err="failed to get container status \"60364c8e42fdae8f7e3a7843c59bb73a14b32f72db889110f9b2332a04514907\": rpc error: code = NotFound desc = could not find container \"60364c8e42fdae8f7e3a7843c59bb73a14b32f72db889110f9b2332a04514907\": container with ID starting with 60364c8e42fdae8f7e3a7843c59bb73a14b32f72db889110f9b2332a04514907 not found: ID does not exist" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.518077 4809 scope.go:117] "RemoveContainer" containerID="809d2993c84e05dd2c952a29844b069e574e6bcc602d224076d7784b43533448" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.518449 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"809d2993c84e05dd2c952a29844b069e574e6bcc602d224076d7784b43533448"} err="failed to get container status \"809d2993c84e05dd2c952a29844b069e574e6bcc602d224076d7784b43533448\": rpc error: code = NotFound desc = could not find container \"809d2993c84e05dd2c952a29844b069e574e6bcc602d224076d7784b43533448\": container with ID starting with 809d2993c84e05dd2c952a29844b069e574e6bcc602d224076d7784b43533448 not found: ID does not exist" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.518470 4809 scope.go:117] "RemoveContainer" containerID="0c0dd3b18a329afa92b650647e32f202a2912c7cf59e2eb5b336f7721fdbca2f" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.519062 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c0dd3b18a329afa92b650647e32f202a2912c7cf59e2eb5b336f7721fdbca2f"} err="failed to get container status \"0c0dd3b18a329afa92b650647e32f202a2912c7cf59e2eb5b336f7721fdbca2f\": rpc error: code = NotFound desc = could not find container \"0c0dd3b18a329afa92b650647e32f202a2912c7cf59e2eb5b336f7721fdbca2f\": container with ID starting with 0c0dd3b18a329afa92b650647e32f202a2912c7cf59e2eb5b336f7721fdbca2f not found: ID does not exist" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.650114 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.661921 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.681468 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:20:10 crc kubenswrapper[4809]: E1206 06:20:10.682863 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd3f000e-9b7c-4f78-aadb-3a580c1441ff" containerName="sg-core" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.682918 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd3f000e-9b7c-4f78-aadb-3a580c1441ff" containerName="sg-core" Dec 06 06:20:10 crc kubenswrapper[4809]: E1206 06:20:10.682956 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd3f000e-9b7c-4f78-aadb-3a580c1441ff" containerName="ceilometer-notification-agent" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.682966 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd3f000e-9b7c-4f78-aadb-3a580c1441ff" containerName="ceilometer-notification-agent" Dec 06 06:20:10 crc kubenswrapper[4809]: E1206 06:20:10.683001 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd3f000e-9b7c-4f78-aadb-3a580c1441ff" containerName="ceilometer-central-agent" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.683008 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd3f000e-9b7c-4f78-aadb-3a580c1441ff" containerName="ceilometer-central-agent" Dec 06 06:20:10 crc kubenswrapper[4809]: E1206 06:20:10.683044 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd3f000e-9b7c-4f78-aadb-3a580c1441ff" containerName="proxy-httpd" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.683053 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd3f000e-9b7c-4f78-aadb-3a580c1441ff" containerName="proxy-httpd" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.683388 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd3f000e-9b7c-4f78-aadb-3a580c1441ff" containerName="ceilometer-notification-agent" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.683428 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd3f000e-9b7c-4f78-aadb-3a580c1441ff" containerName="proxy-httpd" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.683445 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd3f000e-9b7c-4f78-aadb-3a580c1441ff" containerName="sg-core" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.683464 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd3f000e-9b7c-4f78-aadb-3a580c1441ff" containerName="ceilometer-central-agent" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.686244 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.689630 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.690497 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.703303 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.807568 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43d072b1-a024-430b-ad1f-4c669e35349c-scripts\") pod \"ceilometer-0\" (UID: \"43d072b1-a024-430b-ad1f-4c669e35349c\") " pod="openstack/ceilometer-0" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.807736 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43d072b1-a024-430b-ad1f-4c669e35349c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"43d072b1-a024-430b-ad1f-4c669e35349c\") " pod="openstack/ceilometer-0" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.807807 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/43d072b1-a024-430b-ad1f-4c669e35349c-run-httpd\") pod \"ceilometer-0\" (UID: \"43d072b1-a024-430b-ad1f-4c669e35349c\") " pod="openstack/ceilometer-0" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.808067 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zc59m\" (UniqueName: \"kubernetes.io/projected/43d072b1-a024-430b-ad1f-4c669e35349c-kube-api-access-zc59m\") pod \"ceilometer-0\" (UID: \"43d072b1-a024-430b-ad1f-4c669e35349c\") " pod="openstack/ceilometer-0" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.808222 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/43d072b1-a024-430b-ad1f-4c669e35349c-log-httpd\") pod \"ceilometer-0\" (UID: \"43d072b1-a024-430b-ad1f-4c669e35349c\") " pod="openstack/ceilometer-0" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.808272 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/43d072b1-a024-430b-ad1f-4c669e35349c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"43d072b1-a024-430b-ad1f-4c669e35349c\") " pod="openstack/ceilometer-0" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.808414 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43d072b1-a024-430b-ad1f-4c669e35349c-config-data\") pod \"ceilometer-0\" (UID: \"43d072b1-a024-430b-ad1f-4c669e35349c\") " pod="openstack/ceilometer-0" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.910730 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43d072b1-a024-430b-ad1f-4c669e35349c-scripts\") pod \"ceilometer-0\" (UID: \"43d072b1-a024-430b-ad1f-4c669e35349c\") " pod="openstack/ceilometer-0" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.910857 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43d072b1-a024-430b-ad1f-4c669e35349c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"43d072b1-a024-430b-ad1f-4c669e35349c\") " pod="openstack/ceilometer-0" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.910909 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/43d072b1-a024-430b-ad1f-4c669e35349c-run-httpd\") pod \"ceilometer-0\" (UID: \"43d072b1-a024-430b-ad1f-4c669e35349c\") " pod="openstack/ceilometer-0" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.911813 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zc59m\" (UniqueName: \"kubernetes.io/projected/43d072b1-a024-430b-ad1f-4c669e35349c-kube-api-access-zc59m\") pod \"ceilometer-0\" (UID: \"43d072b1-a024-430b-ad1f-4c669e35349c\") " pod="openstack/ceilometer-0" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.913830 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/43d072b1-a024-430b-ad1f-4c669e35349c-log-httpd\") pod \"ceilometer-0\" (UID: \"43d072b1-a024-430b-ad1f-4c669e35349c\") " pod="openstack/ceilometer-0" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.913903 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/43d072b1-a024-430b-ad1f-4c669e35349c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"43d072b1-a024-430b-ad1f-4c669e35349c\") " pod="openstack/ceilometer-0" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.914967 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43d072b1-a024-430b-ad1f-4c669e35349c-config-data\") pod \"ceilometer-0\" (UID: \"43d072b1-a024-430b-ad1f-4c669e35349c\") " pod="openstack/ceilometer-0" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.915267 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/43d072b1-a024-430b-ad1f-4c669e35349c-log-httpd\") pod \"ceilometer-0\" (UID: \"43d072b1-a024-430b-ad1f-4c669e35349c\") " pod="openstack/ceilometer-0" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.918414 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/43d072b1-a024-430b-ad1f-4c669e35349c-run-httpd\") pod \"ceilometer-0\" (UID: \"43d072b1-a024-430b-ad1f-4c669e35349c\") " pod="openstack/ceilometer-0" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.921005 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43d072b1-a024-430b-ad1f-4c669e35349c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"43d072b1-a024-430b-ad1f-4c669e35349c\") " pod="openstack/ceilometer-0" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.921080 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/43d072b1-a024-430b-ad1f-4c669e35349c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"43d072b1-a024-430b-ad1f-4c669e35349c\") " pod="openstack/ceilometer-0" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.921158 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43d072b1-a024-430b-ad1f-4c669e35349c-scripts\") pod \"ceilometer-0\" (UID: \"43d072b1-a024-430b-ad1f-4c669e35349c\") " pod="openstack/ceilometer-0" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.925896 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43d072b1-a024-430b-ad1f-4c669e35349c-config-data\") pod \"ceilometer-0\" (UID: \"43d072b1-a024-430b-ad1f-4c669e35349c\") " pod="openstack/ceilometer-0" Dec 06 06:20:10 crc kubenswrapper[4809]: I1206 06:20:10.929454 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zc59m\" (UniqueName: \"kubernetes.io/projected/43d072b1-a024-430b-ad1f-4c669e35349c-kube-api-access-zc59m\") pod \"ceilometer-0\" (UID: \"43d072b1-a024-430b-ad1f-4c669e35349c\") " pod="openstack/ceilometer-0" Dec 06 06:20:11 crc kubenswrapper[4809]: I1206 06:20:11.045891 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:20:11 crc kubenswrapper[4809]: I1206 06:20:11.313899 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:20:11 crc kubenswrapper[4809]: I1206 06:20:11.417105 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd3f000e-9b7c-4f78-aadb-3a580c1441ff" path="/var/lib/kubelet/pods/fd3f000e-9b7c-4f78-aadb-3a580c1441ff/volumes" Dec 06 06:20:11 crc kubenswrapper[4809]: I1206 06:20:11.588269 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:20:11 crc kubenswrapper[4809]: W1206 06:20:11.592779 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod43d072b1_a024_430b_ad1f_4c669e35349c.slice/crio-c9e4c20b7e79cbbf431f007277c31ad0157b5dabfa7b745e6a7cbe2012bddec5 WatchSource:0}: Error finding container c9e4c20b7e79cbbf431f007277c31ad0157b5dabfa7b745e6a7cbe2012bddec5: Status 404 returned error can't find the container with id c9e4c20b7e79cbbf431f007277c31ad0157b5dabfa7b745e6a7cbe2012bddec5 Dec 06 06:20:11 crc kubenswrapper[4809]: I1206 06:20:11.597239 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 06:20:12 crc kubenswrapper[4809]: I1206 06:20:12.355061 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"43d072b1-a024-430b-ad1f-4c669e35349c","Type":"ContainerStarted","Data":"c9e4c20b7e79cbbf431f007277c31ad0157b5dabfa7b745e6a7cbe2012bddec5"} Dec 06 06:20:12 crc kubenswrapper[4809]: I1206 06:20:12.611710 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 06 06:20:12 crc kubenswrapper[4809]: I1206 06:20:12.612100 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 06 06:20:12 crc kubenswrapper[4809]: I1206 06:20:12.666136 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:20:12 crc kubenswrapper[4809]: I1206 06:20:12.704278 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:20:13 crc kubenswrapper[4809]: I1206 06:20:13.088491 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 06:20:13 crc kubenswrapper[4809]: I1206 06:20:13.282189 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdzsr\" (UniqueName: \"kubernetes.io/projected/215a53e4-4135-45b5-912c-e25d50e5c2d3-kube-api-access-sdzsr\") pod \"215a53e4-4135-45b5-912c-e25d50e5c2d3\" (UID: \"215a53e4-4135-45b5-912c-e25d50e5c2d3\") " Dec 06 06:20:13 crc kubenswrapper[4809]: I1206 06:20:13.283223 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/215a53e4-4135-45b5-912c-e25d50e5c2d3-logs\") pod \"215a53e4-4135-45b5-912c-e25d50e5c2d3\" (UID: \"215a53e4-4135-45b5-912c-e25d50e5c2d3\") " Dec 06 06:20:13 crc kubenswrapper[4809]: I1206 06:20:13.283272 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/215a53e4-4135-45b5-912c-e25d50e5c2d3-combined-ca-bundle\") pod \"215a53e4-4135-45b5-912c-e25d50e5c2d3\" (UID: \"215a53e4-4135-45b5-912c-e25d50e5c2d3\") " Dec 06 06:20:13 crc kubenswrapper[4809]: I1206 06:20:13.283319 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/215a53e4-4135-45b5-912c-e25d50e5c2d3-config-data\") pod \"215a53e4-4135-45b5-912c-e25d50e5c2d3\" (UID: \"215a53e4-4135-45b5-912c-e25d50e5c2d3\") " Dec 06 06:20:13 crc kubenswrapper[4809]: I1206 06:20:13.284109 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/215a53e4-4135-45b5-912c-e25d50e5c2d3-logs" (OuterVolumeSpecName: "logs") pod "215a53e4-4135-45b5-912c-e25d50e5c2d3" (UID: "215a53e4-4135-45b5-912c-e25d50e5c2d3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:20:13 crc kubenswrapper[4809]: I1206 06:20:13.286058 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/215a53e4-4135-45b5-912c-e25d50e5c2d3-kube-api-access-sdzsr" (OuterVolumeSpecName: "kube-api-access-sdzsr") pod "215a53e4-4135-45b5-912c-e25d50e5c2d3" (UID: "215a53e4-4135-45b5-912c-e25d50e5c2d3"). InnerVolumeSpecName "kube-api-access-sdzsr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:20:13 crc kubenswrapper[4809]: I1206 06:20:13.332168 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/215a53e4-4135-45b5-912c-e25d50e5c2d3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "215a53e4-4135-45b5-912c-e25d50e5c2d3" (UID: "215a53e4-4135-45b5-912c-e25d50e5c2d3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:20:13 crc kubenswrapper[4809]: I1206 06:20:13.364026 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/215a53e4-4135-45b5-912c-e25d50e5c2d3-config-data" (OuterVolumeSpecName: "config-data") pod "215a53e4-4135-45b5-912c-e25d50e5c2d3" (UID: "215a53e4-4135-45b5-912c-e25d50e5c2d3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:20:13 crc kubenswrapper[4809]: I1206 06:20:13.391394 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdzsr\" (UniqueName: \"kubernetes.io/projected/215a53e4-4135-45b5-912c-e25d50e5c2d3-kube-api-access-sdzsr\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:13 crc kubenswrapper[4809]: I1206 06:20:13.392914 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/215a53e4-4135-45b5-912c-e25d50e5c2d3-logs\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:13 crc kubenswrapper[4809]: I1206 06:20:13.392967 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/215a53e4-4135-45b5-912c-e25d50e5c2d3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:13 crc kubenswrapper[4809]: I1206 06:20:13.392993 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/215a53e4-4135-45b5-912c-e25d50e5c2d3-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:13 crc kubenswrapper[4809]: I1206 06:20:13.432524 4809 generic.go:334] "Generic (PLEG): container finished" podID="215a53e4-4135-45b5-912c-e25d50e5c2d3" containerID="87401fb28e1de011b3a112da08d30919cd13ce434f4e76261389e19eb091b1b8" exitCode=0 Dec 06 06:20:13 crc kubenswrapper[4809]: I1206 06:20:13.433103 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 06:20:13 crc kubenswrapper[4809]: I1206 06:20:13.452090 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"43d072b1-a024-430b-ad1f-4c669e35349c","Type":"ContainerStarted","Data":"fa43abeec7d8474fb8e3bb23fb13aa511eb4a0eca9ed5617eb0c302009e3fe85"} Dec 06 06:20:13 crc kubenswrapper[4809]: I1206 06:20:13.452130 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"215a53e4-4135-45b5-912c-e25d50e5c2d3","Type":"ContainerDied","Data":"87401fb28e1de011b3a112da08d30919cd13ce434f4e76261389e19eb091b1b8"} Dec 06 06:20:13 crc kubenswrapper[4809]: I1206 06:20:13.452156 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"215a53e4-4135-45b5-912c-e25d50e5c2d3","Type":"ContainerDied","Data":"4233a597f87bf500704b3c0e4974db08db0b180d93dd0e63c40543e7ecd4dbaf"} Dec 06 06:20:13 crc kubenswrapper[4809]: I1206 06:20:13.452202 4809 scope.go:117] "RemoveContainer" containerID="87401fb28e1de011b3a112da08d30919cd13ce434f4e76261389e19eb091b1b8" Dec 06 06:20:13 crc kubenswrapper[4809]: I1206 06:20:13.480472 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:20:13 crc kubenswrapper[4809]: I1206 06:20:13.599655 4809 scope.go:117] "RemoveContainer" containerID="38942159c5dfe1eeb521ee0e2e56ac8d200f1be759cd9896b948b276d2094f2e" Dec 06 06:20:13 crc kubenswrapper[4809]: I1206 06:20:13.630379 4809 scope.go:117] "RemoveContainer" containerID="87401fb28e1de011b3a112da08d30919cd13ce434f4e76261389e19eb091b1b8" Dec 06 06:20:13 crc kubenswrapper[4809]: I1206 06:20:13.636178 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f17fe211-3570-4b69-866c-9607c6489341" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.248:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 06:20:13 crc kubenswrapper[4809]: I1206 06:20:13.636301 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f17fe211-3570-4b69-866c-9607c6489341" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.248:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 06:20:13 crc kubenswrapper[4809]: E1206 06:20:13.638534 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87401fb28e1de011b3a112da08d30919cd13ce434f4e76261389e19eb091b1b8\": container with ID starting with 87401fb28e1de011b3a112da08d30919cd13ce434f4e76261389e19eb091b1b8 not found: ID does not exist" containerID="87401fb28e1de011b3a112da08d30919cd13ce434f4e76261389e19eb091b1b8" Dec 06 06:20:13 crc kubenswrapper[4809]: I1206 06:20:13.638592 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87401fb28e1de011b3a112da08d30919cd13ce434f4e76261389e19eb091b1b8"} err="failed to get container status \"87401fb28e1de011b3a112da08d30919cd13ce434f4e76261389e19eb091b1b8\": rpc error: code = NotFound desc = could not find container \"87401fb28e1de011b3a112da08d30919cd13ce434f4e76261389e19eb091b1b8\": container with ID starting with 87401fb28e1de011b3a112da08d30919cd13ce434f4e76261389e19eb091b1b8 not found: ID does not exist" Dec 06 06:20:13 crc kubenswrapper[4809]: I1206 06:20:13.638622 4809 scope.go:117] "RemoveContainer" containerID="38942159c5dfe1eeb521ee0e2e56ac8d200f1be759cd9896b948b276d2094f2e" Dec 06 06:20:13 crc kubenswrapper[4809]: E1206 06:20:13.639129 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38942159c5dfe1eeb521ee0e2e56ac8d200f1be759cd9896b948b276d2094f2e\": container with ID starting with 38942159c5dfe1eeb521ee0e2e56ac8d200f1be759cd9896b948b276d2094f2e not found: ID does not exist" containerID="38942159c5dfe1eeb521ee0e2e56ac8d200f1be759cd9896b948b276d2094f2e" Dec 06 06:20:13 crc kubenswrapper[4809]: I1206 06:20:13.639228 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38942159c5dfe1eeb521ee0e2e56ac8d200f1be759cd9896b948b276d2094f2e"} err="failed to get container status \"38942159c5dfe1eeb521ee0e2e56ac8d200f1be759cd9896b948b276d2094f2e\": rpc error: code = NotFound desc = could not find container \"38942159c5dfe1eeb521ee0e2e56ac8d200f1be759cd9896b948b276d2094f2e\": container with ID starting with 38942159c5dfe1eeb521ee0e2e56ac8d200f1be759cd9896b948b276d2094f2e not found: ID does not exist" Dec 06 06:20:13 crc kubenswrapper[4809]: I1206 06:20:13.776330 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-tp9tt"] Dec 06 06:20:13 crc kubenswrapper[4809]: E1206 06:20:13.777322 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="215a53e4-4135-45b5-912c-e25d50e5c2d3" containerName="nova-api-log" Dec 06 06:20:13 crc kubenswrapper[4809]: I1206 06:20:13.777343 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="215a53e4-4135-45b5-912c-e25d50e5c2d3" containerName="nova-api-log" Dec 06 06:20:13 crc kubenswrapper[4809]: E1206 06:20:13.777378 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="215a53e4-4135-45b5-912c-e25d50e5c2d3" containerName="nova-api-api" Dec 06 06:20:13 crc kubenswrapper[4809]: I1206 06:20:13.777385 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="215a53e4-4135-45b5-912c-e25d50e5c2d3" containerName="nova-api-api" Dec 06 06:20:13 crc kubenswrapper[4809]: I1206 06:20:13.777705 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="215a53e4-4135-45b5-912c-e25d50e5c2d3" containerName="nova-api-log" Dec 06 06:20:13 crc kubenswrapper[4809]: I1206 06:20:13.777761 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="215a53e4-4135-45b5-912c-e25d50e5c2d3" containerName="nova-api-api" Dec 06 06:20:13 crc kubenswrapper[4809]: I1206 06:20:13.778951 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-tp9tt" Dec 06 06:20:13 crc kubenswrapper[4809]: I1206 06:20:13.784273 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 06 06:20:13 crc kubenswrapper[4809]: I1206 06:20:13.784410 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 06 06:20:13 crc kubenswrapper[4809]: I1206 06:20:13.791766 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-tp9tt"] Dec 06 06:20:13 crc kubenswrapper[4809]: I1206 06:20:13.932658 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e106bab-59d8-4d2e-89f9-1f46e9b37f43-scripts\") pod \"nova-cell1-cell-mapping-tp9tt\" (UID: \"4e106bab-59d8-4d2e-89f9-1f46e9b37f43\") " pod="openstack/nova-cell1-cell-mapping-tp9tt" Dec 06 06:20:13 crc kubenswrapper[4809]: I1206 06:20:13.932853 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e106bab-59d8-4d2e-89f9-1f46e9b37f43-config-data\") pod \"nova-cell1-cell-mapping-tp9tt\" (UID: \"4e106bab-59d8-4d2e-89f9-1f46e9b37f43\") " pod="openstack/nova-cell1-cell-mapping-tp9tt" Dec 06 06:20:13 crc kubenswrapper[4809]: I1206 06:20:13.932913 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e106bab-59d8-4d2e-89f9-1f46e9b37f43-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-tp9tt\" (UID: \"4e106bab-59d8-4d2e-89f9-1f46e9b37f43\") " pod="openstack/nova-cell1-cell-mapping-tp9tt" Dec 06 06:20:13 crc kubenswrapper[4809]: I1206 06:20:13.933686 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h797q\" (UniqueName: \"kubernetes.io/projected/4e106bab-59d8-4d2e-89f9-1f46e9b37f43-kube-api-access-h797q\") pod \"nova-cell1-cell-mapping-tp9tt\" (UID: \"4e106bab-59d8-4d2e-89f9-1f46e9b37f43\") " pod="openstack/nova-cell1-cell-mapping-tp9tt" Dec 06 06:20:14 crc kubenswrapper[4809]: I1206 06:20:14.036199 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e106bab-59d8-4d2e-89f9-1f46e9b37f43-config-data\") pod \"nova-cell1-cell-mapping-tp9tt\" (UID: \"4e106bab-59d8-4d2e-89f9-1f46e9b37f43\") " pod="openstack/nova-cell1-cell-mapping-tp9tt" Dec 06 06:20:14 crc kubenswrapper[4809]: I1206 06:20:14.036247 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e106bab-59d8-4d2e-89f9-1f46e9b37f43-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-tp9tt\" (UID: \"4e106bab-59d8-4d2e-89f9-1f46e9b37f43\") " pod="openstack/nova-cell1-cell-mapping-tp9tt" Dec 06 06:20:14 crc kubenswrapper[4809]: I1206 06:20:14.036488 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h797q\" (UniqueName: \"kubernetes.io/projected/4e106bab-59d8-4d2e-89f9-1f46e9b37f43-kube-api-access-h797q\") pod \"nova-cell1-cell-mapping-tp9tt\" (UID: \"4e106bab-59d8-4d2e-89f9-1f46e9b37f43\") " pod="openstack/nova-cell1-cell-mapping-tp9tt" Dec 06 06:20:14 crc kubenswrapper[4809]: I1206 06:20:14.036561 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e106bab-59d8-4d2e-89f9-1f46e9b37f43-scripts\") pod \"nova-cell1-cell-mapping-tp9tt\" (UID: \"4e106bab-59d8-4d2e-89f9-1f46e9b37f43\") " pod="openstack/nova-cell1-cell-mapping-tp9tt" Dec 06 06:20:14 crc kubenswrapper[4809]: I1206 06:20:14.040634 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e106bab-59d8-4d2e-89f9-1f46e9b37f43-config-data\") pod \"nova-cell1-cell-mapping-tp9tt\" (UID: \"4e106bab-59d8-4d2e-89f9-1f46e9b37f43\") " pod="openstack/nova-cell1-cell-mapping-tp9tt" Dec 06 06:20:14 crc kubenswrapper[4809]: I1206 06:20:14.044419 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e106bab-59d8-4d2e-89f9-1f46e9b37f43-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-tp9tt\" (UID: \"4e106bab-59d8-4d2e-89f9-1f46e9b37f43\") " pod="openstack/nova-cell1-cell-mapping-tp9tt" Dec 06 06:20:14 crc kubenswrapper[4809]: I1206 06:20:14.044835 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e106bab-59d8-4d2e-89f9-1f46e9b37f43-scripts\") pod \"nova-cell1-cell-mapping-tp9tt\" (UID: \"4e106bab-59d8-4d2e-89f9-1f46e9b37f43\") " pod="openstack/nova-cell1-cell-mapping-tp9tt" Dec 06 06:20:14 crc kubenswrapper[4809]: I1206 06:20:14.055828 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h797q\" (UniqueName: \"kubernetes.io/projected/4e106bab-59d8-4d2e-89f9-1f46e9b37f43-kube-api-access-h797q\") pod \"nova-cell1-cell-mapping-tp9tt\" (UID: \"4e106bab-59d8-4d2e-89f9-1f46e9b37f43\") " pod="openstack/nova-cell1-cell-mapping-tp9tt" Dec 06 06:20:14 crc kubenswrapper[4809]: I1206 06:20:14.121185 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-tp9tt" Dec 06 06:20:14 crc kubenswrapper[4809]: I1206 06:20:14.473676 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"43d072b1-a024-430b-ad1f-4c669e35349c","Type":"ContainerStarted","Data":"b32efe19327ba089385c3fe5c6dc3f5d761d142991d09b630cdd902293d67fa2"} Dec 06 06:20:14 crc kubenswrapper[4809]: I1206 06:20:14.473713 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"43d072b1-a024-430b-ad1f-4c669e35349c","Type":"ContainerStarted","Data":"2265a4be89a538489e7ad0f7ad150ccb03be06746a68da7229d145e50868a0b2"} Dec 06 06:20:14 crc kubenswrapper[4809]: I1206 06:20:14.707837 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-tp9tt"] Dec 06 06:20:14 crc kubenswrapper[4809]: W1206 06:20:14.716482 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4e106bab_59d8_4d2e_89f9_1f46e9b37f43.slice/crio-49caf496c4a72e06508b2c1db28e820afd3acf01b5a41876151f9c1e40491321 WatchSource:0}: Error finding container 49caf496c4a72e06508b2c1db28e820afd3acf01b5a41876151f9c1e40491321: Status 404 returned error can't find the container with id 49caf496c4a72e06508b2c1db28e820afd3acf01b5a41876151f9c1e40491321 Dec 06 06:20:15 crc kubenswrapper[4809]: I1206 06:20:15.492845 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-tp9tt" event={"ID":"4e106bab-59d8-4d2e-89f9-1f46e9b37f43","Type":"ContainerStarted","Data":"ee227d98802b9bb43d7d6c3599870ec807f20cfa515b0f11e3a0a8d474c3cb99"} Dec 06 06:20:15 crc kubenswrapper[4809]: I1206 06:20:15.493578 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-tp9tt" event={"ID":"4e106bab-59d8-4d2e-89f9-1f46e9b37f43","Type":"ContainerStarted","Data":"49caf496c4a72e06508b2c1db28e820afd3acf01b5a41876151f9c1e40491321"} Dec 06 06:20:15 crc kubenswrapper[4809]: I1206 06:20:15.536966 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-tp9tt" podStartSLOduration=2.5369127860000003 podStartE2EDuration="2.536912786s" podCreationTimestamp="2025-12-06 06:20:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:20:15.511881121 +0000 UTC m=+1740.400864053" watchObservedRunningTime="2025-12-06 06:20:15.536912786 +0000 UTC m=+1740.425895728" Dec 06 06:20:16 crc kubenswrapper[4809]: I1206 06:20:16.507352 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"43d072b1-a024-430b-ad1f-4c669e35349c","Type":"ContainerStarted","Data":"23b27c62d969be5b621f105b93f7f5846923671a71afda236f53559ab6a8a31a"} Dec 06 06:20:16 crc kubenswrapper[4809]: I1206 06:20:16.507562 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="43d072b1-a024-430b-ad1f-4c669e35349c" containerName="ceilometer-central-agent" containerID="cri-o://fa43abeec7d8474fb8e3bb23fb13aa511eb4a0eca9ed5617eb0c302009e3fe85" gracePeriod=30 Dec 06 06:20:16 crc kubenswrapper[4809]: I1206 06:20:16.507632 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="43d072b1-a024-430b-ad1f-4c669e35349c" containerName="proxy-httpd" containerID="cri-o://23b27c62d969be5b621f105b93f7f5846923671a71afda236f53559ab6a8a31a" gracePeriod=30 Dec 06 06:20:16 crc kubenswrapper[4809]: I1206 06:20:16.507620 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="43d072b1-a024-430b-ad1f-4c669e35349c" containerName="sg-core" containerID="cri-o://b32efe19327ba089385c3fe5c6dc3f5d761d142991d09b630cdd902293d67fa2" gracePeriod=30 Dec 06 06:20:16 crc kubenswrapper[4809]: I1206 06:20:16.509490 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="43d072b1-a024-430b-ad1f-4c669e35349c" containerName="ceilometer-notification-agent" containerID="cri-o://2265a4be89a538489e7ad0f7ad150ccb03be06746a68da7229d145e50868a0b2" gracePeriod=30 Dec 06 06:20:16 crc kubenswrapper[4809]: I1206 06:20:16.543146 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.724842648 podStartE2EDuration="6.543126183s" podCreationTimestamp="2025-12-06 06:20:10 +0000 UTC" firstStartedPulling="2025-12-06 06:20:11.596860946 +0000 UTC m=+1736.485843888" lastFinishedPulling="2025-12-06 06:20:15.415144481 +0000 UTC m=+1740.304127423" observedRunningTime="2025-12-06 06:20:16.528027376 +0000 UTC m=+1741.417010318" watchObservedRunningTime="2025-12-06 06:20:16.543126183 +0000 UTC m=+1741.432109125" Dec 06 06:20:16 crc kubenswrapper[4809]: I1206 06:20:16.823917 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6b7bbf7cf9-2kbmw" Dec 06 06:20:16 crc kubenswrapper[4809]: I1206 06:20:16.907519 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-fg9jq"] Dec 06 06:20:16 crc kubenswrapper[4809]: I1206 06:20:16.907820 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-9b86998b5-fg9jq" podUID="37c89662-6b35-4e73-bf37-f6bb01978c5f" containerName="dnsmasq-dns" containerID="cri-o://54bc7c704b1b4e133a8e1806521b4431560c2ad2f82669b44524c18ffb5d443c" gracePeriod=10 Dec 06 06:20:17 crc kubenswrapper[4809]: I1206 06:20:17.517383 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-fg9jq" Dec 06 06:20:17 crc kubenswrapper[4809]: I1206 06:20:17.520193 4809 generic.go:334] "Generic (PLEG): container finished" podID="43d072b1-a024-430b-ad1f-4c669e35349c" containerID="23b27c62d969be5b621f105b93f7f5846923671a71afda236f53559ab6a8a31a" exitCode=0 Dec 06 06:20:17 crc kubenswrapper[4809]: I1206 06:20:17.520248 4809 generic.go:334] "Generic (PLEG): container finished" podID="43d072b1-a024-430b-ad1f-4c669e35349c" containerID="b32efe19327ba089385c3fe5c6dc3f5d761d142991d09b630cdd902293d67fa2" exitCode=2 Dec 06 06:20:17 crc kubenswrapper[4809]: I1206 06:20:17.520259 4809 generic.go:334] "Generic (PLEG): container finished" podID="43d072b1-a024-430b-ad1f-4c669e35349c" containerID="2265a4be89a538489e7ad0f7ad150ccb03be06746a68da7229d145e50868a0b2" exitCode=0 Dec 06 06:20:17 crc kubenswrapper[4809]: I1206 06:20:17.520200 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"43d072b1-a024-430b-ad1f-4c669e35349c","Type":"ContainerDied","Data":"23b27c62d969be5b621f105b93f7f5846923671a71afda236f53559ab6a8a31a"} Dec 06 06:20:17 crc kubenswrapper[4809]: I1206 06:20:17.520332 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"43d072b1-a024-430b-ad1f-4c669e35349c","Type":"ContainerDied","Data":"b32efe19327ba089385c3fe5c6dc3f5d761d142991d09b630cdd902293d67fa2"} Dec 06 06:20:17 crc kubenswrapper[4809]: I1206 06:20:17.520395 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"43d072b1-a024-430b-ad1f-4c669e35349c","Type":"ContainerDied","Data":"2265a4be89a538489e7ad0f7ad150ccb03be06746a68da7229d145e50868a0b2"} Dec 06 06:20:17 crc kubenswrapper[4809]: I1206 06:20:17.523084 4809 generic.go:334] "Generic (PLEG): container finished" podID="37c89662-6b35-4e73-bf37-f6bb01978c5f" containerID="54bc7c704b1b4e133a8e1806521b4431560c2ad2f82669b44524c18ffb5d443c" exitCode=0 Dec 06 06:20:17 crc kubenswrapper[4809]: I1206 06:20:17.523225 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-fg9jq" event={"ID":"37c89662-6b35-4e73-bf37-f6bb01978c5f","Type":"ContainerDied","Data":"54bc7c704b1b4e133a8e1806521b4431560c2ad2f82669b44524c18ffb5d443c"} Dec 06 06:20:17 crc kubenswrapper[4809]: I1206 06:20:17.523252 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-fg9jq" Dec 06 06:20:17 crc kubenswrapper[4809]: I1206 06:20:17.523263 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-fg9jq" event={"ID":"37c89662-6b35-4e73-bf37-f6bb01978c5f","Type":"ContainerDied","Data":"a3b2aeabf80e82647cf1efb7c757498fecab574b5cc8fa45f0d14668ecea2846"} Dec 06 06:20:17 crc kubenswrapper[4809]: I1206 06:20:17.523283 4809 scope.go:117] "RemoveContainer" containerID="54bc7c704b1b4e133a8e1806521b4431560c2ad2f82669b44524c18ffb5d443c" Dec 06 06:20:17 crc kubenswrapper[4809]: I1206 06:20:17.554536 4809 scope.go:117] "RemoveContainer" containerID="4608c06dd7d29f002c9151795ccc44dba1da393d792ee6eb2cdbba4b65ea7485" Dec 06 06:20:17 crc kubenswrapper[4809]: I1206 06:20:17.584531 4809 scope.go:117] "RemoveContainer" containerID="54bc7c704b1b4e133a8e1806521b4431560c2ad2f82669b44524c18ffb5d443c" Dec 06 06:20:17 crc kubenswrapper[4809]: E1206 06:20:17.585815 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54bc7c704b1b4e133a8e1806521b4431560c2ad2f82669b44524c18ffb5d443c\": container with ID starting with 54bc7c704b1b4e133a8e1806521b4431560c2ad2f82669b44524c18ffb5d443c not found: ID does not exist" containerID="54bc7c704b1b4e133a8e1806521b4431560c2ad2f82669b44524c18ffb5d443c" Dec 06 06:20:17 crc kubenswrapper[4809]: I1206 06:20:17.585957 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54bc7c704b1b4e133a8e1806521b4431560c2ad2f82669b44524c18ffb5d443c"} err="failed to get container status \"54bc7c704b1b4e133a8e1806521b4431560c2ad2f82669b44524c18ffb5d443c\": rpc error: code = NotFound desc = could not find container \"54bc7c704b1b4e133a8e1806521b4431560c2ad2f82669b44524c18ffb5d443c\": container with ID starting with 54bc7c704b1b4e133a8e1806521b4431560c2ad2f82669b44524c18ffb5d443c not found: ID does not exist" Dec 06 06:20:17 crc kubenswrapper[4809]: I1206 06:20:17.586037 4809 scope.go:117] "RemoveContainer" containerID="4608c06dd7d29f002c9151795ccc44dba1da393d792ee6eb2cdbba4b65ea7485" Dec 06 06:20:17 crc kubenswrapper[4809]: E1206 06:20:17.586432 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4608c06dd7d29f002c9151795ccc44dba1da393d792ee6eb2cdbba4b65ea7485\": container with ID starting with 4608c06dd7d29f002c9151795ccc44dba1da393d792ee6eb2cdbba4b65ea7485 not found: ID does not exist" containerID="4608c06dd7d29f002c9151795ccc44dba1da393d792ee6eb2cdbba4b65ea7485" Dec 06 06:20:17 crc kubenswrapper[4809]: I1206 06:20:17.586475 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4608c06dd7d29f002c9151795ccc44dba1da393d792ee6eb2cdbba4b65ea7485"} err="failed to get container status \"4608c06dd7d29f002c9151795ccc44dba1da393d792ee6eb2cdbba4b65ea7485\": rpc error: code = NotFound desc = could not find container \"4608c06dd7d29f002c9151795ccc44dba1da393d792ee6eb2cdbba4b65ea7485\": container with ID starting with 4608c06dd7d29f002c9151795ccc44dba1da393d792ee6eb2cdbba4b65ea7485 not found: ID does not exist" Dec 06 06:20:17 crc kubenswrapper[4809]: I1206 06:20:17.635993 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37c89662-6b35-4e73-bf37-f6bb01978c5f-dns-svc\") pod \"37c89662-6b35-4e73-bf37-f6bb01978c5f\" (UID: \"37c89662-6b35-4e73-bf37-f6bb01978c5f\") " Dec 06 06:20:17 crc kubenswrapper[4809]: I1206 06:20:17.636205 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37c89662-6b35-4e73-bf37-f6bb01978c5f-ovsdbserver-nb\") pod \"37c89662-6b35-4e73-bf37-f6bb01978c5f\" (UID: \"37c89662-6b35-4e73-bf37-f6bb01978c5f\") " Dec 06 06:20:17 crc kubenswrapper[4809]: I1206 06:20:17.636298 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37c89662-6b35-4e73-bf37-f6bb01978c5f-config\") pod \"37c89662-6b35-4e73-bf37-f6bb01978c5f\" (UID: \"37c89662-6b35-4e73-bf37-f6bb01978c5f\") " Dec 06 06:20:17 crc kubenswrapper[4809]: I1206 06:20:17.636334 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27mh5\" (UniqueName: \"kubernetes.io/projected/37c89662-6b35-4e73-bf37-f6bb01978c5f-kube-api-access-27mh5\") pod \"37c89662-6b35-4e73-bf37-f6bb01978c5f\" (UID: \"37c89662-6b35-4e73-bf37-f6bb01978c5f\") " Dec 06 06:20:17 crc kubenswrapper[4809]: I1206 06:20:17.636465 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/37c89662-6b35-4e73-bf37-f6bb01978c5f-dns-swift-storage-0\") pod \"37c89662-6b35-4e73-bf37-f6bb01978c5f\" (UID: \"37c89662-6b35-4e73-bf37-f6bb01978c5f\") " Dec 06 06:20:17 crc kubenswrapper[4809]: I1206 06:20:17.636495 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37c89662-6b35-4e73-bf37-f6bb01978c5f-ovsdbserver-sb\") pod \"37c89662-6b35-4e73-bf37-f6bb01978c5f\" (UID: \"37c89662-6b35-4e73-bf37-f6bb01978c5f\") " Dec 06 06:20:17 crc kubenswrapper[4809]: I1206 06:20:17.672240 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37c89662-6b35-4e73-bf37-f6bb01978c5f-kube-api-access-27mh5" (OuterVolumeSpecName: "kube-api-access-27mh5") pod "37c89662-6b35-4e73-bf37-f6bb01978c5f" (UID: "37c89662-6b35-4e73-bf37-f6bb01978c5f"). InnerVolumeSpecName "kube-api-access-27mh5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:20:17 crc kubenswrapper[4809]: I1206 06:20:17.709305 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37c89662-6b35-4e73-bf37-f6bb01978c5f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "37c89662-6b35-4e73-bf37-f6bb01978c5f" (UID: "37c89662-6b35-4e73-bf37-f6bb01978c5f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:20:17 crc kubenswrapper[4809]: I1206 06:20:17.735248 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37c89662-6b35-4e73-bf37-f6bb01978c5f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "37c89662-6b35-4e73-bf37-f6bb01978c5f" (UID: "37c89662-6b35-4e73-bf37-f6bb01978c5f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:20:17 crc kubenswrapper[4809]: I1206 06:20:17.740184 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37c89662-6b35-4e73-bf37-f6bb01978c5f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:17 crc kubenswrapper[4809]: I1206 06:20:17.740389 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27mh5\" (UniqueName: \"kubernetes.io/projected/37c89662-6b35-4e73-bf37-f6bb01978c5f-kube-api-access-27mh5\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:17 crc kubenswrapper[4809]: I1206 06:20:17.740636 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37c89662-6b35-4e73-bf37-f6bb01978c5f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:17 crc kubenswrapper[4809]: I1206 06:20:17.741551 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37c89662-6b35-4e73-bf37-f6bb01978c5f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "37c89662-6b35-4e73-bf37-f6bb01978c5f" (UID: "37c89662-6b35-4e73-bf37-f6bb01978c5f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:20:17 crc kubenswrapper[4809]: I1206 06:20:17.753863 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37c89662-6b35-4e73-bf37-f6bb01978c5f-config" (OuterVolumeSpecName: "config") pod "37c89662-6b35-4e73-bf37-f6bb01978c5f" (UID: "37c89662-6b35-4e73-bf37-f6bb01978c5f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:20:17 crc kubenswrapper[4809]: I1206 06:20:17.780170 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37c89662-6b35-4e73-bf37-f6bb01978c5f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "37c89662-6b35-4e73-bf37-f6bb01978c5f" (UID: "37c89662-6b35-4e73-bf37-f6bb01978c5f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:20:17 crc kubenswrapper[4809]: I1206 06:20:17.843359 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37c89662-6b35-4e73-bf37-f6bb01978c5f-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:17 crc kubenswrapper[4809]: I1206 06:20:17.843406 4809 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/37c89662-6b35-4e73-bf37-f6bb01978c5f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:17 crc kubenswrapper[4809]: I1206 06:20:17.843417 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37c89662-6b35-4e73-bf37-f6bb01978c5f-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:17 crc kubenswrapper[4809]: I1206 06:20:17.887208 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-fg9jq"] Dec 06 06:20:17 crc kubenswrapper[4809]: I1206 06:20:17.900700 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-fg9jq"] Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.303227 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.380893 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43d072b1-a024-430b-ad1f-4c669e35349c-scripts\") pod \"43d072b1-a024-430b-ad1f-4c669e35349c\" (UID: \"43d072b1-a024-430b-ad1f-4c669e35349c\") " Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.381055 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/43d072b1-a024-430b-ad1f-4c669e35349c-run-httpd\") pod \"43d072b1-a024-430b-ad1f-4c669e35349c\" (UID: \"43d072b1-a024-430b-ad1f-4c669e35349c\") " Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.381170 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/43d072b1-a024-430b-ad1f-4c669e35349c-log-httpd\") pod \"43d072b1-a024-430b-ad1f-4c669e35349c\" (UID: \"43d072b1-a024-430b-ad1f-4c669e35349c\") " Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.381216 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43d072b1-a024-430b-ad1f-4c669e35349c-combined-ca-bundle\") pod \"43d072b1-a024-430b-ad1f-4c669e35349c\" (UID: \"43d072b1-a024-430b-ad1f-4c669e35349c\") " Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.381278 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/43d072b1-a024-430b-ad1f-4c669e35349c-sg-core-conf-yaml\") pod \"43d072b1-a024-430b-ad1f-4c669e35349c\" (UID: \"43d072b1-a024-430b-ad1f-4c669e35349c\") " Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.381320 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zc59m\" (UniqueName: \"kubernetes.io/projected/43d072b1-a024-430b-ad1f-4c669e35349c-kube-api-access-zc59m\") pod \"43d072b1-a024-430b-ad1f-4c669e35349c\" (UID: \"43d072b1-a024-430b-ad1f-4c669e35349c\") " Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.381411 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43d072b1-a024-430b-ad1f-4c669e35349c-config-data\") pod \"43d072b1-a024-430b-ad1f-4c669e35349c\" (UID: \"43d072b1-a024-430b-ad1f-4c669e35349c\") " Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.381467 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43d072b1-a024-430b-ad1f-4c669e35349c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "43d072b1-a024-430b-ad1f-4c669e35349c" (UID: "43d072b1-a024-430b-ad1f-4c669e35349c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.382278 4809 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/43d072b1-a024-430b-ad1f-4c669e35349c-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.382925 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43d072b1-a024-430b-ad1f-4c669e35349c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "43d072b1-a024-430b-ad1f-4c669e35349c" (UID: "43d072b1-a024-430b-ad1f-4c669e35349c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.391293 4809 scope.go:117] "RemoveContainer" containerID="9bc2dc88b3b9f309ce0169aa510326dd77022dece4c69cab3141df00fe5c0106" Dec 06 06:20:19 crc kubenswrapper[4809]: E1206 06:20:19.391635 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.408264 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43d072b1-a024-430b-ad1f-4c669e35349c-kube-api-access-zc59m" (OuterVolumeSpecName: "kube-api-access-zc59m") pod "43d072b1-a024-430b-ad1f-4c669e35349c" (UID: "43d072b1-a024-430b-ad1f-4c669e35349c"). InnerVolumeSpecName "kube-api-access-zc59m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.408363 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43d072b1-a024-430b-ad1f-4c669e35349c-scripts" (OuterVolumeSpecName: "scripts") pod "43d072b1-a024-430b-ad1f-4c669e35349c" (UID: "43d072b1-a024-430b-ad1f-4c669e35349c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.439794 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37c89662-6b35-4e73-bf37-f6bb01978c5f" path="/var/lib/kubelet/pods/37c89662-6b35-4e73-bf37-f6bb01978c5f/volumes" Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.475174 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43d072b1-a024-430b-ad1f-4c669e35349c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "43d072b1-a024-430b-ad1f-4c669e35349c" (UID: "43d072b1-a024-430b-ad1f-4c669e35349c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.487916 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zc59m\" (UniqueName: \"kubernetes.io/projected/43d072b1-a024-430b-ad1f-4c669e35349c-kube-api-access-zc59m\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.487979 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43d072b1-a024-430b-ad1f-4c669e35349c-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.487992 4809 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/43d072b1-a024-430b-ad1f-4c669e35349c-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.488005 4809 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/43d072b1-a024-430b-ad1f-4c669e35349c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.529120 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43d072b1-a024-430b-ad1f-4c669e35349c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "43d072b1-a024-430b-ad1f-4c669e35349c" (UID: "43d072b1-a024-430b-ad1f-4c669e35349c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.555501 4809 generic.go:334] "Generic (PLEG): container finished" podID="43d072b1-a024-430b-ad1f-4c669e35349c" containerID="fa43abeec7d8474fb8e3bb23fb13aa511eb4a0eca9ed5617eb0c302009e3fe85" exitCode=0 Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.555559 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"43d072b1-a024-430b-ad1f-4c669e35349c","Type":"ContainerDied","Data":"fa43abeec7d8474fb8e3bb23fb13aa511eb4a0eca9ed5617eb0c302009e3fe85"} Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.555596 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"43d072b1-a024-430b-ad1f-4c669e35349c","Type":"ContainerDied","Data":"c9e4c20b7e79cbbf431f007277c31ad0157b5dabfa7b745e6a7cbe2012bddec5"} Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.555619 4809 scope.go:117] "RemoveContainer" containerID="23b27c62d969be5b621f105b93f7f5846923671a71afda236f53559ab6a8a31a" Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.555617 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.589611 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43d072b1-a024-430b-ad1f-4c669e35349c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.592137 4809 scope.go:117] "RemoveContainer" containerID="b32efe19327ba089385c3fe5c6dc3f5d761d142991d09b630cdd902293d67fa2" Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.596036 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43d072b1-a024-430b-ad1f-4c669e35349c-config-data" (OuterVolumeSpecName: "config-data") pod "43d072b1-a024-430b-ad1f-4c669e35349c" (UID: "43d072b1-a024-430b-ad1f-4c669e35349c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.611139 4809 scope.go:117] "RemoveContainer" containerID="2265a4be89a538489e7ad0f7ad150ccb03be06746a68da7229d145e50868a0b2" Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.631604 4809 scope.go:117] "RemoveContainer" containerID="fa43abeec7d8474fb8e3bb23fb13aa511eb4a0eca9ed5617eb0c302009e3fe85" Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.653464 4809 scope.go:117] "RemoveContainer" containerID="23b27c62d969be5b621f105b93f7f5846923671a71afda236f53559ab6a8a31a" Dec 06 06:20:19 crc kubenswrapper[4809]: E1206 06:20:19.653857 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23b27c62d969be5b621f105b93f7f5846923671a71afda236f53559ab6a8a31a\": container with ID starting with 23b27c62d969be5b621f105b93f7f5846923671a71afda236f53559ab6a8a31a not found: ID does not exist" containerID="23b27c62d969be5b621f105b93f7f5846923671a71afda236f53559ab6a8a31a" Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.653922 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23b27c62d969be5b621f105b93f7f5846923671a71afda236f53559ab6a8a31a"} err="failed to get container status \"23b27c62d969be5b621f105b93f7f5846923671a71afda236f53559ab6a8a31a\": rpc error: code = NotFound desc = could not find container \"23b27c62d969be5b621f105b93f7f5846923671a71afda236f53559ab6a8a31a\": container with ID starting with 23b27c62d969be5b621f105b93f7f5846923671a71afda236f53559ab6a8a31a not found: ID does not exist" Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.653996 4809 scope.go:117] "RemoveContainer" containerID="b32efe19327ba089385c3fe5c6dc3f5d761d142991d09b630cdd902293d67fa2" Dec 06 06:20:19 crc kubenswrapper[4809]: E1206 06:20:19.654612 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b32efe19327ba089385c3fe5c6dc3f5d761d142991d09b630cdd902293d67fa2\": container with ID starting with b32efe19327ba089385c3fe5c6dc3f5d761d142991d09b630cdd902293d67fa2 not found: ID does not exist" containerID="b32efe19327ba089385c3fe5c6dc3f5d761d142991d09b630cdd902293d67fa2" Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.654668 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b32efe19327ba089385c3fe5c6dc3f5d761d142991d09b630cdd902293d67fa2"} err="failed to get container status \"b32efe19327ba089385c3fe5c6dc3f5d761d142991d09b630cdd902293d67fa2\": rpc error: code = NotFound desc = could not find container \"b32efe19327ba089385c3fe5c6dc3f5d761d142991d09b630cdd902293d67fa2\": container with ID starting with b32efe19327ba089385c3fe5c6dc3f5d761d142991d09b630cdd902293d67fa2 not found: ID does not exist" Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.654704 4809 scope.go:117] "RemoveContainer" containerID="2265a4be89a538489e7ad0f7ad150ccb03be06746a68da7229d145e50868a0b2" Dec 06 06:20:19 crc kubenswrapper[4809]: E1206 06:20:19.655130 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2265a4be89a538489e7ad0f7ad150ccb03be06746a68da7229d145e50868a0b2\": container with ID starting with 2265a4be89a538489e7ad0f7ad150ccb03be06746a68da7229d145e50868a0b2 not found: ID does not exist" containerID="2265a4be89a538489e7ad0f7ad150ccb03be06746a68da7229d145e50868a0b2" Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.655176 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2265a4be89a538489e7ad0f7ad150ccb03be06746a68da7229d145e50868a0b2"} err="failed to get container status \"2265a4be89a538489e7ad0f7ad150ccb03be06746a68da7229d145e50868a0b2\": rpc error: code = NotFound desc = could not find container \"2265a4be89a538489e7ad0f7ad150ccb03be06746a68da7229d145e50868a0b2\": container with ID starting with 2265a4be89a538489e7ad0f7ad150ccb03be06746a68da7229d145e50868a0b2 not found: ID does not exist" Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.655208 4809 scope.go:117] "RemoveContainer" containerID="fa43abeec7d8474fb8e3bb23fb13aa511eb4a0eca9ed5617eb0c302009e3fe85" Dec 06 06:20:19 crc kubenswrapper[4809]: E1206 06:20:19.655683 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa43abeec7d8474fb8e3bb23fb13aa511eb4a0eca9ed5617eb0c302009e3fe85\": container with ID starting with fa43abeec7d8474fb8e3bb23fb13aa511eb4a0eca9ed5617eb0c302009e3fe85 not found: ID does not exist" containerID="fa43abeec7d8474fb8e3bb23fb13aa511eb4a0eca9ed5617eb0c302009e3fe85" Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.655716 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa43abeec7d8474fb8e3bb23fb13aa511eb4a0eca9ed5617eb0c302009e3fe85"} err="failed to get container status \"fa43abeec7d8474fb8e3bb23fb13aa511eb4a0eca9ed5617eb0c302009e3fe85\": rpc error: code = NotFound desc = could not find container \"fa43abeec7d8474fb8e3bb23fb13aa511eb4a0eca9ed5617eb0c302009e3fe85\": container with ID starting with fa43abeec7d8474fb8e3bb23fb13aa511eb4a0eca9ed5617eb0c302009e3fe85 not found: ID does not exist" Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.692288 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43d072b1-a024-430b-ad1f-4c669e35349c-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.929633 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.939825 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.958059 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:20:19 crc kubenswrapper[4809]: E1206 06:20:19.958629 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37c89662-6b35-4e73-bf37-f6bb01978c5f" containerName="dnsmasq-dns" Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.958656 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="37c89662-6b35-4e73-bf37-f6bb01978c5f" containerName="dnsmasq-dns" Dec 06 06:20:19 crc kubenswrapper[4809]: E1206 06:20:19.958675 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43d072b1-a024-430b-ad1f-4c669e35349c" containerName="proxy-httpd" Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.958686 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="43d072b1-a024-430b-ad1f-4c669e35349c" containerName="proxy-httpd" Dec 06 06:20:19 crc kubenswrapper[4809]: E1206 06:20:19.958704 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43d072b1-a024-430b-ad1f-4c669e35349c" containerName="sg-core" Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.958711 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="43d072b1-a024-430b-ad1f-4c669e35349c" containerName="sg-core" Dec 06 06:20:19 crc kubenswrapper[4809]: E1206 06:20:19.958727 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37c89662-6b35-4e73-bf37-f6bb01978c5f" containerName="init" Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.958736 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="37c89662-6b35-4e73-bf37-f6bb01978c5f" containerName="init" Dec 06 06:20:19 crc kubenswrapper[4809]: E1206 06:20:19.958770 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43d072b1-a024-430b-ad1f-4c669e35349c" containerName="ceilometer-notification-agent" Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.958778 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="43d072b1-a024-430b-ad1f-4c669e35349c" containerName="ceilometer-notification-agent" Dec 06 06:20:19 crc kubenswrapper[4809]: E1206 06:20:19.958796 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43d072b1-a024-430b-ad1f-4c669e35349c" containerName="ceilometer-central-agent" Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.958806 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="43d072b1-a024-430b-ad1f-4c669e35349c" containerName="ceilometer-central-agent" Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.959071 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="43d072b1-a024-430b-ad1f-4c669e35349c" containerName="proxy-httpd" Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.959110 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="37c89662-6b35-4e73-bf37-f6bb01978c5f" containerName="dnsmasq-dns" Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.959135 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="43d072b1-a024-430b-ad1f-4c669e35349c" containerName="ceilometer-central-agent" Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.959150 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="43d072b1-a024-430b-ad1f-4c669e35349c" containerName="sg-core" Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.959169 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="43d072b1-a024-430b-ad1f-4c669e35349c" containerName="ceilometer-notification-agent" Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.962011 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.969046 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.969093 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 06:20:19 crc kubenswrapper[4809]: I1206 06:20:19.977756 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:20:20 crc kubenswrapper[4809]: I1206 06:20:20.100104 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6be13a68-b3a2-4627-a468-bfe138f8e271-log-httpd\") pod \"ceilometer-0\" (UID: \"6be13a68-b3a2-4627-a468-bfe138f8e271\") " pod="openstack/ceilometer-0" Dec 06 06:20:20 crc kubenswrapper[4809]: I1206 06:20:20.100174 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6be13a68-b3a2-4627-a468-bfe138f8e271-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6be13a68-b3a2-4627-a468-bfe138f8e271\") " pod="openstack/ceilometer-0" Dec 06 06:20:20 crc kubenswrapper[4809]: I1206 06:20:20.100204 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6be13a68-b3a2-4627-a468-bfe138f8e271-scripts\") pod \"ceilometer-0\" (UID: \"6be13a68-b3a2-4627-a468-bfe138f8e271\") " pod="openstack/ceilometer-0" Dec 06 06:20:20 crc kubenswrapper[4809]: I1206 06:20:20.100241 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6be13a68-b3a2-4627-a468-bfe138f8e271-config-data\") pod \"ceilometer-0\" (UID: \"6be13a68-b3a2-4627-a468-bfe138f8e271\") " pod="openstack/ceilometer-0" Dec 06 06:20:20 crc kubenswrapper[4809]: I1206 06:20:20.100293 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6be13a68-b3a2-4627-a468-bfe138f8e271-run-httpd\") pod \"ceilometer-0\" (UID: \"6be13a68-b3a2-4627-a468-bfe138f8e271\") " pod="openstack/ceilometer-0" Dec 06 06:20:20 crc kubenswrapper[4809]: I1206 06:20:20.100357 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5qcf\" (UniqueName: \"kubernetes.io/projected/6be13a68-b3a2-4627-a468-bfe138f8e271-kube-api-access-q5qcf\") pod \"ceilometer-0\" (UID: \"6be13a68-b3a2-4627-a468-bfe138f8e271\") " pod="openstack/ceilometer-0" Dec 06 06:20:20 crc kubenswrapper[4809]: I1206 06:20:20.100420 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6be13a68-b3a2-4627-a468-bfe138f8e271-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6be13a68-b3a2-4627-a468-bfe138f8e271\") " pod="openstack/ceilometer-0" Dec 06 06:20:20 crc kubenswrapper[4809]: I1206 06:20:20.202507 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6be13a68-b3a2-4627-a468-bfe138f8e271-run-httpd\") pod \"ceilometer-0\" (UID: \"6be13a68-b3a2-4627-a468-bfe138f8e271\") " pod="openstack/ceilometer-0" Dec 06 06:20:20 crc kubenswrapper[4809]: I1206 06:20:20.202614 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5qcf\" (UniqueName: \"kubernetes.io/projected/6be13a68-b3a2-4627-a468-bfe138f8e271-kube-api-access-q5qcf\") pod \"ceilometer-0\" (UID: \"6be13a68-b3a2-4627-a468-bfe138f8e271\") " pod="openstack/ceilometer-0" Dec 06 06:20:20 crc kubenswrapper[4809]: I1206 06:20:20.202704 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6be13a68-b3a2-4627-a468-bfe138f8e271-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6be13a68-b3a2-4627-a468-bfe138f8e271\") " pod="openstack/ceilometer-0" Dec 06 06:20:20 crc kubenswrapper[4809]: I1206 06:20:20.202744 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6be13a68-b3a2-4627-a468-bfe138f8e271-log-httpd\") pod \"ceilometer-0\" (UID: \"6be13a68-b3a2-4627-a468-bfe138f8e271\") " pod="openstack/ceilometer-0" Dec 06 06:20:20 crc kubenswrapper[4809]: I1206 06:20:20.202789 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6be13a68-b3a2-4627-a468-bfe138f8e271-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6be13a68-b3a2-4627-a468-bfe138f8e271\") " pod="openstack/ceilometer-0" Dec 06 06:20:20 crc kubenswrapper[4809]: I1206 06:20:20.202821 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6be13a68-b3a2-4627-a468-bfe138f8e271-scripts\") pod \"ceilometer-0\" (UID: \"6be13a68-b3a2-4627-a468-bfe138f8e271\") " pod="openstack/ceilometer-0" Dec 06 06:20:20 crc kubenswrapper[4809]: I1206 06:20:20.202886 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6be13a68-b3a2-4627-a468-bfe138f8e271-config-data\") pod \"ceilometer-0\" (UID: \"6be13a68-b3a2-4627-a468-bfe138f8e271\") " pod="openstack/ceilometer-0" Dec 06 06:20:20 crc kubenswrapper[4809]: I1206 06:20:20.203035 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6be13a68-b3a2-4627-a468-bfe138f8e271-run-httpd\") pod \"ceilometer-0\" (UID: \"6be13a68-b3a2-4627-a468-bfe138f8e271\") " pod="openstack/ceilometer-0" Dec 06 06:20:20 crc kubenswrapper[4809]: I1206 06:20:20.203273 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6be13a68-b3a2-4627-a468-bfe138f8e271-log-httpd\") pod \"ceilometer-0\" (UID: \"6be13a68-b3a2-4627-a468-bfe138f8e271\") " pod="openstack/ceilometer-0" Dec 06 06:20:20 crc kubenswrapper[4809]: I1206 06:20:20.207051 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6be13a68-b3a2-4627-a468-bfe138f8e271-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6be13a68-b3a2-4627-a468-bfe138f8e271\") " pod="openstack/ceilometer-0" Dec 06 06:20:20 crc kubenswrapper[4809]: I1206 06:20:20.207758 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6be13a68-b3a2-4627-a468-bfe138f8e271-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6be13a68-b3a2-4627-a468-bfe138f8e271\") " pod="openstack/ceilometer-0" Dec 06 06:20:20 crc kubenswrapper[4809]: I1206 06:20:20.209756 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6be13a68-b3a2-4627-a468-bfe138f8e271-scripts\") pod \"ceilometer-0\" (UID: \"6be13a68-b3a2-4627-a468-bfe138f8e271\") " pod="openstack/ceilometer-0" Dec 06 06:20:20 crc kubenswrapper[4809]: I1206 06:20:20.210470 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6be13a68-b3a2-4627-a468-bfe138f8e271-config-data\") pod \"ceilometer-0\" (UID: \"6be13a68-b3a2-4627-a468-bfe138f8e271\") " pod="openstack/ceilometer-0" Dec 06 06:20:20 crc kubenswrapper[4809]: I1206 06:20:20.221521 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5qcf\" (UniqueName: \"kubernetes.io/projected/6be13a68-b3a2-4627-a468-bfe138f8e271-kube-api-access-q5qcf\") pod \"ceilometer-0\" (UID: \"6be13a68-b3a2-4627-a468-bfe138f8e271\") " pod="openstack/ceilometer-0" Dec 06 06:20:20 crc kubenswrapper[4809]: I1206 06:20:20.280848 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:20:20 crc kubenswrapper[4809]: I1206 06:20:20.774479 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:20:21 crc kubenswrapper[4809]: I1206 06:20:21.404569 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43d072b1-a024-430b-ad1f-4c669e35349c" path="/var/lib/kubelet/pods/43d072b1-a024-430b-ad1f-4c669e35349c/volumes" Dec 06 06:20:21 crc kubenswrapper[4809]: I1206 06:20:21.583988 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6be13a68-b3a2-4627-a468-bfe138f8e271","Type":"ContainerStarted","Data":"9e2715000c91670c1179b414e60ad5823615d33970710722c170c0c92af55aed"} Dec 06 06:20:21 crc kubenswrapper[4809]: I1206 06:20:21.585873 4809 generic.go:334] "Generic (PLEG): container finished" podID="4e106bab-59d8-4d2e-89f9-1f46e9b37f43" containerID="ee227d98802b9bb43d7d6c3599870ec807f20cfa515b0f11e3a0a8d474c3cb99" exitCode=0 Dec 06 06:20:21 crc kubenswrapper[4809]: I1206 06:20:21.585956 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-tp9tt" event={"ID":"4e106bab-59d8-4d2e-89f9-1f46e9b37f43","Type":"ContainerDied","Data":"ee227d98802b9bb43d7d6c3599870ec807f20cfa515b0f11e3a0a8d474c3cb99"} Dec 06 06:20:22 crc kubenswrapper[4809]: I1206 06:20:22.619174 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 06 06:20:22 crc kubenswrapper[4809]: I1206 06:20:22.622320 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 06 06:20:22 crc kubenswrapper[4809]: I1206 06:20:22.634927 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 06 06:20:23 crc kubenswrapper[4809]: I1206 06:20:23.028773 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-tp9tt" Dec 06 06:20:23 crc kubenswrapper[4809]: I1206 06:20:23.069647 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h797q\" (UniqueName: \"kubernetes.io/projected/4e106bab-59d8-4d2e-89f9-1f46e9b37f43-kube-api-access-h797q\") pod \"4e106bab-59d8-4d2e-89f9-1f46e9b37f43\" (UID: \"4e106bab-59d8-4d2e-89f9-1f46e9b37f43\") " Dec 06 06:20:23 crc kubenswrapper[4809]: I1206 06:20:23.069801 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e106bab-59d8-4d2e-89f9-1f46e9b37f43-combined-ca-bundle\") pod \"4e106bab-59d8-4d2e-89f9-1f46e9b37f43\" (UID: \"4e106bab-59d8-4d2e-89f9-1f46e9b37f43\") " Dec 06 06:20:23 crc kubenswrapper[4809]: I1206 06:20:23.069894 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e106bab-59d8-4d2e-89f9-1f46e9b37f43-scripts\") pod \"4e106bab-59d8-4d2e-89f9-1f46e9b37f43\" (UID: \"4e106bab-59d8-4d2e-89f9-1f46e9b37f43\") " Dec 06 06:20:23 crc kubenswrapper[4809]: I1206 06:20:23.069982 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e106bab-59d8-4d2e-89f9-1f46e9b37f43-config-data\") pod \"4e106bab-59d8-4d2e-89f9-1f46e9b37f43\" (UID: \"4e106bab-59d8-4d2e-89f9-1f46e9b37f43\") " Dec 06 06:20:23 crc kubenswrapper[4809]: I1206 06:20:23.075293 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e106bab-59d8-4d2e-89f9-1f46e9b37f43-scripts" (OuterVolumeSpecName: "scripts") pod "4e106bab-59d8-4d2e-89f9-1f46e9b37f43" (UID: "4e106bab-59d8-4d2e-89f9-1f46e9b37f43"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:20:23 crc kubenswrapper[4809]: I1206 06:20:23.075502 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e106bab-59d8-4d2e-89f9-1f46e9b37f43-kube-api-access-h797q" (OuterVolumeSpecName: "kube-api-access-h797q") pod "4e106bab-59d8-4d2e-89f9-1f46e9b37f43" (UID: "4e106bab-59d8-4d2e-89f9-1f46e9b37f43"). InnerVolumeSpecName "kube-api-access-h797q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:20:23 crc kubenswrapper[4809]: W1206 06:20:23.118787 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod43d072b1_a024_430b_ad1f_4c669e35349c.slice/crio-23b27c62d969be5b621f105b93f7f5846923671a71afda236f53559ab6a8a31a.scope WatchSource:0}: Error finding container 23b27c62d969be5b621f105b93f7f5846923671a71afda236f53559ab6a8a31a: Status 404 returned error can't find the container with id 23b27c62d969be5b621f105b93f7f5846923671a71afda236f53559ab6a8a31a Dec 06 06:20:23 crc kubenswrapper[4809]: I1206 06:20:23.121072 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e106bab-59d8-4d2e-89f9-1f46e9b37f43-config-data" (OuterVolumeSpecName: "config-data") pod "4e106bab-59d8-4d2e-89f9-1f46e9b37f43" (UID: "4e106bab-59d8-4d2e-89f9-1f46e9b37f43"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:20:23 crc kubenswrapper[4809]: I1206 06:20:23.134243 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e106bab-59d8-4d2e-89f9-1f46e9b37f43-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4e106bab-59d8-4d2e-89f9-1f46e9b37f43" (UID: "4e106bab-59d8-4d2e-89f9-1f46e9b37f43"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:20:23 crc kubenswrapper[4809]: I1206 06:20:23.172362 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e106bab-59d8-4d2e-89f9-1f46e9b37f43-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:23 crc kubenswrapper[4809]: I1206 06:20:23.172431 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e106bab-59d8-4d2e-89f9-1f46e9b37f43-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:23 crc kubenswrapper[4809]: I1206 06:20:23.172447 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h797q\" (UniqueName: \"kubernetes.io/projected/4e106bab-59d8-4d2e-89f9-1f46e9b37f43-kube-api-access-h797q\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:23 crc kubenswrapper[4809]: I1206 06:20:23.172886 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e106bab-59d8-4d2e-89f9-1f46e9b37f43-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:23 crc kubenswrapper[4809]: E1206 06:20:23.281851 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod413494e9_0401_4a6c_b85e_8e6be9997d20.slice/crio-conmon-2e551eb383880763e22b099bbea7d3b3355007a1cc0163bc7551e4a9006d34a3.scope\": RecentStats: unable to find data in memory cache]" Dec 06 06:20:23 crc kubenswrapper[4809]: E1206 06:20:23.281917 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod43d072b1_a024_430b_ad1f_4c669e35349c.slice/crio-b32efe19327ba089385c3fe5c6dc3f5d761d142991d09b630cdd902293d67fa2.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod43d072b1_a024_430b_ad1f_4c669e35349c.slice/crio-conmon-b32efe19327ba089385c3fe5c6dc3f5d761d142991d09b630cdd902293d67fa2.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4e106bab_59d8_4d2e_89f9_1f46e9b37f43.slice/crio-ee227d98802b9bb43d7d6c3599870ec807f20cfa515b0f11e3a0a8d474c3cb99.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod37c89662_6b35_4e73_bf37_f6bb01978c5f.slice/crio-conmon-54bc7c704b1b4e133a8e1806521b4431560c2ad2f82669b44524c18ffb5d443c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod37c89662_6b35_4e73_bf37_f6bb01978c5f.slice/crio-a3b2aeabf80e82647cf1efb7c757498fecab574b5cc8fa45f0d14668ecea2846\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod43d072b1_a024_430b_ad1f_4c669e35349c.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod43d072b1_a024_430b_ad1f_4c669e35349c.slice/crio-conmon-fa43abeec7d8474fb8e3bb23fb13aa511eb4a0eca9ed5617eb0c302009e3fe85.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod37c89662_6b35_4e73_bf37_f6bb01978c5f.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod43d072b1_a024_430b_ad1f_4c669e35349c.slice/crio-conmon-2265a4be89a538489e7ad0f7ad150ccb03be06746a68da7229d145e50868a0b2.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod413494e9_0401_4a6c_b85e_8e6be9997d20.slice/crio-2e551eb383880763e22b099bbea7d3b3355007a1cc0163bc7551e4a9006d34a3.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod43d072b1_a024_430b_ad1f_4c669e35349c.slice/crio-fa43abeec7d8474fb8e3bb23fb13aa511eb4a0eca9ed5617eb0c302009e3fe85.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4e106bab_59d8_4d2e_89f9_1f46e9b37f43.slice/crio-conmon-ee227d98802b9bb43d7d6c3599870ec807f20cfa515b0f11e3a0a8d474c3cb99.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod37c89662_6b35_4e73_bf37_f6bb01978c5f.slice/crio-54bc7c704b1b4e133a8e1806521b4431560c2ad2f82669b44524c18ffb5d443c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod413494e9_0401_4a6c_b85e_8e6be9997d20.slice/crio-conmon-2e551eb383880763e22b099bbea7d3b3355007a1cc0163bc7551e4a9006d34a3.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod43d072b1_a024_430b_ad1f_4c669e35349c.slice/crio-2265a4be89a538489e7ad0f7ad150ccb03be06746a68da7229d145e50868a0b2.scope\": RecentStats: unable to find data in memory cache]" Dec 06 06:20:23 crc kubenswrapper[4809]: I1206 06:20:23.610092 4809 generic.go:334] "Generic (PLEG): container finished" podID="413494e9-0401-4a6c-b85e-8e6be9997d20" containerID="2e551eb383880763e22b099bbea7d3b3355007a1cc0163bc7551e4a9006d34a3" exitCode=137 Dec 06 06:20:23 crc kubenswrapper[4809]: I1206 06:20:23.610162 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"413494e9-0401-4a6c-b85e-8e6be9997d20","Type":"ContainerDied","Data":"2e551eb383880763e22b099bbea7d3b3355007a1cc0163bc7551e4a9006d34a3"} Dec 06 06:20:23 crc kubenswrapper[4809]: I1206 06:20:23.612337 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6be13a68-b3a2-4627-a468-bfe138f8e271","Type":"ContainerStarted","Data":"395cee0fe2d65139489165e1abcab94073496fc112f29c7a2421e942a24d1c7a"} Dec 06 06:20:23 crc kubenswrapper[4809]: I1206 06:20:23.613977 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-tp9tt" event={"ID":"4e106bab-59d8-4d2e-89f9-1f46e9b37f43","Type":"ContainerDied","Data":"49caf496c4a72e06508b2c1db28e820afd3acf01b5a41876151f9c1e40491321"} Dec 06 06:20:23 crc kubenswrapper[4809]: I1206 06:20:23.614064 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="49caf496c4a72e06508b2c1db28e820afd3acf01b5a41876151f9c1e40491321" Dec 06 06:20:23 crc kubenswrapper[4809]: I1206 06:20:23.614029 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-tp9tt" Dec 06 06:20:23 crc kubenswrapper[4809]: I1206 06:20:23.624606 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 06 06:20:23 crc kubenswrapper[4809]: I1206 06:20:23.797602 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 06:20:23 crc kubenswrapper[4809]: I1206 06:20:23.797818 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="65d1f54f-e7cd-467f-a385-e081568dc6f6" containerName="nova-scheduler-scheduler" containerID="cri-o://651a13e2264e20f136248c2269f20b93f1368a8b7e29b9bd41af0c3c1ed835b8" gracePeriod=30 Dec 06 06:20:23 crc kubenswrapper[4809]: I1206 06:20:23.846159 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.214487 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 06 06:20:24 crc kubenswrapper[4809]: E1206 06:20:24.268175 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="651a13e2264e20f136248c2269f20b93f1368a8b7e29b9bd41af0c3c1ed835b8" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 06 06:20:24 crc kubenswrapper[4809]: E1206 06:20:24.272842 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="651a13e2264e20f136248c2269f20b93f1368a8b7e29b9bd41af0c3c1ed835b8" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 06 06:20:24 crc kubenswrapper[4809]: E1206 06:20:24.274976 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="651a13e2264e20f136248c2269f20b93f1368a8b7e29b9bd41af0c3c1ed835b8" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 06 06:20:24 crc kubenswrapper[4809]: E1206 06:20:24.275044 4809 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="65d1f54f-e7cd-467f-a385-e081568dc6f6" containerName="nova-scheduler-scheduler" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.303421 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/413494e9-0401-4a6c-b85e-8e6be9997d20-combined-ca-bundle\") pod \"413494e9-0401-4a6c-b85e-8e6be9997d20\" (UID: \"413494e9-0401-4a6c-b85e-8e6be9997d20\") " Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.303504 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/413494e9-0401-4a6c-b85e-8e6be9997d20-config-data\") pod \"413494e9-0401-4a6c-b85e-8e6be9997d20\" (UID: \"413494e9-0401-4a6c-b85e-8e6be9997d20\") " Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.303695 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/413494e9-0401-4a6c-b85e-8e6be9997d20-scripts\") pod \"413494e9-0401-4a6c-b85e-8e6be9997d20\" (UID: \"413494e9-0401-4a6c-b85e-8e6be9997d20\") " Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.303800 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ss5v6\" (UniqueName: \"kubernetes.io/projected/413494e9-0401-4a6c-b85e-8e6be9997d20-kube-api-access-ss5v6\") pod \"413494e9-0401-4a6c-b85e-8e6be9997d20\" (UID: \"413494e9-0401-4a6c-b85e-8e6be9997d20\") " Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.310575 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/413494e9-0401-4a6c-b85e-8e6be9997d20-scripts" (OuterVolumeSpecName: "scripts") pod "413494e9-0401-4a6c-b85e-8e6be9997d20" (UID: "413494e9-0401-4a6c-b85e-8e6be9997d20"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.312081 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/413494e9-0401-4a6c-b85e-8e6be9997d20-kube-api-access-ss5v6" (OuterVolumeSpecName: "kube-api-access-ss5v6") pod "413494e9-0401-4a6c-b85e-8e6be9997d20" (UID: "413494e9-0401-4a6c-b85e-8e6be9997d20"). InnerVolumeSpecName "kube-api-access-ss5v6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.409039 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/413494e9-0401-4a6c-b85e-8e6be9997d20-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.409355 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ss5v6\" (UniqueName: \"kubernetes.io/projected/413494e9-0401-4a6c-b85e-8e6be9997d20-kube-api-access-ss5v6\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.461078 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/413494e9-0401-4a6c-b85e-8e6be9997d20-config-data" (OuterVolumeSpecName: "config-data") pod "413494e9-0401-4a6c-b85e-8e6be9997d20" (UID: "413494e9-0401-4a6c-b85e-8e6be9997d20"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.466431 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/413494e9-0401-4a6c-b85e-8e6be9997d20-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "413494e9-0401-4a6c-b85e-8e6be9997d20" (UID: "413494e9-0401-4a6c-b85e-8e6be9997d20"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.512001 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/413494e9-0401-4a6c-b85e-8e6be9997d20-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.512040 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/413494e9-0401-4a6c-b85e-8e6be9997d20-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.640540 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"413494e9-0401-4a6c-b85e-8e6be9997d20","Type":"ContainerDied","Data":"f352c0ef812ecfcd4d21ad8a64c31e41640ca38385fb27da3ea01c9cfb3c88d0"} Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.640598 4809 scope.go:117] "RemoveContainer" containerID="2e551eb383880763e22b099bbea7d3b3355007a1cc0163bc7551e4a9006d34a3" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.640751 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.653793 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6be13a68-b3a2-4627-a468-bfe138f8e271","Type":"ContainerStarted","Data":"4924bbf255027594929d8c8877a450eb6f34c2b6f12af3f9d584a9d58ac98b13"} Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.673205 4809 scope.go:117] "RemoveContainer" containerID="c7a165ffefc33081d973a91c041184fc11186d1219545d5ae8e3ca366b218b2a" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.742620 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.749358 4809 scope.go:117] "RemoveContainer" containerID="49a05cb8e9cda351bdd9cd29857e7bf2579eab07f1408ebc2efd1ce510d592c6" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.778776 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.786685 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 06 06:20:24 crc kubenswrapper[4809]: E1206 06:20:24.787452 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e106bab-59d8-4d2e-89f9-1f46e9b37f43" containerName="nova-manage" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.787475 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e106bab-59d8-4d2e-89f9-1f46e9b37f43" containerName="nova-manage" Dec 06 06:20:24 crc kubenswrapper[4809]: E1206 06:20:24.787496 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="413494e9-0401-4a6c-b85e-8e6be9997d20" containerName="aodh-evaluator" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.787509 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="413494e9-0401-4a6c-b85e-8e6be9997d20" containerName="aodh-evaluator" Dec 06 06:20:24 crc kubenswrapper[4809]: E1206 06:20:24.787533 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="413494e9-0401-4a6c-b85e-8e6be9997d20" containerName="aodh-listener" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.787546 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="413494e9-0401-4a6c-b85e-8e6be9997d20" containerName="aodh-listener" Dec 06 06:20:24 crc kubenswrapper[4809]: E1206 06:20:24.787563 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="413494e9-0401-4a6c-b85e-8e6be9997d20" containerName="aodh-api" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.787569 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="413494e9-0401-4a6c-b85e-8e6be9997d20" containerName="aodh-api" Dec 06 06:20:24 crc kubenswrapper[4809]: E1206 06:20:24.787585 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="413494e9-0401-4a6c-b85e-8e6be9997d20" containerName="aodh-notifier" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.787593 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="413494e9-0401-4a6c-b85e-8e6be9997d20" containerName="aodh-notifier" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.788061 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="413494e9-0401-4a6c-b85e-8e6be9997d20" containerName="aodh-notifier" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.788084 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="413494e9-0401-4a6c-b85e-8e6be9997d20" containerName="aodh-listener" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.788110 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="413494e9-0401-4a6c-b85e-8e6be9997d20" containerName="aodh-api" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.788126 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="413494e9-0401-4a6c-b85e-8e6be9997d20" containerName="aodh-evaluator" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.788138 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e106bab-59d8-4d2e-89f9-1f46e9b37f43" containerName="nova-manage" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.790420 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.798601 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-f84tp" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.799070 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.799246 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.799438 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.799584 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.799707 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.831043 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/133d9795-adf6-44a2-908b-9790b2b6ed60-public-tls-certs\") pod \"aodh-0\" (UID: \"133d9795-adf6-44a2-908b-9790b2b6ed60\") " pod="openstack/aodh-0" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.831237 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/133d9795-adf6-44a2-908b-9790b2b6ed60-internal-tls-certs\") pod \"aodh-0\" (UID: \"133d9795-adf6-44a2-908b-9790b2b6ed60\") " pod="openstack/aodh-0" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.831383 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9ksb\" (UniqueName: \"kubernetes.io/projected/133d9795-adf6-44a2-908b-9790b2b6ed60-kube-api-access-j9ksb\") pod \"aodh-0\" (UID: \"133d9795-adf6-44a2-908b-9790b2b6ed60\") " pod="openstack/aodh-0" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.831464 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/133d9795-adf6-44a2-908b-9790b2b6ed60-config-data\") pod \"aodh-0\" (UID: \"133d9795-adf6-44a2-908b-9790b2b6ed60\") " pod="openstack/aodh-0" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.831503 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/133d9795-adf6-44a2-908b-9790b2b6ed60-combined-ca-bundle\") pod \"aodh-0\" (UID: \"133d9795-adf6-44a2-908b-9790b2b6ed60\") " pod="openstack/aodh-0" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.831609 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/133d9795-adf6-44a2-908b-9790b2b6ed60-scripts\") pod \"aodh-0\" (UID: \"133d9795-adf6-44a2-908b-9790b2b6ed60\") " pod="openstack/aodh-0" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.923613 4809 scope.go:117] "RemoveContainer" containerID="7e1832c98b7bbb803ca4af4d248b206551c35844e83b9caceeeff49ade9e0215" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.936879 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/133d9795-adf6-44a2-908b-9790b2b6ed60-scripts\") pod \"aodh-0\" (UID: \"133d9795-adf6-44a2-908b-9790b2b6ed60\") " pod="openstack/aodh-0" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.937328 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/133d9795-adf6-44a2-908b-9790b2b6ed60-public-tls-certs\") pod \"aodh-0\" (UID: \"133d9795-adf6-44a2-908b-9790b2b6ed60\") " pod="openstack/aodh-0" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.937420 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/133d9795-adf6-44a2-908b-9790b2b6ed60-internal-tls-certs\") pod \"aodh-0\" (UID: \"133d9795-adf6-44a2-908b-9790b2b6ed60\") " pod="openstack/aodh-0" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.937490 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9ksb\" (UniqueName: \"kubernetes.io/projected/133d9795-adf6-44a2-908b-9790b2b6ed60-kube-api-access-j9ksb\") pod \"aodh-0\" (UID: \"133d9795-adf6-44a2-908b-9790b2b6ed60\") " pod="openstack/aodh-0" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.937537 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/133d9795-adf6-44a2-908b-9790b2b6ed60-config-data\") pod \"aodh-0\" (UID: \"133d9795-adf6-44a2-908b-9790b2b6ed60\") " pod="openstack/aodh-0" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.937564 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/133d9795-adf6-44a2-908b-9790b2b6ed60-combined-ca-bundle\") pod \"aodh-0\" (UID: \"133d9795-adf6-44a2-908b-9790b2b6ed60\") " pod="openstack/aodh-0" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.944925 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/133d9795-adf6-44a2-908b-9790b2b6ed60-internal-tls-certs\") pod \"aodh-0\" (UID: \"133d9795-adf6-44a2-908b-9790b2b6ed60\") " pod="openstack/aodh-0" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.948060 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/133d9795-adf6-44a2-908b-9790b2b6ed60-scripts\") pod \"aodh-0\" (UID: \"133d9795-adf6-44a2-908b-9790b2b6ed60\") " pod="openstack/aodh-0" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.962115 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/133d9795-adf6-44a2-908b-9790b2b6ed60-combined-ca-bundle\") pod \"aodh-0\" (UID: \"133d9795-adf6-44a2-908b-9790b2b6ed60\") " pod="openstack/aodh-0" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.965389 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/133d9795-adf6-44a2-908b-9790b2b6ed60-public-tls-certs\") pod \"aodh-0\" (UID: \"133d9795-adf6-44a2-908b-9790b2b6ed60\") " pod="openstack/aodh-0" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.969440 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9ksb\" (UniqueName: \"kubernetes.io/projected/133d9795-adf6-44a2-908b-9790b2b6ed60-kube-api-access-j9ksb\") pod \"aodh-0\" (UID: \"133d9795-adf6-44a2-908b-9790b2b6ed60\") " pod="openstack/aodh-0" Dec 06 06:20:24 crc kubenswrapper[4809]: I1206 06:20:24.974031 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/133d9795-adf6-44a2-908b-9790b2b6ed60-config-data\") pod \"aodh-0\" (UID: \"133d9795-adf6-44a2-908b-9790b2b6ed60\") " pod="openstack/aodh-0" Dec 06 06:20:25 crc kubenswrapper[4809]: I1206 06:20:25.212853 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 06 06:20:25 crc kubenswrapper[4809]: I1206 06:20:25.266512 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 06:20:25 crc kubenswrapper[4809]: I1206 06:20:25.353139 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65d1f54f-e7cd-467f-a385-e081568dc6f6-config-data\") pod \"65d1f54f-e7cd-467f-a385-e081568dc6f6\" (UID: \"65d1f54f-e7cd-467f-a385-e081568dc6f6\") " Dec 06 06:20:25 crc kubenswrapper[4809]: I1206 06:20:25.353553 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jk7qq\" (UniqueName: \"kubernetes.io/projected/65d1f54f-e7cd-467f-a385-e081568dc6f6-kube-api-access-jk7qq\") pod \"65d1f54f-e7cd-467f-a385-e081568dc6f6\" (UID: \"65d1f54f-e7cd-467f-a385-e081568dc6f6\") " Dec 06 06:20:25 crc kubenswrapper[4809]: I1206 06:20:25.353767 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65d1f54f-e7cd-467f-a385-e081568dc6f6-combined-ca-bundle\") pod \"65d1f54f-e7cd-467f-a385-e081568dc6f6\" (UID: \"65d1f54f-e7cd-467f-a385-e081568dc6f6\") " Dec 06 06:20:25 crc kubenswrapper[4809]: I1206 06:20:25.369066 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65d1f54f-e7cd-467f-a385-e081568dc6f6-kube-api-access-jk7qq" (OuterVolumeSpecName: "kube-api-access-jk7qq") pod "65d1f54f-e7cd-467f-a385-e081568dc6f6" (UID: "65d1f54f-e7cd-467f-a385-e081568dc6f6"). InnerVolumeSpecName "kube-api-access-jk7qq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:20:25 crc kubenswrapper[4809]: I1206 06:20:25.411289 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="413494e9-0401-4a6c-b85e-8e6be9997d20" path="/var/lib/kubelet/pods/413494e9-0401-4a6c-b85e-8e6be9997d20/volumes" Dec 06 06:20:25 crc kubenswrapper[4809]: I1206 06:20:25.421110 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65d1f54f-e7cd-467f-a385-e081568dc6f6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "65d1f54f-e7cd-467f-a385-e081568dc6f6" (UID: "65d1f54f-e7cd-467f-a385-e081568dc6f6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:20:25 crc kubenswrapper[4809]: I1206 06:20:25.421315 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65d1f54f-e7cd-467f-a385-e081568dc6f6-config-data" (OuterVolumeSpecName: "config-data") pod "65d1f54f-e7cd-467f-a385-e081568dc6f6" (UID: "65d1f54f-e7cd-467f-a385-e081568dc6f6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:20:25 crc kubenswrapper[4809]: I1206 06:20:25.456703 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jk7qq\" (UniqueName: \"kubernetes.io/projected/65d1f54f-e7cd-467f-a385-e081568dc6f6-kube-api-access-jk7qq\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:25 crc kubenswrapper[4809]: I1206 06:20:25.456768 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65d1f54f-e7cd-467f-a385-e081568dc6f6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:25 crc kubenswrapper[4809]: I1206 06:20:25.456784 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65d1f54f-e7cd-467f-a385-e081568dc6f6-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:25 crc kubenswrapper[4809]: I1206 06:20:25.670666 4809 generic.go:334] "Generic (PLEG): container finished" podID="65d1f54f-e7cd-467f-a385-e081568dc6f6" containerID="651a13e2264e20f136248c2269f20b93f1368a8b7e29b9bd41af0c3c1ed835b8" exitCode=0 Dec 06 06:20:25 crc kubenswrapper[4809]: I1206 06:20:25.670722 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 06:20:25 crc kubenswrapper[4809]: I1206 06:20:25.670743 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"65d1f54f-e7cd-467f-a385-e081568dc6f6","Type":"ContainerDied","Data":"651a13e2264e20f136248c2269f20b93f1368a8b7e29b9bd41af0c3c1ed835b8"} Dec 06 06:20:25 crc kubenswrapper[4809]: I1206 06:20:25.671236 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"65d1f54f-e7cd-467f-a385-e081568dc6f6","Type":"ContainerDied","Data":"0f1fff03deeaa44564bdcaba4ee99ac1995319be786f9060dfd15b96289e4b21"} Dec 06 06:20:25 crc kubenswrapper[4809]: I1206 06:20:25.671255 4809 scope.go:117] "RemoveContainer" containerID="651a13e2264e20f136248c2269f20b93f1368a8b7e29b9bd41af0c3c1ed835b8" Dec 06 06:20:25 crc kubenswrapper[4809]: I1206 06:20:25.678923 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6be13a68-b3a2-4627-a468-bfe138f8e271","Type":"ContainerStarted","Data":"b6ec38c05e0256409801129a53fbc8fbf46cce01ce3dfd344bc30e17646b7936"} Dec 06 06:20:25 crc kubenswrapper[4809]: I1206 06:20:25.679151 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f17fe211-3570-4b69-866c-9607c6489341" containerName="nova-metadata-log" containerID="cri-o://5375b409ca2dbb8cc53e5c42961b8fbe2e4b34d2ddaf260aae13f98079f6fffd" gracePeriod=30 Dec 06 06:20:25 crc kubenswrapper[4809]: I1206 06:20:25.679240 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f17fe211-3570-4b69-866c-9607c6489341" containerName="nova-metadata-metadata" containerID="cri-o://9399ca5d70e9db739895e6db6766cd6ad4c43d81e9ba06135277d222b541cc75" gracePeriod=30 Dec 06 06:20:25 crc kubenswrapper[4809]: I1206 06:20:25.711842 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 06:20:25 crc kubenswrapper[4809]: I1206 06:20:25.719545 4809 scope.go:117] "RemoveContainer" containerID="651a13e2264e20f136248c2269f20b93f1368a8b7e29b9bd41af0c3c1ed835b8" Dec 06 06:20:25 crc kubenswrapper[4809]: E1206 06:20:25.720117 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"651a13e2264e20f136248c2269f20b93f1368a8b7e29b9bd41af0c3c1ed835b8\": container with ID starting with 651a13e2264e20f136248c2269f20b93f1368a8b7e29b9bd41af0c3c1ed835b8 not found: ID does not exist" containerID="651a13e2264e20f136248c2269f20b93f1368a8b7e29b9bd41af0c3c1ed835b8" Dec 06 06:20:25 crc kubenswrapper[4809]: I1206 06:20:25.720154 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"651a13e2264e20f136248c2269f20b93f1368a8b7e29b9bd41af0c3c1ed835b8"} err="failed to get container status \"651a13e2264e20f136248c2269f20b93f1368a8b7e29b9bd41af0c3c1ed835b8\": rpc error: code = NotFound desc = could not find container \"651a13e2264e20f136248c2269f20b93f1368a8b7e29b9bd41af0c3c1ed835b8\": container with ID starting with 651a13e2264e20f136248c2269f20b93f1368a8b7e29b9bd41af0c3c1ed835b8 not found: ID does not exist" Dec 06 06:20:25 crc kubenswrapper[4809]: I1206 06:20:25.730717 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 06:20:25 crc kubenswrapper[4809]: I1206 06:20:25.748166 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 06:20:25 crc kubenswrapper[4809]: E1206 06:20:25.748823 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65d1f54f-e7cd-467f-a385-e081568dc6f6" containerName="nova-scheduler-scheduler" Dec 06 06:20:25 crc kubenswrapper[4809]: I1206 06:20:25.748851 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="65d1f54f-e7cd-467f-a385-e081568dc6f6" containerName="nova-scheduler-scheduler" Dec 06 06:20:25 crc kubenswrapper[4809]: I1206 06:20:25.749186 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="65d1f54f-e7cd-467f-a385-e081568dc6f6" containerName="nova-scheduler-scheduler" Dec 06 06:20:25 crc kubenswrapper[4809]: I1206 06:20:25.750126 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 06:20:25 crc kubenswrapper[4809]: I1206 06:20:25.753156 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 06 06:20:25 crc kubenswrapper[4809]: I1206 06:20:25.767517 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 06:20:25 crc kubenswrapper[4809]: I1206 06:20:25.782555 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 06 06:20:25 crc kubenswrapper[4809]: I1206 06:20:25.865521 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjzcq\" (UniqueName: \"kubernetes.io/projected/43f53993-991a-450e-a663-acd9f070d9b6-kube-api-access-xjzcq\") pod \"nova-scheduler-0\" (UID: \"43f53993-991a-450e-a663-acd9f070d9b6\") " pod="openstack/nova-scheduler-0" Dec 06 06:20:25 crc kubenswrapper[4809]: I1206 06:20:25.865592 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43f53993-991a-450e-a663-acd9f070d9b6-config-data\") pod \"nova-scheduler-0\" (UID: \"43f53993-991a-450e-a663-acd9f070d9b6\") " pod="openstack/nova-scheduler-0" Dec 06 06:20:25 crc kubenswrapper[4809]: I1206 06:20:25.865761 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43f53993-991a-450e-a663-acd9f070d9b6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"43f53993-991a-450e-a663-acd9f070d9b6\") " pod="openstack/nova-scheduler-0" Dec 06 06:20:25 crc kubenswrapper[4809]: I1206 06:20:25.967323 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43f53993-991a-450e-a663-acd9f070d9b6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"43f53993-991a-450e-a663-acd9f070d9b6\") " pod="openstack/nova-scheduler-0" Dec 06 06:20:25 crc kubenswrapper[4809]: I1206 06:20:25.967553 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjzcq\" (UniqueName: \"kubernetes.io/projected/43f53993-991a-450e-a663-acd9f070d9b6-kube-api-access-xjzcq\") pod \"nova-scheduler-0\" (UID: \"43f53993-991a-450e-a663-acd9f070d9b6\") " pod="openstack/nova-scheduler-0" Dec 06 06:20:25 crc kubenswrapper[4809]: I1206 06:20:25.967588 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43f53993-991a-450e-a663-acd9f070d9b6-config-data\") pod \"nova-scheduler-0\" (UID: \"43f53993-991a-450e-a663-acd9f070d9b6\") " pod="openstack/nova-scheduler-0" Dec 06 06:20:25 crc kubenswrapper[4809]: I1206 06:20:25.975765 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43f53993-991a-450e-a663-acd9f070d9b6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"43f53993-991a-450e-a663-acd9f070d9b6\") " pod="openstack/nova-scheduler-0" Dec 06 06:20:25 crc kubenswrapper[4809]: I1206 06:20:25.977448 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43f53993-991a-450e-a663-acd9f070d9b6-config-data\") pod \"nova-scheduler-0\" (UID: \"43f53993-991a-450e-a663-acd9f070d9b6\") " pod="openstack/nova-scheduler-0" Dec 06 06:20:25 crc kubenswrapper[4809]: I1206 06:20:25.983981 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjzcq\" (UniqueName: \"kubernetes.io/projected/43f53993-991a-450e-a663-acd9f070d9b6-kube-api-access-xjzcq\") pod \"nova-scheduler-0\" (UID: \"43f53993-991a-450e-a663-acd9f070d9b6\") " pod="openstack/nova-scheduler-0" Dec 06 06:20:26 crc kubenswrapper[4809]: I1206 06:20:26.273857 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 06:20:26 crc kubenswrapper[4809]: I1206 06:20:26.689529 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"133d9795-adf6-44a2-908b-9790b2b6ed60","Type":"ContainerStarted","Data":"5564afd45ab40683a86fabb94556e733086104d51ea1565ca54f162edeebc0e0"} Dec 06 06:20:26 crc kubenswrapper[4809]: I1206 06:20:26.689996 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"133d9795-adf6-44a2-908b-9790b2b6ed60","Type":"ContainerStarted","Data":"c3ab1b43518b68e6638a3f9d4a602bbd3c914665a7180c79bc44368b20cd3de3"} Dec 06 06:20:26 crc kubenswrapper[4809]: I1206 06:20:26.692812 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6be13a68-b3a2-4627-a468-bfe138f8e271","Type":"ContainerStarted","Data":"1f3429a9e96b760fffceb1e08061b1428e6750a42a79dd8274be7e5e09bedb8d"} Dec 06 06:20:26 crc kubenswrapper[4809]: I1206 06:20:26.692979 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 06:20:26 crc kubenswrapper[4809]: I1206 06:20:26.694986 4809 generic.go:334] "Generic (PLEG): container finished" podID="f17fe211-3570-4b69-866c-9607c6489341" containerID="5375b409ca2dbb8cc53e5c42961b8fbe2e4b34d2ddaf260aae13f98079f6fffd" exitCode=143 Dec 06 06:20:26 crc kubenswrapper[4809]: I1206 06:20:26.695053 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f17fe211-3570-4b69-866c-9607c6489341","Type":"ContainerDied","Data":"5375b409ca2dbb8cc53e5c42961b8fbe2e4b34d2ddaf260aae13f98079f6fffd"} Dec 06 06:20:26 crc kubenswrapper[4809]: I1206 06:20:26.720757 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.704373081 podStartE2EDuration="7.720729858s" podCreationTimestamp="2025-12-06 06:20:19 +0000 UTC" firstStartedPulling="2025-12-06 06:20:20.773107426 +0000 UTC m=+1745.662090368" lastFinishedPulling="2025-12-06 06:20:25.789464203 +0000 UTC m=+1750.678447145" observedRunningTime="2025-12-06 06:20:26.712387793 +0000 UTC m=+1751.601370735" watchObservedRunningTime="2025-12-06 06:20:26.720729858 +0000 UTC m=+1751.609712800" Dec 06 06:20:26 crc kubenswrapper[4809]: W1206 06:20:26.770262 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod43f53993_991a_450e_a663_acd9f070d9b6.slice/crio-fb8cabe67b75dee8460a9cdc5e43b8fbe3556c5931e1fdcc9cf3328255ef5e2d WatchSource:0}: Error finding container fb8cabe67b75dee8460a9cdc5e43b8fbe3556c5931e1fdcc9cf3328255ef5e2d: Status 404 returned error can't find the container with id fb8cabe67b75dee8460a9cdc5e43b8fbe3556c5931e1fdcc9cf3328255ef5e2d Dec 06 06:20:26 crc kubenswrapper[4809]: I1206 06:20:26.778053 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 06:20:27 crc kubenswrapper[4809]: I1206 06:20:27.403534 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65d1f54f-e7cd-467f-a385-e081568dc6f6" path="/var/lib/kubelet/pods/65d1f54f-e7cd-467f-a385-e081568dc6f6/volumes" Dec 06 06:20:27 crc kubenswrapper[4809]: I1206 06:20:27.755107 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"133d9795-adf6-44a2-908b-9790b2b6ed60","Type":"ContainerStarted","Data":"05acd4adbf641b564c0376628972d4b4613caec5cf5354ad2ac88856db4f45a8"} Dec 06 06:20:27 crc kubenswrapper[4809]: I1206 06:20:27.763443 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"43f53993-991a-450e-a663-acd9f070d9b6","Type":"ContainerStarted","Data":"111952d76f211ad7fb7f1a32c3189c9da61fc647502732bde16496dc82928b6b"} Dec 06 06:20:27 crc kubenswrapper[4809]: I1206 06:20:27.763486 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"43f53993-991a-450e-a663-acd9f070d9b6","Type":"ContainerStarted","Data":"fb8cabe67b75dee8460a9cdc5e43b8fbe3556c5931e1fdcc9cf3328255ef5e2d"} Dec 06 06:20:28 crc kubenswrapper[4809]: I1206 06:20:28.772614 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"133d9795-adf6-44a2-908b-9790b2b6ed60","Type":"ContainerStarted","Data":"8121ca8bcd2847fcc69f3adfca90a523aa1dfd3a8953789f71ff030b1f35b5ed"} Dec 06 06:20:28 crc kubenswrapper[4809]: I1206 06:20:28.773298 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"133d9795-adf6-44a2-908b-9790b2b6ed60","Type":"ContainerStarted","Data":"ce59f044d20e36aae4146ae186fa291d71e4671b5e2b8253ce5c2d95beb99058"} Dec 06 06:20:28 crc kubenswrapper[4809]: I1206 06:20:28.797684 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.7976614939999997 podStartE2EDuration="3.797661494s" podCreationTimestamp="2025-12-06 06:20:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:20:27.798351412 +0000 UTC m=+1752.687334354" watchObservedRunningTime="2025-12-06 06:20:28.797661494 +0000 UTC m=+1753.686644436" Dec 06 06:20:28 crc kubenswrapper[4809]: I1206 06:20:28.845148 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="f17fe211-3570-4b69-866c-9607c6489341" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.248:8775/\": read tcp 10.217.0.2:44364->10.217.0.248:8775: read: connection reset by peer" Dec 06 06:20:28 crc kubenswrapper[4809]: I1206 06:20:28.845202 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="f17fe211-3570-4b69-866c-9607c6489341" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.248:8775/\": read tcp 10.217.0.2:44362->10.217.0.248:8775: read: connection reset by peer" Dec 06 06:20:29 crc kubenswrapper[4809]: I1206 06:20:29.404913 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 06:20:29 crc kubenswrapper[4809]: I1206 06:20:29.425434 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=3.049457077 podStartE2EDuration="5.425410959s" podCreationTimestamp="2025-12-06 06:20:24 +0000 UTC" firstStartedPulling="2025-12-06 06:20:25.803848412 +0000 UTC m=+1750.692831344" lastFinishedPulling="2025-12-06 06:20:28.179802284 +0000 UTC m=+1753.068785226" observedRunningTime="2025-12-06 06:20:28.794513509 +0000 UTC m=+1753.683496451" watchObservedRunningTime="2025-12-06 06:20:29.425410959 +0000 UTC m=+1754.314393901" Dec 06 06:20:29 crc kubenswrapper[4809]: I1206 06:20:29.476712 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f17fe211-3570-4b69-866c-9607c6489341-combined-ca-bundle\") pod \"f17fe211-3570-4b69-866c-9607c6489341\" (UID: \"f17fe211-3570-4b69-866c-9607c6489341\") " Dec 06 06:20:29 crc kubenswrapper[4809]: I1206 06:20:29.476841 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f17fe211-3570-4b69-866c-9607c6489341-nova-metadata-tls-certs\") pod \"f17fe211-3570-4b69-866c-9607c6489341\" (UID: \"f17fe211-3570-4b69-866c-9607c6489341\") " Dec 06 06:20:29 crc kubenswrapper[4809]: I1206 06:20:29.476960 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xpk4p\" (UniqueName: \"kubernetes.io/projected/f17fe211-3570-4b69-866c-9607c6489341-kube-api-access-xpk4p\") pod \"f17fe211-3570-4b69-866c-9607c6489341\" (UID: \"f17fe211-3570-4b69-866c-9607c6489341\") " Dec 06 06:20:29 crc kubenswrapper[4809]: I1206 06:20:29.477185 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f17fe211-3570-4b69-866c-9607c6489341-config-data\") pod \"f17fe211-3570-4b69-866c-9607c6489341\" (UID: \"f17fe211-3570-4b69-866c-9607c6489341\") " Dec 06 06:20:29 crc kubenswrapper[4809]: I1206 06:20:29.477282 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f17fe211-3570-4b69-866c-9607c6489341-logs\") pod \"f17fe211-3570-4b69-866c-9607c6489341\" (UID: \"f17fe211-3570-4b69-866c-9607c6489341\") " Dec 06 06:20:29 crc kubenswrapper[4809]: I1206 06:20:29.478656 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f17fe211-3570-4b69-866c-9607c6489341-logs" (OuterVolumeSpecName: "logs") pod "f17fe211-3570-4b69-866c-9607c6489341" (UID: "f17fe211-3570-4b69-866c-9607c6489341"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:20:29 crc kubenswrapper[4809]: I1206 06:20:29.483356 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f17fe211-3570-4b69-866c-9607c6489341-kube-api-access-xpk4p" (OuterVolumeSpecName: "kube-api-access-xpk4p") pod "f17fe211-3570-4b69-866c-9607c6489341" (UID: "f17fe211-3570-4b69-866c-9607c6489341"). InnerVolumeSpecName "kube-api-access-xpk4p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:20:29 crc kubenswrapper[4809]: I1206 06:20:29.537906 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f17fe211-3570-4b69-866c-9607c6489341-config-data" (OuterVolumeSpecName: "config-data") pod "f17fe211-3570-4b69-866c-9607c6489341" (UID: "f17fe211-3570-4b69-866c-9607c6489341"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:20:29 crc kubenswrapper[4809]: I1206 06:20:29.555199 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f17fe211-3570-4b69-866c-9607c6489341-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f17fe211-3570-4b69-866c-9607c6489341" (UID: "f17fe211-3570-4b69-866c-9607c6489341"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:20:29 crc kubenswrapper[4809]: I1206 06:20:29.579969 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xpk4p\" (UniqueName: \"kubernetes.io/projected/f17fe211-3570-4b69-866c-9607c6489341-kube-api-access-xpk4p\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:29 crc kubenswrapper[4809]: I1206 06:20:29.582427 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f17fe211-3570-4b69-866c-9607c6489341-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:29 crc kubenswrapper[4809]: I1206 06:20:29.582441 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f17fe211-3570-4b69-866c-9607c6489341-logs\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:29 crc kubenswrapper[4809]: I1206 06:20:29.582449 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f17fe211-3570-4b69-866c-9607c6489341-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:29 crc kubenswrapper[4809]: I1206 06:20:29.628453 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f17fe211-3570-4b69-866c-9607c6489341-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "f17fe211-3570-4b69-866c-9607c6489341" (UID: "f17fe211-3570-4b69-866c-9607c6489341"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:20:29 crc kubenswrapper[4809]: I1206 06:20:29.684195 4809 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f17fe211-3570-4b69-866c-9607c6489341-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:29 crc kubenswrapper[4809]: I1206 06:20:29.783960 4809 generic.go:334] "Generic (PLEG): container finished" podID="f17fe211-3570-4b69-866c-9607c6489341" containerID="9399ca5d70e9db739895e6db6766cd6ad4c43d81e9ba06135277d222b541cc75" exitCode=0 Dec 06 06:20:29 crc kubenswrapper[4809]: I1206 06:20:29.784048 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 06:20:29 crc kubenswrapper[4809]: I1206 06:20:29.784077 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f17fe211-3570-4b69-866c-9607c6489341","Type":"ContainerDied","Data":"9399ca5d70e9db739895e6db6766cd6ad4c43d81e9ba06135277d222b541cc75"} Dec 06 06:20:29 crc kubenswrapper[4809]: I1206 06:20:29.784104 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f17fe211-3570-4b69-866c-9607c6489341","Type":"ContainerDied","Data":"87bc3a89585fba86c856829307710e58092b7643b3cecb821d9055b4a507c3a6"} Dec 06 06:20:29 crc kubenswrapper[4809]: I1206 06:20:29.784122 4809 scope.go:117] "RemoveContainer" containerID="9399ca5d70e9db739895e6db6766cd6ad4c43d81e9ba06135277d222b541cc75" Dec 06 06:20:29 crc kubenswrapper[4809]: I1206 06:20:29.811232 4809 scope.go:117] "RemoveContainer" containerID="5375b409ca2dbb8cc53e5c42961b8fbe2e4b34d2ddaf260aae13f98079f6fffd" Dec 06 06:20:29 crc kubenswrapper[4809]: I1206 06:20:29.846094 4809 scope.go:117] "RemoveContainer" containerID="9399ca5d70e9db739895e6db6766cd6ad4c43d81e9ba06135277d222b541cc75" Dec 06 06:20:29 crc kubenswrapper[4809]: E1206 06:20:29.846569 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9399ca5d70e9db739895e6db6766cd6ad4c43d81e9ba06135277d222b541cc75\": container with ID starting with 9399ca5d70e9db739895e6db6766cd6ad4c43d81e9ba06135277d222b541cc75 not found: ID does not exist" containerID="9399ca5d70e9db739895e6db6766cd6ad4c43d81e9ba06135277d222b541cc75" Dec 06 06:20:29 crc kubenswrapper[4809]: I1206 06:20:29.846605 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9399ca5d70e9db739895e6db6766cd6ad4c43d81e9ba06135277d222b541cc75"} err="failed to get container status \"9399ca5d70e9db739895e6db6766cd6ad4c43d81e9ba06135277d222b541cc75\": rpc error: code = NotFound desc = could not find container \"9399ca5d70e9db739895e6db6766cd6ad4c43d81e9ba06135277d222b541cc75\": container with ID starting with 9399ca5d70e9db739895e6db6766cd6ad4c43d81e9ba06135277d222b541cc75 not found: ID does not exist" Dec 06 06:20:29 crc kubenswrapper[4809]: I1206 06:20:29.846630 4809 scope.go:117] "RemoveContainer" containerID="5375b409ca2dbb8cc53e5c42961b8fbe2e4b34d2ddaf260aae13f98079f6fffd" Dec 06 06:20:29 crc kubenswrapper[4809]: E1206 06:20:29.846845 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5375b409ca2dbb8cc53e5c42961b8fbe2e4b34d2ddaf260aae13f98079f6fffd\": container with ID starting with 5375b409ca2dbb8cc53e5c42961b8fbe2e4b34d2ddaf260aae13f98079f6fffd not found: ID does not exist" containerID="5375b409ca2dbb8cc53e5c42961b8fbe2e4b34d2ddaf260aae13f98079f6fffd" Dec 06 06:20:29 crc kubenswrapper[4809]: I1206 06:20:29.846869 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5375b409ca2dbb8cc53e5c42961b8fbe2e4b34d2ddaf260aae13f98079f6fffd"} err="failed to get container status \"5375b409ca2dbb8cc53e5c42961b8fbe2e4b34d2ddaf260aae13f98079f6fffd\": rpc error: code = NotFound desc = could not find container \"5375b409ca2dbb8cc53e5c42961b8fbe2e4b34d2ddaf260aae13f98079f6fffd\": container with ID starting with 5375b409ca2dbb8cc53e5c42961b8fbe2e4b34d2ddaf260aae13f98079f6fffd not found: ID does not exist" Dec 06 06:20:29 crc kubenswrapper[4809]: I1206 06:20:29.850988 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 06:20:29 crc kubenswrapper[4809]: I1206 06:20:29.880139 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 06:20:29 crc kubenswrapper[4809]: I1206 06:20:29.904989 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 06 06:20:29 crc kubenswrapper[4809]: E1206 06:20:29.905581 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f17fe211-3570-4b69-866c-9607c6489341" containerName="nova-metadata-log" Dec 06 06:20:29 crc kubenswrapper[4809]: I1206 06:20:29.905606 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f17fe211-3570-4b69-866c-9607c6489341" containerName="nova-metadata-log" Dec 06 06:20:29 crc kubenswrapper[4809]: E1206 06:20:29.905650 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f17fe211-3570-4b69-866c-9607c6489341" containerName="nova-metadata-metadata" Dec 06 06:20:29 crc kubenswrapper[4809]: I1206 06:20:29.905657 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f17fe211-3570-4b69-866c-9607c6489341" containerName="nova-metadata-metadata" Dec 06 06:20:29 crc kubenswrapper[4809]: I1206 06:20:29.905880 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f17fe211-3570-4b69-866c-9607c6489341" containerName="nova-metadata-log" Dec 06 06:20:29 crc kubenswrapper[4809]: I1206 06:20:29.905898 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f17fe211-3570-4b69-866c-9607c6489341" containerName="nova-metadata-metadata" Dec 06 06:20:29 crc kubenswrapper[4809]: I1206 06:20:29.907287 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 06:20:29 crc kubenswrapper[4809]: I1206 06:20:29.911240 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 06 06:20:29 crc kubenswrapper[4809]: I1206 06:20:29.914246 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 06 06:20:29 crc kubenswrapper[4809]: I1206 06:20:29.919864 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 06:20:29 crc kubenswrapper[4809]: I1206 06:20:29.991125 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0909590-844d-41ec-9443-5caacd682230-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e0909590-844d-41ec-9443-5caacd682230\") " pod="openstack/nova-metadata-0" Dec 06 06:20:29 crc kubenswrapper[4809]: I1206 06:20:29.991182 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0909590-844d-41ec-9443-5caacd682230-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e0909590-844d-41ec-9443-5caacd682230\") " pod="openstack/nova-metadata-0" Dec 06 06:20:29 crc kubenswrapper[4809]: I1206 06:20:29.991254 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8nw4\" (UniqueName: \"kubernetes.io/projected/e0909590-844d-41ec-9443-5caacd682230-kube-api-access-p8nw4\") pod \"nova-metadata-0\" (UID: \"e0909590-844d-41ec-9443-5caacd682230\") " pod="openstack/nova-metadata-0" Dec 06 06:20:29 crc kubenswrapper[4809]: I1206 06:20:29.991286 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0909590-844d-41ec-9443-5caacd682230-config-data\") pod \"nova-metadata-0\" (UID: \"e0909590-844d-41ec-9443-5caacd682230\") " pod="openstack/nova-metadata-0" Dec 06 06:20:29 crc kubenswrapper[4809]: I1206 06:20:29.991570 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0909590-844d-41ec-9443-5caacd682230-logs\") pod \"nova-metadata-0\" (UID: \"e0909590-844d-41ec-9443-5caacd682230\") " pod="openstack/nova-metadata-0" Dec 06 06:20:30 crc kubenswrapper[4809]: I1206 06:20:30.094315 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0909590-844d-41ec-9443-5caacd682230-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e0909590-844d-41ec-9443-5caacd682230\") " pod="openstack/nova-metadata-0" Dec 06 06:20:30 crc kubenswrapper[4809]: I1206 06:20:30.094358 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0909590-844d-41ec-9443-5caacd682230-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e0909590-844d-41ec-9443-5caacd682230\") " pod="openstack/nova-metadata-0" Dec 06 06:20:30 crc kubenswrapper[4809]: I1206 06:20:30.094397 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8nw4\" (UniqueName: \"kubernetes.io/projected/e0909590-844d-41ec-9443-5caacd682230-kube-api-access-p8nw4\") pod \"nova-metadata-0\" (UID: \"e0909590-844d-41ec-9443-5caacd682230\") " pod="openstack/nova-metadata-0" Dec 06 06:20:30 crc kubenswrapper[4809]: I1206 06:20:30.094423 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0909590-844d-41ec-9443-5caacd682230-config-data\") pod \"nova-metadata-0\" (UID: \"e0909590-844d-41ec-9443-5caacd682230\") " pod="openstack/nova-metadata-0" Dec 06 06:20:30 crc kubenswrapper[4809]: I1206 06:20:30.094535 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0909590-844d-41ec-9443-5caacd682230-logs\") pod \"nova-metadata-0\" (UID: \"e0909590-844d-41ec-9443-5caacd682230\") " pod="openstack/nova-metadata-0" Dec 06 06:20:30 crc kubenswrapper[4809]: I1206 06:20:30.095093 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0909590-844d-41ec-9443-5caacd682230-logs\") pod \"nova-metadata-0\" (UID: \"e0909590-844d-41ec-9443-5caacd682230\") " pod="openstack/nova-metadata-0" Dec 06 06:20:30 crc kubenswrapper[4809]: I1206 06:20:30.098495 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0909590-844d-41ec-9443-5caacd682230-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e0909590-844d-41ec-9443-5caacd682230\") " pod="openstack/nova-metadata-0" Dec 06 06:20:30 crc kubenswrapper[4809]: I1206 06:20:30.099414 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0909590-844d-41ec-9443-5caacd682230-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e0909590-844d-41ec-9443-5caacd682230\") " pod="openstack/nova-metadata-0" Dec 06 06:20:30 crc kubenswrapper[4809]: I1206 06:20:30.106753 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0909590-844d-41ec-9443-5caacd682230-config-data\") pod \"nova-metadata-0\" (UID: \"e0909590-844d-41ec-9443-5caacd682230\") " pod="openstack/nova-metadata-0" Dec 06 06:20:30 crc kubenswrapper[4809]: I1206 06:20:30.115556 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8nw4\" (UniqueName: \"kubernetes.io/projected/e0909590-844d-41ec-9443-5caacd682230-kube-api-access-p8nw4\") pod \"nova-metadata-0\" (UID: \"e0909590-844d-41ec-9443-5caacd682230\") " pod="openstack/nova-metadata-0" Dec 06 06:20:30 crc kubenswrapper[4809]: I1206 06:20:30.276698 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 06:20:30 crc kubenswrapper[4809]: I1206 06:20:30.388589 4809 scope.go:117] "RemoveContainer" containerID="9bc2dc88b3b9f309ce0169aa510326dd77022dece4c69cab3141df00fe5c0106" Dec 06 06:20:30 crc kubenswrapper[4809]: E1206 06:20:30.389213 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:20:30 crc kubenswrapper[4809]: I1206 06:20:30.922155 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 06:20:31 crc kubenswrapper[4809]: I1206 06:20:31.274607 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 06 06:20:31 crc kubenswrapper[4809]: I1206 06:20:31.414785 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f17fe211-3570-4b69-866c-9607c6489341" path="/var/lib/kubelet/pods/f17fe211-3570-4b69-866c-9607c6489341/volumes" Dec 06 06:20:31 crc kubenswrapper[4809]: I1206 06:20:31.829064 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e0909590-844d-41ec-9443-5caacd682230","Type":"ContainerStarted","Data":"76e0b9b1405db08480044440fc46ac91ee26accb953e1c1e13f105a6500cb68b"} Dec 06 06:20:31 crc kubenswrapper[4809]: I1206 06:20:31.829140 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e0909590-844d-41ec-9443-5caacd682230","Type":"ContainerStarted","Data":"11298ab5a07ebddd66d3e0198b47487d783f566279278bb88a4941517b968884"} Dec 06 06:20:31 crc kubenswrapper[4809]: I1206 06:20:31.829151 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e0909590-844d-41ec-9443-5caacd682230","Type":"ContainerStarted","Data":"3e76d8178acaa90e73685b688fbd46140a04eda03fbd7547a1bebd2057347341"} Dec 06 06:20:31 crc kubenswrapper[4809]: I1206 06:20:31.856590 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.856574051 podStartE2EDuration="2.856574051s" podCreationTimestamp="2025-12-06 06:20:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:20:31.853278052 +0000 UTC m=+1756.742260994" watchObservedRunningTime="2025-12-06 06:20:31.856574051 +0000 UTC m=+1756.745556993" Dec 06 06:20:35 crc kubenswrapper[4809]: I1206 06:20:35.277788 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 06 06:20:35 crc kubenswrapper[4809]: I1206 06:20:35.279080 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 06 06:20:36 crc kubenswrapper[4809]: I1206 06:20:36.275171 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 06 06:20:36 crc kubenswrapper[4809]: I1206 06:20:36.314018 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 06 06:20:36 crc kubenswrapper[4809]: I1206 06:20:36.941567 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 06 06:20:40 crc kubenswrapper[4809]: I1206 06:20:40.280512 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 06 06:20:40 crc kubenswrapper[4809]: I1206 06:20:40.281073 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 06 06:20:41 crc kubenswrapper[4809]: I1206 06:20:41.295248 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="e0909590-844d-41ec-9443-5caacd682230" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.0:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 06:20:41 crc kubenswrapper[4809]: I1206 06:20:41.295235 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="e0909590-844d-41ec-9443-5caacd682230" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.0:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 06:20:43 crc kubenswrapper[4809]: I1206 06:20:43.557344 4809 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod215a53e4-4135-45b5-912c-e25d50e5c2d3"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod215a53e4-4135-45b5-912c-e25d50e5c2d3] : Timed out while waiting for systemd to remove kubepods-besteffort-pod215a53e4_4135_45b5_912c_e25d50e5c2d3.slice" Dec 06 06:20:43 crc kubenswrapper[4809]: E1206 06:20:43.557430 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort pod215a53e4-4135-45b5-912c-e25d50e5c2d3] : unable to destroy cgroup paths for cgroup [kubepods besteffort pod215a53e4-4135-45b5-912c-e25d50e5c2d3] : Timed out while waiting for systemd to remove kubepods-besteffort-pod215a53e4_4135_45b5_912c_e25d50e5c2d3.slice" pod="openstack/nova-api-0" podUID="215a53e4-4135-45b5-912c-e25d50e5c2d3" Dec 06 06:20:43 crc kubenswrapper[4809]: I1206 06:20:43.970756 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 06:20:44 crc kubenswrapper[4809]: I1206 06:20:44.011774 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 06 06:20:44 crc kubenswrapper[4809]: I1206 06:20:44.024273 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 06 06:20:44 crc kubenswrapper[4809]: I1206 06:20:44.049133 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 06 06:20:44 crc kubenswrapper[4809]: I1206 06:20:44.053155 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 06:20:44 crc kubenswrapper[4809]: I1206 06:20:44.055849 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 06 06:20:44 crc kubenswrapper[4809]: I1206 06:20:44.056136 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 06 06:20:44 crc kubenswrapper[4809]: I1206 06:20:44.057175 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 06 06:20:44 crc kubenswrapper[4809]: I1206 06:20:44.064606 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 06:20:44 crc kubenswrapper[4809]: I1206 06:20:44.176242 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61c6cc41-4c7d-422e-987c-e05e77ae463d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"61c6cc41-4c7d-422e-987c-e05e77ae463d\") " pod="openstack/nova-api-0" Dec 06 06:20:44 crc kubenswrapper[4809]: I1206 06:20:44.177023 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/61c6cc41-4c7d-422e-987c-e05e77ae463d-public-tls-certs\") pod \"nova-api-0\" (UID: \"61c6cc41-4c7d-422e-987c-e05e77ae463d\") " pod="openstack/nova-api-0" Dec 06 06:20:44 crc kubenswrapper[4809]: I1206 06:20:44.177171 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61c6cc41-4c7d-422e-987c-e05e77ae463d-logs\") pod \"nova-api-0\" (UID: \"61c6cc41-4c7d-422e-987c-e05e77ae463d\") " pod="openstack/nova-api-0" Dec 06 06:20:44 crc kubenswrapper[4809]: I1206 06:20:44.177308 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61c6cc41-4c7d-422e-987c-e05e77ae463d-config-data\") pod \"nova-api-0\" (UID: \"61c6cc41-4c7d-422e-987c-e05e77ae463d\") " pod="openstack/nova-api-0" Dec 06 06:20:44 crc kubenswrapper[4809]: I1206 06:20:44.177417 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crtrt\" (UniqueName: \"kubernetes.io/projected/61c6cc41-4c7d-422e-987c-e05e77ae463d-kube-api-access-crtrt\") pod \"nova-api-0\" (UID: \"61c6cc41-4c7d-422e-987c-e05e77ae463d\") " pod="openstack/nova-api-0" Dec 06 06:20:44 crc kubenswrapper[4809]: I1206 06:20:44.177550 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/61c6cc41-4c7d-422e-987c-e05e77ae463d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"61c6cc41-4c7d-422e-987c-e05e77ae463d\") " pod="openstack/nova-api-0" Dec 06 06:20:44 crc kubenswrapper[4809]: I1206 06:20:44.279989 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61c6cc41-4c7d-422e-987c-e05e77ae463d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"61c6cc41-4c7d-422e-987c-e05e77ae463d\") " pod="openstack/nova-api-0" Dec 06 06:20:44 crc kubenswrapper[4809]: I1206 06:20:44.280147 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/61c6cc41-4c7d-422e-987c-e05e77ae463d-public-tls-certs\") pod \"nova-api-0\" (UID: \"61c6cc41-4c7d-422e-987c-e05e77ae463d\") " pod="openstack/nova-api-0" Dec 06 06:20:44 crc kubenswrapper[4809]: I1206 06:20:44.280176 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61c6cc41-4c7d-422e-987c-e05e77ae463d-logs\") pod \"nova-api-0\" (UID: \"61c6cc41-4c7d-422e-987c-e05e77ae463d\") " pod="openstack/nova-api-0" Dec 06 06:20:44 crc kubenswrapper[4809]: I1206 06:20:44.280202 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61c6cc41-4c7d-422e-987c-e05e77ae463d-config-data\") pod \"nova-api-0\" (UID: \"61c6cc41-4c7d-422e-987c-e05e77ae463d\") " pod="openstack/nova-api-0" Dec 06 06:20:44 crc kubenswrapper[4809]: I1206 06:20:44.280226 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crtrt\" (UniqueName: \"kubernetes.io/projected/61c6cc41-4c7d-422e-987c-e05e77ae463d-kube-api-access-crtrt\") pod \"nova-api-0\" (UID: \"61c6cc41-4c7d-422e-987c-e05e77ae463d\") " pod="openstack/nova-api-0" Dec 06 06:20:44 crc kubenswrapper[4809]: I1206 06:20:44.280255 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/61c6cc41-4c7d-422e-987c-e05e77ae463d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"61c6cc41-4c7d-422e-987c-e05e77ae463d\") " pod="openstack/nova-api-0" Dec 06 06:20:44 crc kubenswrapper[4809]: I1206 06:20:44.280919 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61c6cc41-4c7d-422e-987c-e05e77ae463d-logs\") pod \"nova-api-0\" (UID: \"61c6cc41-4c7d-422e-987c-e05e77ae463d\") " pod="openstack/nova-api-0" Dec 06 06:20:44 crc kubenswrapper[4809]: I1206 06:20:44.294154 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61c6cc41-4c7d-422e-987c-e05e77ae463d-config-data\") pod \"nova-api-0\" (UID: \"61c6cc41-4c7d-422e-987c-e05e77ae463d\") " pod="openstack/nova-api-0" Dec 06 06:20:44 crc kubenswrapper[4809]: I1206 06:20:44.294461 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/61c6cc41-4c7d-422e-987c-e05e77ae463d-public-tls-certs\") pod \"nova-api-0\" (UID: \"61c6cc41-4c7d-422e-987c-e05e77ae463d\") " pod="openstack/nova-api-0" Dec 06 06:20:44 crc kubenswrapper[4809]: I1206 06:20:44.294538 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/61c6cc41-4c7d-422e-987c-e05e77ae463d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"61c6cc41-4c7d-422e-987c-e05e77ae463d\") " pod="openstack/nova-api-0" Dec 06 06:20:44 crc kubenswrapper[4809]: I1206 06:20:44.296744 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61c6cc41-4c7d-422e-987c-e05e77ae463d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"61c6cc41-4c7d-422e-987c-e05e77ae463d\") " pod="openstack/nova-api-0" Dec 06 06:20:44 crc kubenswrapper[4809]: I1206 06:20:44.297017 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crtrt\" (UniqueName: \"kubernetes.io/projected/61c6cc41-4c7d-422e-987c-e05e77ae463d-kube-api-access-crtrt\") pod \"nova-api-0\" (UID: \"61c6cc41-4c7d-422e-987c-e05e77ae463d\") " pod="openstack/nova-api-0" Dec 06 06:20:44 crc kubenswrapper[4809]: I1206 06:20:44.388977 4809 scope.go:117] "RemoveContainer" containerID="9bc2dc88b3b9f309ce0169aa510326dd77022dece4c69cab3141df00fe5c0106" Dec 06 06:20:44 crc kubenswrapper[4809]: E1206 06:20:44.389372 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:20:44 crc kubenswrapper[4809]: I1206 06:20:44.389646 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 06:20:44 crc kubenswrapper[4809]: W1206 06:20:44.959539 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod61c6cc41_4c7d_422e_987c_e05e77ae463d.slice/crio-fe03fd39202d5b5cde681fe2dc725866c0e4c2df5e1a51f8b12486e3da069ced WatchSource:0}: Error finding container fe03fd39202d5b5cde681fe2dc725866c0e4c2df5e1a51f8b12486e3da069ced: Status 404 returned error can't find the container with id fe03fd39202d5b5cde681fe2dc725866c0e4c2df5e1a51f8b12486e3da069ced Dec 06 06:20:44 crc kubenswrapper[4809]: I1206 06:20:44.962309 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 06:20:44 crc kubenswrapper[4809]: I1206 06:20:44.987491 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"61c6cc41-4c7d-422e-987c-e05e77ae463d","Type":"ContainerStarted","Data":"fe03fd39202d5b5cde681fe2dc725866c0e4c2df5e1a51f8b12486e3da069ced"} Dec 06 06:20:45 crc kubenswrapper[4809]: I1206 06:20:45.403648 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="215a53e4-4135-45b5-912c-e25d50e5c2d3" path="/var/lib/kubelet/pods/215a53e4-4135-45b5-912c-e25d50e5c2d3/volumes" Dec 06 06:20:45 crc kubenswrapper[4809]: I1206 06:20:45.999563 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"61c6cc41-4c7d-422e-987c-e05e77ae463d","Type":"ContainerStarted","Data":"1a3b7da085699512889ffb94aa1d0a5debc67d06228f239c784580c691860550"} Dec 06 06:20:45 crc kubenswrapper[4809]: I1206 06:20:45.999603 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"61c6cc41-4c7d-422e-987c-e05e77ae463d","Type":"ContainerStarted","Data":"b6fd73aeaba3c81bc99dbc76db9f39dab29377897a98a4c880ea477f51182d88"} Dec 06 06:20:46 crc kubenswrapper[4809]: I1206 06:20:46.024884 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.024865007 podStartE2EDuration="2.024865007s" podCreationTimestamp="2025-12-06 06:20:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:20:46.018424353 +0000 UTC m=+1770.907407285" watchObservedRunningTime="2025-12-06 06:20:46.024865007 +0000 UTC m=+1770.913847939" Dec 06 06:20:50 crc kubenswrapper[4809]: I1206 06:20:50.282170 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 06 06:20:50 crc kubenswrapper[4809]: I1206 06:20:50.282961 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 06 06:20:50 crc kubenswrapper[4809]: I1206 06:20:50.288236 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 06 06:20:50 crc kubenswrapper[4809]: I1206 06:20:50.292612 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 06 06:20:50 crc kubenswrapper[4809]: I1206 06:20:50.293734 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 06 06:20:54 crc kubenswrapper[4809]: I1206 06:20:54.018710 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 06:20:54 crc kubenswrapper[4809]: I1206 06:20:54.019380 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="cdf93098-585a-4391-8419-c5bc7ab1c567" containerName="kube-state-metrics" containerID="cri-o://28674d3c480d379e5c1c3169207c00292adf403404c050cd9ae2bb9fcdc42dc9" gracePeriod=30 Dec 06 06:20:54 crc kubenswrapper[4809]: I1206 06:20:54.174753 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 06 06:20:54 crc kubenswrapper[4809]: I1206 06:20:54.179180 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/mysqld-exporter-0" podUID="3563a545-b965-4c74-a183-9f8da5876aee" containerName="mysqld-exporter" containerID="cri-o://2416c7e9cd8e39fe724afee7991341b25c4e5b7738b872ea62c8a977a8721b0c" gracePeriod=30 Dec 06 06:20:54 crc kubenswrapper[4809]: I1206 06:20:54.394256 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 06:20:54 crc kubenswrapper[4809]: I1206 06:20:54.394954 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 06:20:54 crc kubenswrapper[4809]: I1206 06:20:54.733240 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 06 06:20:54 crc kubenswrapper[4809]: I1206 06:20:54.852500 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85dpb\" (UniqueName: \"kubernetes.io/projected/cdf93098-585a-4391-8419-c5bc7ab1c567-kube-api-access-85dpb\") pod \"cdf93098-585a-4391-8419-c5bc7ab1c567\" (UID: \"cdf93098-585a-4391-8419-c5bc7ab1c567\") " Dec 06 06:20:54 crc kubenswrapper[4809]: I1206 06:20:54.864517 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdf93098-585a-4391-8419-c5bc7ab1c567-kube-api-access-85dpb" (OuterVolumeSpecName: "kube-api-access-85dpb") pod "cdf93098-585a-4391-8419-c5bc7ab1c567" (UID: "cdf93098-585a-4391-8419-c5bc7ab1c567"). InnerVolumeSpecName "kube-api-access-85dpb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:20:54 crc kubenswrapper[4809]: I1206 06:20:54.928571 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 06 06:20:54 crc kubenswrapper[4809]: I1206 06:20:54.955026 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3563a545-b965-4c74-a183-9f8da5876aee-combined-ca-bundle\") pod \"3563a545-b965-4c74-a183-9f8da5876aee\" (UID: \"3563a545-b965-4c74-a183-9f8da5876aee\") " Dec 06 06:20:54 crc kubenswrapper[4809]: I1206 06:20:54.955422 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3563a545-b965-4c74-a183-9f8da5876aee-config-data\") pod \"3563a545-b965-4c74-a183-9f8da5876aee\" (UID: \"3563a545-b965-4c74-a183-9f8da5876aee\") " Dec 06 06:20:54 crc kubenswrapper[4809]: I1206 06:20:54.955608 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lptf5\" (UniqueName: \"kubernetes.io/projected/3563a545-b965-4c74-a183-9f8da5876aee-kube-api-access-lptf5\") pod \"3563a545-b965-4c74-a183-9f8da5876aee\" (UID: \"3563a545-b965-4c74-a183-9f8da5876aee\") " Dec 06 06:20:54 crc kubenswrapper[4809]: I1206 06:20:54.957026 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85dpb\" (UniqueName: \"kubernetes.io/projected/cdf93098-585a-4391-8419-c5bc7ab1c567-kube-api-access-85dpb\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:54 crc kubenswrapper[4809]: I1206 06:20:54.958680 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3563a545-b965-4c74-a183-9f8da5876aee-kube-api-access-lptf5" (OuterVolumeSpecName: "kube-api-access-lptf5") pod "3563a545-b965-4c74-a183-9f8da5876aee" (UID: "3563a545-b965-4c74-a183-9f8da5876aee"). InnerVolumeSpecName "kube-api-access-lptf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.005238 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3563a545-b965-4c74-a183-9f8da5876aee-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3563a545-b965-4c74-a183-9f8da5876aee" (UID: "3563a545-b965-4c74-a183-9f8da5876aee"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.041777 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3563a545-b965-4c74-a183-9f8da5876aee-config-data" (OuterVolumeSpecName: "config-data") pod "3563a545-b965-4c74-a183-9f8da5876aee" (UID: "3563a545-b965-4c74-a183-9f8da5876aee"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.059056 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3563a545-b965-4c74-a183-9f8da5876aee-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.059089 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3563a545-b965-4c74-a183-9f8da5876aee-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.059099 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lptf5\" (UniqueName: \"kubernetes.io/projected/3563a545-b965-4c74-a183-9f8da5876aee-kube-api-access-lptf5\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.092841 4809 generic.go:334] "Generic (PLEG): container finished" podID="3563a545-b965-4c74-a183-9f8da5876aee" containerID="2416c7e9cd8e39fe724afee7991341b25c4e5b7738b872ea62c8a977a8721b0c" exitCode=2 Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.092899 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"3563a545-b965-4c74-a183-9f8da5876aee","Type":"ContainerDied","Data":"2416c7e9cd8e39fe724afee7991341b25c4e5b7738b872ea62c8a977a8721b0c"} Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.093027 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.094166 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"3563a545-b965-4c74-a183-9f8da5876aee","Type":"ContainerDied","Data":"3a302852a8edd1b50ba0f27dab81fb73d1cf4525ada49dec4f7accf4058a1d9b"} Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.094198 4809 scope.go:117] "RemoveContainer" containerID="2416c7e9cd8e39fe724afee7991341b25c4e5b7738b872ea62c8a977a8721b0c" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.094799 4809 generic.go:334] "Generic (PLEG): container finished" podID="cdf93098-585a-4391-8419-c5bc7ab1c567" containerID="28674d3c480d379e5c1c3169207c00292adf403404c050cd9ae2bb9fcdc42dc9" exitCode=2 Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.094837 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"cdf93098-585a-4391-8419-c5bc7ab1c567","Type":"ContainerDied","Data":"28674d3c480d379e5c1c3169207c00292adf403404c050cd9ae2bb9fcdc42dc9"} Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.094862 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.094872 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"cdf93098-585a-4391-8419-c5bc7ab1c567","Type":"ContainerDied","Data":"66f5db3f29fbb91d8689d12339457c220d225d60ff1721cc5985909e88e6d5d1"} Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.142894 4809 scope.go:117] "RemoveContainer" containerID="2416c7e9cd8e39fe724afee7991341b25c4e5b7738b872ea62c8a977a8721b0c" Dec 06 06:20:55 crc kubenswrapper[4809]: E1206 06:20:55.152024 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2416c7e9cd8e39fe724afee7991341b25c4e5b7738b872ea62c8a977a8721b0c\": container with ID starting with 2416c7e9cd8e39fe724afee7991341b25c4e5b7738b872ea62c8a977a8721b0c not found: ID does not exist" containerID="2416c7e9cd8e39fe724afee7991341b25c4e5b7738b872ea62c8a977a8721b0c" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.152086 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2416c7e9cd8e39fe724afee7991341b25c4e5b7738b872ea62c8a977a8721b0c"} err="failed to get container status \"2416c7e9cd8e39fe724afee7991341b25c4e5b7738b872ea62c8a977a8721b0c\": rpc error: code = NotFound desc = could not find container \"2416c7e9cd8e39fe724afee7991341b25c4e5b7738b872ea62c8a977a8721b0c\": container with ID starting with 2416c7e9cd8e39fe724afee7991341b25c4e5b7738b872ea62c8a977a8721b0c not found: ID does not exist" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.152117 4809 scope.go:117] "RemoveContainer" containerID="28674d3c480d379e5c1c3169207c00292adf403404c050cd9ae2bb9fcdc42dc9" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.185095 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.222672 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.238214 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-0"] Dec 06 06:20:55 crc kubenswrapper[4809]: E1206 06:20:55.238900 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3563a545-b965-4c74-a183-9f8da5876aee" containerName="mysqld-exporter" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.238987 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3563a545-b965-4c74-a183-9f8da5876aee" containerName="mysqld-exporter" Dec 06 06:20:55 crc kubenswrapper[4809]: E1206 06:20:55.239032 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdf93098-585a-4391-8419-c5bc7ab1c567" containerName="kube-state-metrics" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.239039 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdf93098-585a-4391-8419-c5bc7ab1c567" containerName="kube-state-metrics" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.239277 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="3563a545-b965-4c74-a183-9f8da5876aee" containerName="mysqld-exporter" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.239316 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdf93098-585a-4391-8419-c5bc7ab1c567" containerName="kube-state-metrics" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.239427 4809 scope.go:117] "RemoveContainer" containerID="28674d3c480d379e5c1c3169207c00292adf403404c050cd9ae2bb9fcdc42dc9" Dec 06 06:20:55 crc kubenswrapper[4809]: E1206 06:20:55.240073 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28674d3c480d379e5c1c3169207c00292adf403404c050cd9ae2bb9fcdc42dc9\": container with ID starting with 28674d3c480d379e5c1c3169207c00292adf403404c050cd9ae2bb9fcdc42dc9 not found: ID does not exist" containerID="28674d3c480d379e5c1c3169207c00292adf403404c050cd9ae2bb9fcdc42dc9" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.240103 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28674d3c480d379e5c1c3169207c00292adf403404c050cd9ae2bb9fcdc42dc9"} err="failed to get container status \"28674d3c480d379e5c1c3169207c00292adf403404c050cd9ae2bb9fcdc42dc9\": rpc error: code = NotFound desc = could not find container \"28674d3c480d379e5c1c3169207c00292adf403404c050cd9ae2bb9fcdc42dc9\": container with ID starting with 28674d3c480d379e5c1c3169207c00292adf403404c050cd9ae2bb9fcdc42dc9 not found: ID does not exist" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.240367 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.244302 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-mysqld-exporter-svc" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.244334 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-config-data" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.265287 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.279398 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.290370 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.301762 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.303261 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.306480 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.308593 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.316962 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.370282 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/36f27192-451f-4ce2-afda-c0e66a6028b2-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"36f27192-451f-4ce2-afda-c0e66a6028b2\") " pod="openstack/kube-state-metrics-0" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.370802 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4b85ab2-e407-4779-bcc8-4ea2d35aee72-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"f4b85ab2-e407-4779-bcc8-4ea2d35aee72\") " pod="openstack/mysqld-exporter-0" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.370839 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlv2b\" (UniqueName: \"kubernetes.io/projected/36f27192-451f-4ce2-afda-c0e66a6028b2-kube-api-access-mlv2b\") pod \"kube-state-metrics-0\" (UID: \"36f27192-451f-4ce2-afda-c0e66a6028b2\") " pod="openstack/kube-state-metrics-0" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.370899 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2spf\" (UniqueName: \"kubernetes.io/projected/f4b85ab2-e407-4779-bcc8-4ea2d35aee72-kube-api-access-z2spf\") pod \"mysqld-exporter-0\" (UID: \"f4b85ab2-e407-4779-bcc8-4ea2d35aee72\") " pod="openstack/mysqld-exporter-0" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.371013 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4b85ab2-e407-4779-bcc8-4ea2d35aee72-config-data\") pod \"mysqld-exporter-0\" (UID: \"f4b85ab2-e407-4779-bcc8-4ea2d35aee72\") " pod="openstack/mysqld-exporter-0" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.371104 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36f27192-451f-4ce2-afda-c0e66a6028b2-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"36f27192-451f-4ce2-afda-c0e66a6028b2\") " pod="openstack/kube-state-metrics-0" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.371199 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4b85ab2-e407-4779-bcc8-4ea2d35aee72-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"f4b85ab2-e407-4779-bcc8-4ea2d35aee72\") " pod="openstack/mysqld-exporter-0" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.371333 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/36f27192-451f-4ce2-afda-c0e66a6028b2-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"36f27192-451f-4ce2-afda-c0e66a6028b2\") " pod="openstack/kube-state-metrics-0" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.405467 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3563a545-b965-4c74-a183-9f8da5876aee" path="/var/lib/kubelet/pods/3563a545-b965-4c74-a183-9f8da5876aee/volumes" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.406262 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cdf93098-585a-4391-8419-c5bc7ab1c567" path="/var/lib/kubelet/pods/cdf93098-585a-4391-8419-c5bc7ab1c567/volumes" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.419184 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="61c6cc41-4c7d-422e-987c-e05e77ae463d" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.1.1:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.419202 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="61c6cc41-4c7d-422e-987c-e05e77ae463d" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.1.1:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.472788 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4b85ab2-e407-4779-bcc8-4ea2d35aee72-config-data\") pod \"mysqld-exporter-0\" (UID: \"f4b85ab2-e407-4779-bcc8-4ea2d35aee72\") " pod="openstack/mysqld-exporter-0" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.472883 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36f27192-451f-4ce2-afda-c0e66a6028b2-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"36f27192-451f-4ce2-afda-c0e66a6028b2\") " pod="openstack/kube-state-metrics-0" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.472914 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4b85ab2-e407-4779-bcc8-4ea2d35aee72-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"f4b85ab2-e407-4779-bcc8-4ea2d35aee72\") " pod="openstack/mysqld-exporter-0" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.473019 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/36f27192-451f-4ce2-afda-c0e66a6028b2-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"36f27192-451f-4ce2-afda-c0e66a6028b2\") " pod="openstack/kube-state-metrics-0" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.473063 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/36f27192-451f-4ce2-afda-c0e66a6028b2-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"36f27192-451f-4ce2-afda-c0e66a6028b2\") " pod="openstack/kube-state-metrics-0" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.473131 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlv2b\" (UniqueName: \"kubernetes.io/projected/36f27192-451f-4ce2-afda-c0e66a6028b2-kube-api-access-mlv2b\") pod \"kube-state-metrics-0\" (UID: \"36f27192-451f-4ce2-afda-c0e66a6028b2\") " pod="openstack/kube-state-metrics-0" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.473146 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4b85ab2-e407-4779-bcc8-4ea2d35aee72-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"f4b85ab2-e407-4779-bcc8-4ea2d35aee72\") " pod="openstack/mysqld-exporter-0" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.473179 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2spf\" (UniqueName: \"kubernetes.io/projected/f4b85ab2-e407-4779-bcc8-4ea2d35aee72-kube-api-access-z2spf\") pod \"mysqld-exporter-0\" (UID: \"f4b85ab2-e407-4779-bcc8-4ea2d35aee72\") " pod="openstack/mysqld-exporter-0" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.481413 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4b85ab2-e407-4779-bcc8-4ea2d35aee72-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"f4b85ab2-e407-4779-bcc8-4ea2d35aee72\") " pod="openstack/mysqld-exporter-0" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.481426 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4b85ab2-e407-4779-bcc8-4ea2d35aee72-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"f4b85ab2-e407-4779-bcc8-4ea2d35aee72\") " pod="openstack/mysqld-exporter-0" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.481536 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/36f27192-451f-4ce2-afda-c0e66a6028b2-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"36f27192-451f-4ce2-afda-c0e66a6028b2\") " pod="openstack/kube-state-metrics-0" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.481963 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36f27192-451f-4ce2-afda-c0e66a6028b2-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"36f27192-451f-4ce2-afda-c0e66a6028b2\") " pod="openstack/kube-state-metrics-0" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.483545 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4b85ab2-e407-4779-bcc8-4ea2d35aee72-config-data\") pod \"mysqld-exporter-0\" (UID: \"f4b85ab2-e407-4779-bcc8-4ea2d35aee72\") " pod="openstack/mysqld-exporter-0" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.484861 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/36f27192-451f-4ce2-afda-c0e66a6028b2-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"36f27192-451f-4ce2-afda-c0e66a6028b2\") " pod="openstack/kube-state-metrics-0" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.493509 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2spf\" (UniqueName: \"kubernetes.io/projected/f4b85ab2-e407-4779-bcc8-4ea2d35aee72-kube-api-access-z2spf\") pod \"mysqld-exporter-0\" (UID: \"f4b85ab2-e407-4779-bcc8-4ea2d35aee72\") " pod="openstack/mysqld-exporter-0" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.502324 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlv2b\" (UniqueName: \"kubernetes.io/projected/36f27192-451f-4ce2-afda-c0e66a6028b2-kube-api-access-mlv2b\") pod \"kube-state-metrics-0\" (UID: \"36f27192-451f-4ce2-afda-c0e66a6028b2\") " pod="openstack/kube-state-metrics-0" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.563853 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 06 06:20:55 crc kubenswrapper[4809]: I1206 06:20:55.624423 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 06 06:20:56 crc kubenswrapper[4809]: W1206 06:20:56.126112 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf4b85ab2_e407_4779_bcc8_4ea2d35aee72.slice/crio-cfc8b5fe16cfab070370e0726ba28720bac9af5ef6fcbb2e95a62e164036382c WatchSource:0}: Error finding container cfc8b5fe16cfab070370e0726ba28720bac9af5ef6fcbb2e95a62e164036382c: Status 404 returned error can't find the container with id cfc8b5fe16cfab070370e0726ba28720bac9af5ef6fcbb2e95a62e164036382c Dec 06 06:20:56 crc kubenswrapper[4809]: I1206 06:20:56.136250 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 06 06:20:56 crc kubenswrapper[4809]: I1206 06:20:56.284951 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:20:56 crc kubenswrapper[4809]: I1206 06:20:56.285258 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6be13a68-b3a2-4627-a468-bfe138f8e271" containerName="ceilometer-central-agent" containerID="cri-o://395cee0fe2d65139489165e1abcab94073496fc112f29c7a2421e942a24d1c7a" gracePeriod=30 Dec 06 06:20:56 crc kubenswrapper[4809]: I1206 06:20:56.287230 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6be13a68-b3a2-4627-a468-bfe138f8e271" containerName="proxy-httpd" containerID="cri-o://1f3429a9e96b760fffceb1e08061b1428e6750a42a79dd8274be7e5e09bedb8d" gracePeriod=30 Dec 06 06:20:56 crc kubenswrapper[4809]: I1206 06:20:56.287398 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6be13a68-b3a2-4627-a468-bfe138f8e271" containerName="ceilometer-notification-agent" containerID="cri-o://4924bbf255027594929d8c8877a450eb6f34c2b6f12af3f9d584a9d58ac98b13" gracePeriod=30 Dec 06 06:20:56 crc kubenswrapper[4809]: I1206 06:20:56.287453 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6be13a68-b3a2-4627-a468-bfe138f8e271" containerName="sg-core" containerID="cri-o://b6ec38c05e0256409801129a53fbc8fbf46cce01ce3dfd344bc30e17646b7936" gracePeriod=30 Dec 06 06:20:56 crc kubenswrapper[4809]: I1206 06:20:56.321032 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 06:20:57 crc kubenswrapper[4809]: I1206 06:20:57.154411 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"f4b85ab2-e407-4779-bcc8-4ea2d35aee72","Type":"ContainerStarted","Data":"189e482cfa1a0aa330c19d40f052413c88ab24acd14e5c9f8dcc618415291b9e"} Dec 06 06:20:57 crc kubenswrapper[4809]: I1206 06:20:57.155995 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"f4b85ab2-e407-4779-bcc8-4ea2d35aee72","Type":"ContainerStarted","Data":"cfc8b5fe16cfab070370e0726ba28720bac9af5ef6fcbb2e95a62e164036382c"} Dec 06 06:20:57 crc kubenswrapper[4809]: I1206 06:20:57.170380 4809 generic.go:334] "Generic (PLEG): container finished" podID="6be13a68-b3a2-4627-a468-bfe138f8e271" containerID="1f3429a9e96b760fffceb1e08061b1428e6750a42a79dd8274be7e5e09bedb8d" exitCode=0 Dec 06 06:20:57 crc kubenswrapper[4809]: I1206 06:20:57.170413 4809 generic.go:334] "Generic (PLEG): container finished" podID="6be13a68-b3a2-4627-a468-bfe138f8e271" containerID="b6ec38c05e0256409801129a53fbc8fbf46cce01ce3dfd344bc30e17646b7936" exitCode=2 Dec 06 06:20:57 crc kubenswrapper[4809]: I1206 06:20:57.170421 4809 generic.go:334] "Generic (PLEG): container finished" podID="6be13a68-b3a2-4627-a468-bfe138f8e271" containerID="395cee0fe2d65139489165e1abcab94073496fc112f29c7a2421e942a24d1c7a" exitCode=0 Dec 06 06:20:57 crc kubenswrapper[4809]: I1206 06:20:57.170464 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6be13a68-b3a2-4627-a468-bfe138f8e271","Type":"ContainerDied","Data":"1f3429a9e96b760fffceb1e08061b1428e6750a42a79dd8274be7e5e09bedb8d"} Dec 06 06:20:57 crc kubenswrapper[4809]: I1206 06:20:57.170490 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6be13a68-b3a2-4627-a468-bfe138f8e271","Type":"ContainerDied","Data":"b6ec38c05e0256409801129a53fbc8fbf46cce01ce3dfd344bc30e17646b7936"} Dec 06 06:20:57 crc kubenswrapper[4809]: I1206 06:20:57.170499 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6be13a68-b3a2-4627-a468-bfe138f8e271","Type":"ContainerDied","Data":"395cee0fe2d65139489165e1abcab94073496fc112f29c7a2421e942a24d1c7a"} Dec 06 06:20:57 crc kubenswrapper[4809]: I1206 06:20:57.172828 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"36f27192-451f-4ce2-afda-c0e66a6028b2","Type":"ContainerStarted","Data":"b2f3204f4908e566576640527cae5ec93f5e360f0125a482c46c8b15ed867b25"} Dec 06 06:20:57 crc kubenswrapper[4809]: I1206 06:20:57.172858 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"36f27192-451f-4ce2-afda-c0e66a6028b2","Type":"ContainerStarted","Data":"46e83b3af4e5312795157a4638cd2ca154619ba021b14a29db51813edd5559c7"} Dec 06 06:20:57 crc kubenswrapper[4809]: I1206 06:20:57.173255 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 06 06:20:57 crc kubenswrapper[4809]: I1206 06:20:57.198330 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-0" podStartSLOduration=1.689700897 podStartE2EDuration="2.19831477s" podCreationTimestamp="2025-12-06 06:20:55 +0000 UTC" firstStartedPulling="2025-12-06 06:20:56.130839249 +0000 UTC m=+1781.019822191" lastFinishedPulling="2025-12-06 06:20:56.639453122 +0000 UTC m=+1781.528436064" observedRunningTime="2025-12-06 06:20:57.197406335 +0000 UTC m=+1782.086389277" watchObservedRunningTime="2025-12-06 06:20:57.19831477 +0000 UTC m=+1782.087297712" Dec 06 06:20:57 crc kubenswrapper[4809]: I1206 06:20:57.235129 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.853872676 podStartE2EDuration="2.235104482s" podCreationTimestamp="2025-12-06 06:20:55 +0000 UTC" firstStartedPulling="2025-12-06 06:20:56.328179824 +0000 UTC m=+1781.217162756" lastFinishedPulling="2025-12-06 06:20:56.70941162 +0000 UTC m=+1781.598394562" observedRunningTime="2025-12-06 06:20:57.217287301 +0000 UTC m=+1782.106270243" watchObservedRunningTime="2025-12-06 06:20:57.235104482 +0000 UTC m=+1782.124087424" Dec 06 06:20:59 crc kubenswrapper[4809]: I1206 06:20:59.388843 4809 scope.go:117] "RemoveContainer" containerID="9bc2dc88b3b9f309ce0169aa510326dd77022dece4c69cab3141df00fe5c0106" Dec 06 06:20:59 crc kubenswrapper[4809]: E1206 06:20:59.390141 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:21:01 crc kubenswrapper[4809]: I1206 06:21:01.234584 4809 generic.go:334] "Generic (PLEG): container finished" podID="6be13a68-b3a2-4627-a468-bfe138f8e271" containerID="4924bbf255027594929d8c8877a450eb6f34c2b6f12af3f9d584a9d58ac98b13" exitCode=0 Dec 06 06:21:01 crc kubenswrapper[4809]: I1206 06:21:01.234796 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6be13a68-b3a2-4627-a468-bfe138f8e271","Type":"ContainerDied","Data":"4924bbf255027594929d8c8877a450eb6f34c2b6f12af3f9d584a9d58ac98b13"} Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.163446 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.248788 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6be13a68-b3a2-4627-a468-bfe138f8e271","Type":"ContainerDied","Data":"9e2715000c91670c1179b414e60ad5823615d33970710722c170c0c92af55aed"} Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.249154 4809 scope.go:117] "RemoveContainer" containerID="1f3429a9e96b760fffceb1e08061b1428e6750a42a79dd8274be7e5e09bedb8d" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.248891 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.266632 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6be13a68-b3a2-4627-a468-bfe138f8e271-combined-ca-bundle\") pod \"6be13a68-b3a2-4627-a468-bfe138f8e271\" (UID: \"6be13a68-b3a2-4627-a468-bfe138f8e271\") " Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.267118 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6be13a68-b3a2-4627-a468-bfe138f8e271-run-httpd\") pod \"6be13a68-b3a2-4627-a468-bfe138f8e271\" (UID: \"6be13a68-b3a2-4627-a468-bfe138f8e271\") " Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.267243 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6be13a68-b3a2-4627-a468-bfe138f8e271-sg-core-conf-yaml\") pod \"6be13a68-b3a2-4627-a468-bfe138f8e271\" (UID: \"6be13a68-b3a2-4627-a468-bfe138f8e271\") " Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.267320 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6be13a68-b3a2-4627-a468-bfe138f8e271-scripts\") pod \"6be13a68-b3a2-4627-a468-bfe138f8e271\" (UID: \"6be13a68-b3a2-4627-a468-bfe138f8e271\") " Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.267383 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6be13a68-b3a2-4627-a468-bfe138f8e271-log-httpd\") pod \"6be13a68-b3a2-4627-a468-bfe138f8e271\" (UID: \"6be13a68-b3a2-4627-a468-bfe138f8e271\") " Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.267419 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6be13a68-b3a2-4627-a468-bfe138f8e271-config-data\") pod \"6be13a68-b3a2-4627-a468-bfe138f8e271\" (UID: \"6be13a68-b3a2-4627-a468-bfe138f8e271\") " Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.267528 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q5qcf\" (UniqueName: \"kubernetes.io/projected/6be13a68-b3a2-4627-a468-bfe138f8e271-kube-api-access-q5qcf\") pod \"6be13a68-b3a2-4627-a468-bfe138f8e271\" (UID: \"6be13a68-b3a2-4627-a468-bfe138f8e271\") " Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.269463 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6be13a68-b3a2-4627-a468-bfe138f8e271-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "6be13a68-b3a2-4627-a468-bfe138f8e271" (UID: "6be13a68-b3a2-4627-a468-bfe138f8e271"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.269732 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6be13a68-b3a2-4627-a468-bfe138f8e271-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "6be13a68-b3a2-4627-a468-bfe138f8e271" (UID: "6be13a68-b3a2-4627-a468-bfe138f8e271"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.273130 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6be13a68-b3a2-4627-a468-bfe138f8e271-kube-api-access-q5qcf" (OuterVolumeSpecName: "kube-api-access-q5qcf") pod "6be13a68-b3a2-4627-a468-bfe138f8e271" (UID: "6be13a68-b3a2-4627-a468-bfe138f8e271"). InnerVolumeSpecName "kube-api-access-q5qcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.273216 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6be13a68-b3a2-4627-a468-bfe138f8e271-scripts" (OuterVolumeSpecName: "scripts") pod "6be13a68-b3a2-4627-a468-bfe138f8e271" (UID: "6be13a68-b3a2-4627-a468-bfe138f8e271"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.275190 4809 scope.go:117] "RemoveContainer" containerID="b6ec38c05e0256409801129a53fbc8fbf46cce01ce3dfd344bc30e17646b7936" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.302797 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6be13a68-b3a2-4627-a468-bfe138f8e271-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "6be13a68-b3a2-4627-a468-bfe138f8e271" (UID: "6be13a68-b3a2-4627-a468-bfe138f8e271"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.362888 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6be13a68-b3a2-4627-a468-bfe138f8e271-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6be13a68-b3a2-4627-a468-bfe138f8e271" (UID: "6be13a68-b3a2-4627-a468-bfe138f8e271"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.371322 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q5qcf\" (UniqueName: \"kubernetes.io/projected/6be13a68-b3a2-4627-a468-bfe138f8e271-kube-api-access-q5qcf\") on node \"crc\" DevicePath \"\"" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.371356 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6be13a68-b3a2-4627-a468-bfe138f8e271-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.371368 4809 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6be13a68-b3a2-4627-a468-bfe138f8e271-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.371377 4809 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6be13a68-b3a2-4627-a468-bfe138f8e271-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.371387 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6be13a68-b3a2-4627-a468-bfe138f8e271-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.371396 4809 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6be13a68-b3a2-4627-a468-bfe138f8e271-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.391769 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6be13a68-b3a2-4627-a468-bfe138f8e271-config-data" (OuterVolumeSpecName: "config-data") pod "6be13a68-b3a2-4627-a468-bfe138f8e271" (UID: "6be13a68-b3a2-4627-a468-bfe138f8e271"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.459514 4809 scope.go:117] "RemoveContainer" containerID="4924bbf255027594929d8c8877a450eb6f34c2b6f12af3f9d584a9d58ac98b13" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.473166 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6be13a68-b3a2-4627-a468-bfe138f8e271-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.480858 4809 scope.go:117] "RemoveContainer" containerID="395cee0fe2d65139489165e1abcab94073496fc112f29c7a2421e942a24d1c7a" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.595348 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.607881 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.621856 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:21:02 crc kubenswrapper[4809]: E1206 06:21:02.622468 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6be13a68-b3a2-4627-a468-bfe138f8e271" containerName="sg-core" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.622491 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="6be13a68-b3a2-4627-a468-bfe138f8e271" containerName="sg-core" Dec 06 06:21:02 crc kubenswrapper[4809]: E1206 06:21:02.622514 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6be13a68-b3a2-4627-a468-bfe138f8e271" containerName="ceilometer-central-agent" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.622521 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="6be13a68-b3a2-4627-a468-bfe138f8e271" containerName="ceilometer-central-agent" Dec 06 06:21:02 crc kubenswrapper[4809]: E1206 06:21:02.622535 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6be13a68-b3a2-4627-a468-bfe138f8e271" containerName="proxy-httpd" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.622542 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="6be13a68-b3a2-4627-a468-bfe138f8e271" containerName="proxy-httpd" Dec 06 06:21:02 crc kubenswrapper[4809]: E1206 06:21:02.622562 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6be13a68-b3a2-4627-a468-bfe138f8e271" containerName="ceilometer-notification-agent" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.622569 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="6be13a68-b3a2-4627-a468-bfe138f8e271" containerName="ceilometer-notification-agent" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.622832 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="6be13a68-b3a2-4627-a468-bfe138f8e271" containerName="sg-core" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.622860 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="6be13a68-b3a2-4627-a468-bfe138f8e271" containerName="proxy-httpd" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.622876 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="6be13a68-b3a2-4627-a468-bfe138f8e271" containerName="ceilometer-central-agent" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.622890 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="6be13a68-b3a2-4627-a468-bfe138f8e271" containerName="ceilometer-notification-agent" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.625109 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.627421 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.627577 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.627701 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.634329 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.676816 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/475f8cc7-1795-455e-a448-734b3bc278a8-log-httpd\") pod \"ceilometer-0\" (UID: \"475f8cc7-1795-455e-a448-734b3bc278a8\") " pod="openstack/ceilometer-0" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.676884 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/475f8cc7-1795-455e-a448-734b3bc278a8-config-data\") pod \"ceilometer-0\" (UID: \"475f8cc7-1795-455e-a448-734b3bc278a8\") " pod="openstack/ceilometer-0" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.676927 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/475f8cc7-1795-455e-a448-734b3bc278a8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"475f8cc7-1795-455e-a448-734b3bc278a8\") " pod="openstack/ceilometer-0" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.676973 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/475f8cc7-1795-455e-a448-734b3bc278a8-scripts\") pod \"ceilometer-0\" (UID: \"475f8cc7-1795-455e-a448-734b3bc278a8\") " pod="openstack/ceilometer-0" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.677053 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-874rm\" (UniqueName: \"kubernetes.io/projected/475f8cc7-1795-455e-a448-734b3bc278a8-kube-api-access-874rm\") pod \"ceilometer-0\" (UID: \"475f8cc7-1795-455e-a448-734b3bc278a8\") " pod="openstack/ceilometer-0" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.677221 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/475f8cc7-1795-455e-a448-734b3bc278a8-run-httpd\") pod \"ceilometer-0\" (UID: \"475f8cc7-1795-455e-a448-734b3bc278a8\") " pod="openstack/ceilometer-0" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.677276 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/475f8cc7-1795-455e-a448-734b3bc278a8-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"475f8cc7-1795-455e-a448-734b3bc278a8\") " pod="openstack/ceilometer-0" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.677371 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/475f8cc7-1795-455e-a448-734b3bc278a8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"475f8cc7-1795-455e-a448-734b3bc278a8\") " pod="openstack/ceilometer-0" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.779202 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/475f8cc7-1795-455e-a448-734b3bc278a8-log-httpd\") pod \"ceilometer-0\" (UID: \"475f8cc7-1795-455e-a448-734b3bc278a8\") " pod="openstack/ceilometer-0" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.779249 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/475f8cc7-1795-455e-a448-734b3bc278a8-config-data\") pod \"ceilometer-0\" (UID: \"475f8cc7-1795-455e-a448-734b3bc278a8\") " pod="openstack/ceilometer-0" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.779283 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/475f8cc7-1795-455e-a448-734b3bc278a8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"475f8cc7-1795-455e-a448-734b3bc278a8\") " pod="openstack/ceilometer-0" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.779301 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/475f8cc7-1795-455e-a448-734b3bc278a8-scripts\") pod \"ceilometer-0\" (UID: \"475f8cc7-1795-455e-a448-734b3bc278a8\") " pod="openstack/ceilometer-0" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.779341 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-874rm\" (UniqueName: \"kubernetes.io/projected/475f8cc7-1795-455e-a448-734b3bc278a8-kube-api-access-874rm\") pod \"ceilometer-0\" (UID: \"475f8cc7-1795-455e-a448-734b3bc278a8\") " pod="openstack/ceilometer-0" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.779371 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/475f8cc7-1795-455e-a448-734b3bc278a8-run-httpd\") pod \"ceilometer-0\" (UID: \"475f8cc7-1795-455e-a448-734b3bc278a8\") " pod="openstack/ceilometer-0" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.779409 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/475f8cc7-1795-455e-a448-734b3bc278a8-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"475f8cc7-1795-455e-a448-734b3bc278a8\") " pod="openstack/ceilometer-0" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.779441 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/475f8cc7-1795-455e-a448-734b3bc278a8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"475f8cc7-1795-455e-a448-734b3bc278a8\") " pod="openstack/ceilometer-0" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.779702 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/475f8cc7-1795-455e-a448-734b3bc278a8-log-httpd\") pod \"ceilometer-0\" (UID: \"475f8cc7-1795-455e-a448-734b3bc278a8\") " pod="openstack/ceilometer-0" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.780122 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/475f8cc7-1795-455e-a448-734b3bc278a8-run-httpd\") pod \"ceilometer-0\" (UID: \"475f8cc7-1795-455e-a448-734b3bc278a8\") " pod="openstack/ceilometer-0" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.783115 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/475f8cc7-1795-455e-a448-734b3bc278a8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"475f8cc7-1795-455e-a448-734b3bc278a8\") " pod="openstack/ceilometer-0" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.783283 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/475f8cc7-1795-455e-a448-734b3bc278a8-scripts\") pod \"ceilometer-0\" (UID: \"475f8cc7-1795-455e-a448-734b3bc278a8\") " pod="openstack/ceilometer-0" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.783732 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/475f8cc7-1795-455e-a448-734b3bc278a8-config-data\") pod \"ceilometer-0\" (UID: \"475f8cc7-1795-455e-a448-734b3bc278a8\") " pod="openstack/ceilometer-0" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.784357 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/475f8cc7-1795-455e-a448-734b3bc278a8-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"475f8cc7-1795-455e-a448-734b3bc278a8\") " pod="openstack/ceilometer-0" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.784535 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/475f8cc7-1795-455e-a448-734b3bc278a8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"475f8cc7-1795-455e-a448-734b3bc278a8\") " pod="openstack/ceilometer-0" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.800716 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-874rm\" (UniqueName: \"kubernetes.io/projected/475f8cc7-1795-455e-a448-734b3bc278a8-kube-api-access-874rm\") pod \"ceilometer-0\" (UID: \"475f8cc7-1795-455e-a448-734b3bc278a8\") " pod="openstack/ceilometer-0" Dec 06 06:21:02 crc kubenswrapper[4809]: I1206 06:21:02.955584 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:21:03 crc kubenswrapper[4809]: I1206 06:21:03.411700 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6be13a68-b3a2-4627-a468-bfe138f8e271" path="/var/lib/kubelet/pods/6be13a68-b3a2-4627-a468-bfe138f8e271/volumes" Dec 06 06:21:03 crc kubenswrapper[4809]: I1206 06:21:03.444368 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:21:04 crc kubenswrapper[4809]: I1206 06:21:04.278204 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"475f8cc7-1795-455e-a448-734b3bc278a8","Type":"ContainerStarted","Data":"1795100d44f0ac579862227ce9537745bad06a5a5e4c6b78eee551775f2cb823"} Dec 06 06:21:04 crc kubenswrapper[4809]: I1206 06:21:04.402201 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 06 06:21:04 crc kubenswrapper[4809]: I1206 06:21:04.402761 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 06 06:21:04 crc kubenswrapper[4809]: I1206 06:21:04.408553 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 06 06:21:04 crc kubenswrapper[4809]: I1206 06:21:04.414088 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 06 06:21:05 crc kubenswrapper[4809]: I1206 06:21:05.287238 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 06 06:21:05 crc kubenswrapper[4809]: I1206 06:21:05.295279 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 06 06:21:05 crc kubenswrapper[4809]: I1206 06:21:05.632087 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 06 06:21:06 crc kubenswrapper[4809]: I1206 06:21:06.301103 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"475f8cc7-1795-455e-a448-734b3bc278a8","Type":"ContainerStarted","Data":"5d8d717f089ed3b927d6066c8f54271cf050ddf80f80dd72782cd64fb182206b"} Dec 06 06:21:07 crc kubenswrapper[4809]: I1206 06:21:07.313075 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"475f8cc7-1795-455e-a448-734b3bc278a8","Type":"ContainerStarted","Data":"569d10abcb4d5547b8dce30fe53ec252da0e9edb9c96e787021ef0b8e65589d6"} Dec 06 06:21:09 crc kubenswrapper[4809]: I1206 06:21:09.344052 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"475f8cc7-1795-455e-a448-734b3bc278a8","Type":"ContainerStarted","Data":"bb98373f270bc6b48302c14e5bb82802f71cddde648d84ddca4feb136955cbbb"} Dec 06 06:21:11 crc kubenswrapper[4809]: I1206 06:21:11.376346 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"475f8cc7-1795-455e-a448-734b3bc278a8","Type":"ContainerStarted","Data":"c005ef42f1cd2a1c17707563dbe907644a57297c90f9f56ad9a2e907a171a8bc"} Dec 06 06:21:11 crc kubenswrapper[4809]: I1206 06:21:11.377110 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 06:21:11 crc kubenswrapper[4809]: I1206 06:21:11.407419 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.243073954 podStartE2EDuration="9.407393613s" podCreationTimestamp="2025-12-06 06:21:02 +0000 UTC" firstStartedPulling="2025-12-06 06:21:03.447252703 +0000 UTC m=+1788.336235645" lastFinishedPulling="2025-12-06 06:21:10.611572362 +0000 UTC m=+1795.500555304" observedRunningTime="2025-12-06 06:21:11.402180112 +0000 UTC m=+1796.291163094" watchObservedRunningTime="2025-12-06 06:21:11.407393613 +0000 UTC m=+1796.296376555" Dec 06 06:21:12 crc kubenswrapper[4809]: I1206 06:21:12.389425 4809 scope.go:117] "RemoveContainer" containerID="9bc2dc88b3b9f309ce0169aa510326dd77022dece4c69cab3141df00fe5c0106" Dec 06 06:21:12 crc kubenswrapper[4809]: E1206 06:21:12.389850 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:21:24 crc kubenswrapper[4809]: I1206 06:21:24.389521 4809 scope.go:117] "RemoveContainer" containerID="9bc2dc88b3b9f309ce0169aa510326dd77022dece4c69cab3141df00fe5c0106" Dec 06 06:21:24 crc kubenswrapper[4809]: E1206 06:21:24.390294 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:21:32 crc kubenswrapper[4809]: I1206 06:21:32.964768 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 06 06:21:36 crc kubenswrapper[4809]: I1206 06:21:36.390156 4809 scope.go:117] "RemoveContainer" containerID="9bc2dc88b3b9f309ce0169aa510326dd77022dece4c69cab3141df00fe5c0106" Dec 06 06:21:36 crc kubenswrapper[4809]: E1206 06:21:36.391964 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:21:46 crc kubenswrapper[4809]: I1206 06:21:46.621017 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-ckhg9"] Dec 06 06:21:46 crc kubenswrapper[4809]: I1206 06:21:46.631631 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-ckhg9"] Dec 06 06:21:46 crc kubenswrapper[4809]: I1206 06:21:46.737418 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-ng4qq"] Dec 06 06:21:46 crc kubenswrapper[4809]: I1206 06:21:46.739817 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-ng4qq" Dec 06 06:21:46 crc kubenswrapper[4809]: I1206 06:21:46.754079 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-ng4qq"] Dec 06 06:21:46 crc kubenswrapper[4809]: I1206 06:21:46.881294 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5vwn\" (UniqueName: \"kubernetes.io/projected/18fa982c-0572-4702-8c51-f0f7af7bf52f-kube-api-access-s5vwn\") pod \"heat-db-sync-ng4qq\" (UID: \"18fa982c-0572-4702-8c51-f0f7af7bf52f\") " pod="openstack/heat-db-sync-ng4qq" Dec 06 06:21:46 crc kubenswrapper[4809]: I1206 06:21:46.881661 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18fa982c-0572-4702-8c51-f0f7af7bf52f-config-data\") pod \"heat-db-sync-ng4qq\" (UID: \"18fa982c-0572-4702-8c51-f0f7af7bf52f\") " pod="openstack/heat-db-sync-ng4qq" Dec 06 06:21:46 crc kubenswrapper[4809]: I1206 06:21:46.881905 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18fa982c-0572-4702-8c51-f0f7af7bf52f-combined-ca-bundle\") pod \"heat-db-sync-ng4qq\" (UID: \"18fa982c-0572-4702-8c51-f0f7af7bf52f\") " pod="openstack/heat-db-sync-ng4qq" Dec 06 06:21:46 crc kubenswrapper[4809]: I1206 06:21:46.984258 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18fa982c-0572-4702-8c51-f0f7af7bf52f-config-data\") pod \"heat-db-sync-ng4qq\" (UID: \"18fa982c-0572-4702-8c51-f0f7af7bf52f\") " pod="openstack/heat-db-sync-ng4qq" Dec 06 06:21:46 crc kubenswrapper[4809]: I1206 06:21:46.984395 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18fa982c-0572-4702-8c51-f0f7af7bf52f-combined-ca-bundle\") pod \"heat-db-sync-ng4qq\" (UID: \"18fa982c-0572-4702-8c51-f0f7af7bf52f\") " pod="openstack/heat-db-sync-ng4qq" Dec 06 06:21:46 crc kubenswrapper[4809]: I1206 06:21:46.984503 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5vwn\" (UniqueName: \"kubernetes.io/projected/18fa982c-0572-4702-8c51-f0f7af7bf52f-kube-api-access-s5vwn\") pod \"heat-db-sync-ng4qq\" (UID: \"18fa982c-0572-4702-8c51-f0f7af7bf52f\") " pod="openstack/heat-db-sync-ng4qq" Dec 06 06:21:47 crc kubenswrapper[4809]: I1206 06:21:47.010590 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18fa982c-0572-4702-8c51-f0f7af7bf52f-combined-ca-bundle\") pod \"heat-db-sync-ng4qq\" (UID: \"18fa982c-0572-4702-8c51-f0f7af7bf52f\") " pod="openstack/heat-db-sync-ng4qq" Dec 06 06:21:47 crc kubenswrapper[4809]: I1206 06:21:47.016228 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5vwn\" (UniqueName: \"kubernetes.io/projected/18fa982c-0572-4702-8c51-f0f7af7bf52f-kube-api-access-s5vwn\") pod \"heat-db-sync-ng4qq\" (UID: \"18fa982c-0572-4702-8c51-f0f7af7bf52f\") " pod="openstack/heat-db-sync-ng4qq" Dec 06 06:21:47 crc kubenswrapper[4809]: I1206 06:21:47.022341 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18fa982c-0572-4702-8c51-f0f7af7bf52f-config-data\") pod \"heat-db-sync-ng4qq\" (UID: \"18fa982c-0572-4702-8c51-f0f7af7bf52f\") " pod="openstack/heat-db-sync-ng4qq" Dec 06 06:21:47 crc kubenswrapper[4809]: I1206 06:21:47.065598 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-ng4qq" Dec 06 06:21:47 crc kubenswrapper[4809]: I1206 06:21:47.416216 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="edf1ab66-b483-4172-81fd-dd3f3b9b44aa" path="/var/lib/kubelet/pods/edf1ab66-b483-4172-81fd-dd3f3b9b44aa/volumes" Dec 06 06:21:47 crc kubenswrapper[4809]: I1206 06:21:47.575005 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-ng4qq"] Dec 06 06:21:47 crc kubenswrapper[4809]: I1206 06:21:47.833081 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-ng4qq" event={"ID":"18fa982c-0572-4702-8c51-f0f7af7bf52f","Type":"ContainerStarted","Data":"a2c904fd7e9d2a74328c28fce664dc5fb7678da34d29f5b08c18e40cfe4bdcc4"} Dec 06 06:21:49 crc kubenswrapper[4809]: I1206 06:21:49.023034 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:21:49 crc kubenswrapper[4809]: I1206 06:21:49.023503 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="475f8cc7-1795-455e-a448-734b3bc278a8" containerName="ceilometer-central-agent" containerID="cri-o://5d8d717f089ed3b927d6066c8f54271cf050ddf80f80dd72782cd64fb182206b" gracePeriod=30 Dec 06 06:21:49 crc kubenswrapper[4809]: I1206 06:21:49.024036 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="475f8cc7-1795-455e-a448-734b3bc278a8" containerName="proxy-httpd" containerID="cri-o://c005ef42f1cd2a1c17707563dbe907644a57297c90f9f56ad9a2e907a171a8bc" gracePeriod=30 Dec 06 06:21:49 crc kubenswrapper[4809]: I1206 06:21:49.024079 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="475f8cc7-1795-455e-a448-734b3bc278a8" containerName="sg-core" containerID="cri-o://bb98373f270bc6b48302c14e5bb82802f71cddde648d84ddca4feb136955cbbb" gracePeriod=30 Dec 06 06:21:49 crc kubenswrapper[4809]: I1206 06:21:49.024113 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="475f8cc7-1795-455e-a448-734b3bc278a8" containerName="ceilometer-notification-agent" containerID="cri-o://569d10abcb4d5547b8dce30fe53ec252da0e9edb9c96e787021ef0b8e65589d6" gracePeriod=30 Dec 06 06:21:49 crc kubenswrapper[4809]: I1206 06:21:49.863389 4809 generic.go:334] "Generic (PLEG): container finished" podID="475f8cc7-1795-455e-a448-734b3bc278a8" containerID="c005ef42f1cd2a1c17707563dbe907644a57297c90f9f56ad9a2e907a171a8bc" exitCode=0 Dec 06 06:21:49 crc kubenswrapper[4809]: I1206 06:21:49.863966 4809 generic.go:334] "Generic (PLEG): container finished" podID="475f8cc7-1795-455e-a448-734b3bc278a8" containerID="bb98373f270bc6b48302c14e5bb82802f71cddde648d84ddca4feb136955cbbb" exitCode=2 Dec 06 06:21:49 crc kubenswrapper[4809]: I1206 06:21:49.863980 4809 generic.go:334] "Generic (PLEG): container finished" podID="475f8cc7-1795-455e-a448-734b3bc278a8" containerID="5d8d717f089ed3b927d6066c8f54271cf050ddf80f80dd72782cd64fb182206b" exitCode=0 Dec 06 06:21:49 crc kubenswrapper[4809]: I1206 06:21:49.864006 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"475f8cc7-1795-455e-a448-734b3bc278a8","Type":"ContainerDied","Data":"c005ef42f1cd2a1c17707563dbe907644a57297c90f9f56ad9a2e907a171a8bc"} Dec 06 06:21:49 crc kubenswrapper[4809]: I1206 06:21:49.864038 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"475f8cc7-1795-455e-a448-734b3bc278a8","Type":"ContainerDied","Data":"bb98373f270bc6b48302c14e5bb82802f71cddde648d84ddca4feb136955cbbb"} Dec 06 06:21:49 crc kubenswrapper[4809]: I1206 06:21:49.864055 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"475f8cc7-1795-455e-a448-734b3bc278a8","Type":"ContainerDied","Data":"5d8d717f089ed3b927d6066c8f54271cf050ddf80f80dd72782cd64fb182206b"} Dec 06 06:21:50 crc kubenswrapper[4809]: I1206 06:21:50.351166 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 06:21:51 crc kubenswrapper[4809]: I1206 06:21:51.389256 4809 scope.go:117] "RemoveContainer" containerID="9bc2dc88b3b9f309ce0169aa510326dd77022dece4c69cab3141df00fe5c0106" Dec 06 06:21:51 crc kubenswrapper[4809]: E1206 06:21:51.389855 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:21:51 crc kubenswrapper[4809]: I1206 06:21:51.428602 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 06:21:51 crc kubenswrapper[4809]: I1206 06:21:51.466368 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lkm6p"] Dec 06 06:21:51 crc kubenswrapper[4809]: I1206 06:21:51.470830 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lkm6p" Dec 06 06:21:51 crc kubenswrapper[4809]: I1206 06:21:51.483194 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lkm6p"] Dec 06 06:21:51 crc kubenswrapper[4809]: I1206 06:21:51.633648 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgzqt\" (UniqueName: \"kubernetes.io/projected/81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf-kube-api-access-bgzqt\") pod \"redhat-operators-lkm6p\" (UID: \"81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf\") " pod="openshift-marketplace/redhat-operators-lkm6p" Dec 06 06:21:51 crc kubenswrapper[4809]: I1206 06:21:51.633817 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf-catalog-content\") pod \"redhat-operators-lkm6p\" (UID: \"81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf\") " pod="openshift-marketplace/redhat-operators-lkm6p" Dec 06 06:21:51 crc kubenswrapper[4809]: I1206 06:21:51.633869 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf-utilities\") pod \"redhat-operators-lkm6p\" (UID: \"81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf\") " pod="openshift-marketplace/redhat-operators-lkm6p" Dec 06 06:21:51 crc kubenswrapper[4809]: I1206 06:21:51.735902 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf-catalog-content\") pod \"redhat-operators-lkm6p\" (UID: \"81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf\") " pod="openshift-marketplace/redhat-operators-lkm6p" Dec 06 06:21:51 crc kubenswrapper[4809]: I1206 06:21:51.736040 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf-utilities\") pod \"redhat-operators-lkm6p\" (UID: \"81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf\") " pod="openshift-marketplace/redhat-operators-lkm6p" Dec 06 06:21:51 crc kubenswrapper[4809]: I1206 06:21:51.736149 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgzqt\" (UniqueName: \"kubernetes.io/projected/81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf-kube-api-access-bgzqt\") pod \"redhat-operators-lkm6p\" (UID: \"81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf\") " pod="openshift-marketplace/redhat-operators-lkm6p" Dec 06 06:21:51 crc kubenswrapper[4809]: I1206 06:21:51.736403 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf-catalog-content\") pod \"redhat-operators-lkm6p\" (UID: \"81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf\") " pod="openshift-marketplace/redhat-operators-lkm6p" Dec 06 06:21:51 crc kubenswrapper[4809]: I1206 06:21:51.736561 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf-utilities\") pod \"redhat-operators-lkm6p\" (UID: \"81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf\") " pod="openshift-marketplace/redhat-operators-lkm6p" Dec 06 06:21:51 crc kubenswrapper[4809]: I1206 06:21:51.766153 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgzqt\" (UniqueName: \"kubernetes.io/projected/81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf-kube-api-access-bgzqt\") pod \"redhat-operators-lkm6p\" (UID: \"81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf\") " pod="openshift-marketplace/redhat-operators-lkm6p" Dec 06 06:21:51 crc kubenswrapper[4809]: I1206 06:21:51.852676 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lkm6p" Dec 06 06:21:52 crc kubenswrapper[4809]: I1206 06:21:52.637470 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lkm6p"] Dec 06 06:21:52 crc kubenswrapper[4809]: I1206 06:21:52.931985 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lkm6p" event={"ID":"81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf","Type":"ContainerStarted","Data":"b9a7d47a334a975be6b37523c4b63f20f2de3910a5e8f9094bc13442d7102cfe"} Dec 06 06:21:53 crc kubenswrapper[4809]: I1206 06:21:53.951659 4809 generic.go:334] "Generic (PLEG): container finished" podID="81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf" containerID="799aefafe4febe7c505514f10f778adba637b52eda146b500661551502382ec5" exitCode=0 Dec 06 06:21:53 crc kubenswrapper[4809]: I1206 06:21:53.951866 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lkm6p" event={"ID":"81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf","Type":"ContainerDied","Data":"799aefafe4febe7c505514f10f778adba637b52eda146b500661551502382ec5"} Dec 06 06:21:54 crc kubenswrapper[4809]: I1206 06:21:54.853745 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="eb1bd4de-6456-4817-a85d-ab0176153f2c" containerName="rabbitmq" containerID="cri-o://a90c2aa3e3fe73387e9b264aedea8b1f549bfaa7885df6302588dd069fb8db47" gracePeriod=604796 Dec 06 06:21:54 crc kubenswrapper[4809]: I1206 06:21:54.970292 4809 generic.go:334] "Generic (PLEG): container finished" podID="475f8cc7-1795-455e-a448-734b3bc278a8" containerID="569d10abcb4d5547b8dce30fe53ec252da0e9edb9c96e787021ef0b8e65589d6" exitCode=0 Dec 06 06:21:54 crc kubenswrapper[4809]: I1206 06:21:54.970346 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"475f8cc7-1795-455e-a448-734b3bc278a8","Type":"ContainerDied","Data":"569d10abcb4d5547b8dce30fe53ec252da0e9edb9c96e787021ef0b8e65589d6"} Dec 06 06:21:55 crc kubenswrapper[4809]: I1206 06:21:55.966649 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="1ceabc13-38c5-4425-8d85-c3afa8935243" containerName="rabbitmq" containerID="cri-o://162e28f89b66909f6341a5b02785bb02066b04eb4d2c67d23537c8ea8505ae27" gracePeriod=604796 Dec 06 06:22:02 crc kubenswrapper[4809]: I1206 06:22:02.389151 4809 scope.go:117] "RemoveContainer" containerID="9bc2dc88b3b9f309ce0169aa510326dd77022dece4c69cab3141df00fe5c0106" Dec 06 06:22:02 crc kubenswrapper[4809]: E1206 06:22:02.391855 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:22:02 crc kubenswrapper[4809]: I1206 06:22:02.956833 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="475f8cc7-1795-455e-a448-734b3bc278a8" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.1.4:3000/\": dial tcp 10.217.1.4:3000: connect: connection refused" Dec 06 06:22:03 crc kubenswrapper[4809]: I1206 06:22:03.077599 4809 generic.go:334] "Generic (PLEG): container finished" podID="eb1bd4de-6456-4817-a85d-ab0176153f2c" containerID="a90c2aa3e3fe73387e9b264aedea8b1f549bfaa7885df6302588dd069fb8db47" exitCode=0 Dec 06 06:22:03 crc kubenswrapper[4809]: I1206 06:22:03.077652 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"eb1bd4de-6456-4817-a85d-ab0176153f2c","Type":"ContainerDied","Data":"a90c2aa3e3fe73387e9b264aedea8b1f549bfaa7885df6302588dd069fb8db47"} Dec 06 06:22:04 crc kubenswrapper[4809]: I1206 06:22:04.609860 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="1ceabc13-38c5-4425-8d85-c3afa8935243" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.127:5671: connect: connection refused" Dec 06 06:22:06 crc kubenswrapper[4809]: I1206 06:22:06.880064 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 06 06:22:06 crc kubenswrapper[4809]: I1206 06:22:06.890690 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.079701 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/475f8cc7-1795-455e-a448-734b3bc278a8-sg-core-conf-yaml\") pod \"475f8cc7-1795-455e-a448-734b3bc278a8\" (UID: \"475f8cc7-1795-455e-a448-734b3bc278a8\") " Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.080116 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/475f8cc7-1795-455e-a448-734b3bc278a8-run-httpd\") pod \"475f8cc7-1795-455e-a448-734b3bc278a8\" (UID: \"475f8cc7-1795-455e-a448-734b3bc278a8\") " Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.080194 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-874rm\" (UniqueName: \"kubernetes.io/projected/475f8cc7-1795-455e-a448-734b3bc278a8-kube-api-access-874rm\") pod \"475f8cc7-1795-455e-a448-734b3bc278a8\" (UID: \"475f8cc7-1795-455e-a448-734b3bc278a8\") " Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.080283 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/475f8cc7-1795-455e-a448-734b3bc278a8-scripts\") pod \"475f8cc7-1795-455e-a448-734b3bc278a8\" (UID: \"475f8cc7-1795-455e-a448-734b3bc278a8\") " Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.080423 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/475f8cc7-1795-455e-a448-734b3bc278a8-ceilometer-tls-certs\") pod \"475f8cc7-1795-455e-a448-734b3bc278a8\" (UID: \"475f8cc7-1795-455e-a448-734b3bc278a8\") " Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.080445 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/eb1bd4de-6456-4817-a85d-ab0176153f2c-erlang-cookie-secret\") pod \"eb1bd4de-6456-4817-a85d-ab0176153f2c\" (UID: \"eb1bd4de-6456-4817-a85d-ab0176153f2c\") " Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.080462 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/eb1bd4de-6456-4817-a85d-ab0176153f2c-rabbitmq-confd\") pod \"eb1bd4de-6456-4817-a85d-ab0176153f2c\" (UID: \"eb1bd4de-6456-4817-a85d-ab0176153f2c\") " Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.080479 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/475f8cc7-1795-455e-a448-734b3bc278a8-log-httpd\") pod \"475f8cc7-1795-455e-a448-734b3bc278a8\" (UID: \"475f8cc7-1795-455e-a448-734b3bc278a8\") " Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.080496 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/eb1bd4de-6456-4817-a85d-ab0176153f2c-pod-info\") pod \"eb1bd4de-6456-4817-a85d-ab0176153f2c\" (UID: \"eb1bd4de-6456-4817-a85d-ab0176153f2c\") " Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.080524 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/eb1bd4de-6456-4817-a85d-ab0176153f2c-rabbitmq-plugins\") pod \"eb1bd4de-6456-4817-a85d-ab0176153f2c\" (UID: \"eb1bd4de-6456-4817-a85d-ab0176153f2c\") " Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.080544 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"eb1bd4de-6456-4817-a85d-ab0176153f2c\" (UID: \"eb1bd4de-6456-4817-a85d-ab0176153f2c\") " Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.080566 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eb1bd4de-6456-4817-a85d-ab0176153f2c-config-data\") pod \"eb1bd4de-6456-4817-a85d-ab0176153f2c\" (UID: \"eb1bd4de-6456-4817-a85d-ab0176153f2c\") " Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.080585 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/eb1bd4de-6456-4817-a85d-ab0176153f2c-server-conf\") pod \"eb1bd4de-6456-4817-a85d-ab0176153f2c\" (UID: \"eb1bd4de-6456-4817-a85d-ab0176153f2c\") " Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.080611 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/eb1bd4de-6456-4817-a85d-ab0176153f2c-rabbitmq-tls\") pod \"eb1bd4de-6456-4817-a85d-ab0176153f2c\" (UID: \"eb1bd4de-6456-4817-a85d-ab0176153f2c\") " Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.080630 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/eb1bd4de-6456-4817-a85d-ab0176153f2c-rabbitmq-erlang-cookie\") pod \"eb1bd4de-6456-4817-a85d-ab0176153f2c\" (UID: \"eb1bd4de-6456-4817-a85d-ab0176153f2c\") " Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.080673 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/eb1bd4de-6456-4817-a85d-ab0176153f2c-plugins-conf\") pod \"eb1bd4de-6456-4817-a85d-ab0176153f2c\" (UID: \"eb1bd4de-6456-4817-a85d-ab0176153f2c\") " Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.080700 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/475f8cc7-1795-455e-a448-734b3bc278a8-combined-ca-bundle\") pod \"475f8cc7-1795-455e-a448-734b3bc278a8\" (UID: \"475f8cc7-1795-455e-a448-734b3bc278a8\") " Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.080723 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/475f8cc7-1795-455e-a448-734b3bc278a8-config-data\") pod \"475f8cc7-1795-455e-a448-734b3bc278a8\" (UID: \"475f8cc7-1795-455e-a448-734b3bc278a8\") " Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.080750 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hfn22\" (UniqueName: \"kubernetes.io/projected/eb1bd4de-6456-4817-a85d-ab0176153f2c-kube-api-access-hfn22\") pod \"eb1bd4de-6456-4817-a85d-ab0176153f2c\" (UID: \"eb1bd4de-6456-4817-a85d-ab0176153f2c\") " Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.080900 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/475f8cc7-1795-455e-a448-734b3bc278a8-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "475f8cc7-1795-455e-a448-734b3bc278a8" (UID: "475f8cc7-1795-455e-a448-734b3bc278a8"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.081323 4809 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/475f8cc7-1795-455e-a448-734b3bc278a8-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.082964 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb1bd4de-6456-4817-a85d-ab0176153f2c-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "eb1bd4de-6456-4817-a85d-ab0176153f2c" (UID: "eb1bd4de-6456-4817-a85d-ab0176153f2c"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.084093 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb1bd4de-6456-4817-a85d-ab0176153f2c-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "eb1bd4de-6456-4817-a85d-ab0176153f2c" (UID: "eb1bd4de-6456-4817-a85d-ab0176153f2c"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.086022 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb1bd4de-6456-4817-a85d-ab0176153f2c-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "eb1bd4de-6456-4817-a85d-ab0176153f2c" (UID: "eb1bd4de-6456-4817-a85d-ab0176153f2c"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.087016 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/475f8cc7-1795-455e-a448-734b3bc278a8-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "475f8cc7-1795-455e-a448-734b3bc278a8" (UID: "475f8cc7-1795-455e-a448-734b3bc278a8"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.097724 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb1bd4de-6456-4817-a85d-ab0176153f2c-kube-api-access-hfn22" (OuterVolumeSpecName: "kube-api-access-hfn22") pod "eb1bd4de-6456-4817-a85d-ab0176153f2c" (UID: "eb1bd4de-6456-4817-a85d-ab0176153f2c"). InnerVolumeSpecName "kube-api-access-hfn22". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.105308 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb1bd4de-6456-4817-a85d-ab0176153f2c-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "eb1bd4de-6456-4817-a85d-ab0176153f2c" (UID: "eb1bd4de-6456-4817-a85d-ab0176153f2c"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.106812 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "persistence") pod "eb1bd4de-6456-4817-a85d-ab0176153f2c" (UID: "eb1bd4de-6456-4817-a85d-ab0176153f2c"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.107069 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb1bd4de-6456-4817-a85d-ab0176153f2c-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "eb1bd4de-6456-4817-a85d-ab0176153f2c" (UID: "eb1bd4de-6456-4817-a85d-ab0176153f2c"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.107235 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/475f8cc7-1795-455e-a448-734b3bc278a8-scripts" (OuterVolumeSpecName: "scripts") pod "475f8cc7-1795-455e-a448-734b3bc278a8" (UID: "475f8cc7-1795-455e-a448-734b3bc278a8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.107329 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/475f8cc7-1795-455e-a448-734b3bc278a8-kube-api-access-874rm" (OuterVolumeSpecName: "kube-api-access-874rm") pod "475f8cc7-1795-455e-a448-734b3bc278a8" (UID: "475f8cc7-1795-455e-a448-734b3bc278a8"). InnerVolumeSpecName "kube-api-access-874rm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.107577 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/eb1bd4de-6456-4817-a85d-ab0176153f2c-pod-info" (OuterVolumeSpecName: "pod-info") pod "eb1bd4de-6456-4817-a85d-ab0176153f2c" (UID: "eb1bd4de-6456-4817-a85d-ab0176153f2c"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.135244 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb1bd4de-6456-4817-a85d-ab0176153f2c-config-data" (OuterVolumeSpecName: "config-data") pod "eb1bd4de-6456-4817-a85d-ab0176153f2c" (UID: "eb1bd4de-6456-4817-a85d-ab0176153f2c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.161726 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.161767 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"eb1bd4de-6456-4817-a85d-ab0176153f2c","Type":"ContainerDied","Data":"e04ab83414709672446678fb1f5632ec7fd7e0a434fdd4fe2bad2914e8502fa3"} Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.162449 4809 scope.go:117] "RemoveContainer" containerID="a90c2aa3e3fe73387e9b264aedea8b1f549bfaa7885df6302588dd069fb8db47" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.186874 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hfn22\" (UniqueName: \"kubernetes.io/projected/eb1bd4de-6456-4817-a85d-ab0176153f2c-kube-api-access-hfn22\") on node \"crc\" DevicePath \"\"" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.186947 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-874rm\" (UniqueName: \"kubernetes.io/projected/475f8cc7-1795-455e-a448-734b3bc278a8-kube-api-access-874rm\") on node \"crc\" DevicePath \"\"" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.186962 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/475f8cc7-1795-455e-a448-734b3bc278a8-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.186974 4809 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/eb1bd4de-6456-4817-a85d-ab0176153f2c-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.186985 4809 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/475f8cc7-1795-455e-a448-734b3bc278a8-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.186997 4809 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/eb1bd4de-6456-4817-a85d-ab0176153f2c-pod-info\") on node \"crc\" DevicePath \"\"" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.187011 4809 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/eb1bd4de-6456-4817-a85d-ab0176153f2c-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.187037 4809 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.187051 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eb1bd4de-6456-4817-a85d-ab0176153f2c-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.187064 4809 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/eb1bd4de-6456-4817-a85d-ab0176153f2c-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.187077 4809 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/eb1bd4de-6456-4817-a85d-ab0176153f2c-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.187088 4809 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/eb1bd4de-6456-4817-a85d-ab0176153f2c-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.187715 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"475f8cc7-1795-455e-a448-734b3bc278a8","Type":"ContainerDied","Data":"1795100d44f0ac579862227ce9537745bad06a5a5e4c6b78eee551775f2cb823"} Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.187807 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.204532 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/475f8cc7-1795-455e-a448-734b3bc278a8-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "475f8cc7-1795-455e-a448-734b3bc278a8" (UID: "475f8cc7-1795-455e-a448-734b3bc278a8"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.221895 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/475f8cc7-1795-455e-a448-734b3bc278a8-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "475f8cc7-1795-455e-a448-734b3bc278a8" (UID: "475f8cc7-1795-455e-a448-734b3bc278a8"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.280916 4809 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.282565 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb1bd4de-6456-4817-a85d-ab0176153f2c-server-conf" (OuterVolumeSpecName: "server-conf") pod "eb1bd4de-6456-4817-a85d-ab0176153f2c" (UID: "eb1bd4de-6456-4817-a85d-ab0176153f2c"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.290709 4809 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.290746 4809 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/eb1bd4de-6456-4817-a85d-ab0176153f2c-server-conf\") on node \"crc\" DevicePath \"\"" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.290761 4809 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/475f8cc7-1795-455e-a448-734b3bc278a8-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.290772 4809 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/475f8cc7-1795-455e-a448-734b3bc278a8-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.305776 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/475f8cc7-1795-455e-a448-734b3bc278a8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "475f8cc7-1795-455e-a448-734b3bc278a8" (UID: "475f8cc7-1795-455e-a448-734b3bc278a8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.307397 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb1bd4de-6456-4817-a85d-ab0176153f2c-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "eb1bd4de-6456-4817-a85d-ab0176153f2c" (UID: "eb1bd4de-6456-4817-a85d-ab0176153f2c"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.339542 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/475f8cc7-1795-455e-a448-734b3bc278a8-config-data" (OuterVolumeSpecName: "config-data") pod "475f8cc7-1795-455e-a448-734b3bc278a8" (UID: "475f8cc7-1795-455e-a448-734b3bc278a8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.392313 4809 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/eb1bd4de-6456-4817-a85d-ab0176153f2c-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.392343 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/475f8cc7-1795-455e-a448-734b3bc278a8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.392353 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/475f8cc7-1795-455e-a448-734b3bc278a8-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.500219 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.528261 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.548649 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 06:22:07 crc kubenswrapper[4809]: E1206 06:22:07.549266 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="475f8cc7-1795-455e-a448-734b3bc278a8" containerName="proxy-httpd" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.549289 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="475f8cc7-1795-455e-a448-734b3bc278a8" containerName="proxy-httpd" Dec 06 06:22:07 crc kubenswrapper[4809]: E1206 06:22:07.549340 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="475f8cc7-1795-455e-a448-734b3bc278a8" containerName="sg-core" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.549351 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="475f8cc7-1795-455e-a448-734b3bc278a8" containerName="sg-core" Dec 06 06:22:07 crc kubenswrapper[4809]: E1206 06:22:07.549386 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="475f8cc7-1795-455e-a448-734b3bc278a8" containerName="ceilometer-notification-agent" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.549395 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="475f8cc7-1795-455e-a448-734b3bc278a8" containerName="ceilometer-notification-agent" Dec 06 06:22:07 crc kubenswrapper[4809]: E1206 06:22:07.549411 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb1bd4de-6456-4817-a85d-ab0176153f2c" containerName="setup-container" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.549419 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb1bd4de-6456-4817-a85d-ab0176153f2c" containerName="setup-container" Dec 06 06:22:07 crc kubenswrapper[4809]: E1206 06:22:07.549441 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="475f8cc7-1795-455e-a448-734b3bc278a8" containerName="ceilometer-central-agent" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.549449 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="475f8cc7-1795-455e-a448-734b3bc278a8" containerName="ceilometer-central-agent" Dec 06 06:22:07 crc kubenswrapper[4809]: E1206 06:22:07.549462 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb1bd4de-6456-4817-a85d-ab0176153f2c" containerName="rabbitmq" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.549469 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb1bd4de-6456-4817-a85d-ab0176153f2c" containerName="rabbitmq" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.549741 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="475f8cc7-1795-455e-a448-734b3bc278a8" containerName="ceilometer-central-agent" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.549763 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="475f8cc7-1795-455e-a448-734b3bc278a8" containerName="sg-core" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.549782 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="475f8cc7-1795-455e-a448-734b3bc278a8" containerName="ceilometer-notification-agent" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.549798 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb1bd4de-6456-4817-a85d-ab0176153f2c" containerName="rabbitmq" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.549819 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="475f8cc7-1795-455e-a448-734b3bc278a8" containerName="proxy-httpd" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.551523 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.555825 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.556065 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.556254 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.556410 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.556564 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-r9k2s" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.556717 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.565772 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.582082 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.599125 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2cc53bf7-6e54-427a-9b63-88d694609c75-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"2cc53bf7-6e54-427a-9b63-88d694609c75\") " pod="openstack/rabbitmq-server-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.599220 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2cc53bf7-6e54-427a-9b63-88d694609c75-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"2cc53bf7-6e54-427a-9b63-88d694609c75\") " pod="openstack/rabbitmq-server-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.599276 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2cc53bf7-6e54-427a-9b63-88d694609c75-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"2cc53bf7-6e54-427a-9b63-88d694609c75\") " pod="openstack/rabbitmq-server-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.599423 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2cc53bf7-6e54-427a-9b63-88d694609c75-server-conf\") pod \"rabbitmq-server-0\" (UID: \"2cc53bf7-6e54-427a-9b63-88d694609c75\") " pod="openstack/rabbitmq-server-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.599463 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2cc53bf7-6e54-427a-9b63-88d694609c75-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"2cc53bf7-6e54-427a-9b63-88d694609c75\") " pod="openstack/rabbitmq-server-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.599484 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2cc53bf7-6e54-427a-9b63-88d694609c75-pod-info\") pod \"rabbitmq-server-0\" (UID: \"2cc53bf7-6e54-427a-9b63-88d694609c75\") " pod="openstack/rabbitmq-server-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.599549 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2cc53bf7-6e54-427a-9b63-88d694609c75-config-data\") pod \"rabbitmq-server-0\" (UID: \"2cc53bf7-6e54-427a-9b63-88d694609c75\") " pod="openstack/rabbitmq-server-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.599734 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2cc53bf7-6e54-427a-9b63-88d694609c75-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"2cc53bf7-6e54-427a-9b63-88d694609c75\") " pod="openstack/rabbitmq-server-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.599792 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2cc53bf7-6e54-427a-9b63-88d694609c75-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"2cc53bf7-6e54-427a-9b63-88d694609c75\") " pod="openstack/rabbitmq-server-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.599869 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxrsx\" (UniqueName: \"kubernetes.io/projected/2cc53bf7-6e54-427a-9b63-88d694609c75-kube-api-access-cxrsx\") pod \"rabbitmq-server-0\" (UID: \"2cc53bf7-6e54-427a-9b63-88d694609c75\") " pod="openstack/rabbitmq-server-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.599893 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"2cc53bf7-6e54-427a-9b63-88d694609c75\") " pod="openstack/rabbitmq-server-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.612052 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.656797 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.675473 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.680164 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.683951 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.684066 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.684478 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.699048 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.701664 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2cc53bf7-6e54-427a-9b63-88d694609c75-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"2cc53bf7-6e54-427a-9b63-88d694609c75\") " pod="openstack/rabbitmq-server-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.701740 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2cc53bf7-6e54-427a-9b63-88d694609c75-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"2cc53bf7-6e54-427a-9b63-88d694609c75\") " pod="openstack/rabbitmq-server-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.701841 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxrsx\" (UniqueName: \"kubernetes.io/projected/2cc53bf7-6e54-427a-9b63-88d694609c75-kube-api-access-cxrsx\") pod \"rabbitmq-server-0\" (UID: \"2cc53bf7-6e54-427a-9b63-88d694609c75\") " pod="openstack/rabbitmq-server-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.704421 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"2cc53bf7-6e54-427a-9b63-88d694609c75\") " pod="openstack/rabbitmq-server-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.704512 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2cc53bf7-6e54-427a-9b63-88d694609c75-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"2cc53bf7-6e54-427a-9b63-88d694609c75\") " pod="openstack/rabbitmq-server-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.704545 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2cc53bf7-6e54-427a-9b63-88d694609c75-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"2cc53bf7-6e54-427a-9b63-88d694609c75\") " pod="openstack/rabbitmq-server-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.704609 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2cc53bf7-6e54-427a-9b63-88d694609c75-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"2cc53bf7-6e54-427a-9b63-88d694609c75\") " pod="openstack/rabbitmq-server-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.704657 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2cc53bf7-6e54-427a-9b63-88d694609c75-server-conf\") pod \"rabbitmq-server-0\" (UID: \"2cc53bf7-6e54-427a-9b63-88d694609c75\") " pod="openstack/rabbitmq-server-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.704728 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2cc53bf7-6e54-427a-9b63-88d694609c75-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"2cc53bf7-6e54-427a-9b63-88d694609c75\") " pod="openstack/rabbitmq-server-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.704758 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2cc53bf7-6e54-427a-9b63-88d694609c75-pod-info\") pod \"rabbitmq-server-0\" (UID: \"2cc53bf7-6e54-427a-9b63-88d694609c75\") " pod="openstack/rabbitmq-server-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.704901 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2cc53bf7-6e54-427a-9b63-88d694609c75-config-data\") pod \"rabbitmq-server-0\" (UID: \"2cc53bf7-6e54-427a-9b63-88d694609c75\") " pod="openstack/rabbitmq-server-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.705890 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"2cc53bf7-6e54-427a-9b63-88d694609c75\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-server-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.710964 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2cc53bf7-6e54-427a-9b63-88d694609c75-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"2cc53bf7-6e54-427a-9b63-88d694609c75\") " pod="openstack/rabbitmq-server-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.711044 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2cc53bf7-6e54-427a-9b63-88d694609c75-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"2cc53bf7-6e54-427a-9b63-88d694609c75\") " pod="openstack/rabbitmq-server-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.711858 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2cc53bf7-6e54-427a-9b63-88d694609c75-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"2cc53bf7-6e54-427a-9b63-88d694609c75\") " pod="openstack/rabbitmq-server-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.713057 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2cc53bf7-6e54-427a-9b63-88d694609c75-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"2cc53bf7-6e54-427a-9b63-88d694609c75\") " pod="openstack/rabbitmq-server-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.714840 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2cc53bf7-6e54-427a-9b63-88d694609c75-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"2cc53bf7-6e54-427a-9b63-88d694609c75\") " pod="openstack/rabbitmq-server-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.715302 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2cc53bf7-6e54-427a-9b63-88d694609c75-pod-info\") pod \"rabbitmq-server-0\" (UID: \"2cc53bf7-6e54-427a-9b63-88d694609c75\") " pod="openstack/rabbitmq-server-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.715627 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2cc53bf7-6e54-427a-9b63-88d694609c75-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"2cc53bf7-6e54-427a-9b63-88d694609c75\") " pod="openstack/rabbitmq-server-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.716005 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2cc53bf7-6e54-427a-9b63-88d694609c75-config-data\") pod \"rabbitmq-server-0\" (UID: \"2cc53bf7-6e54-427a-9b63-88d694609c75\") " pod="openstack/rabbitmq-server-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.717011 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2cc53bf7-6e54-427a-9b63-88d694609c75-server-conf\") pod \"rabbitmq-server-0\" (UID: \"2cc53bf7-6e54-427a-9b63-88d694609c75\") " pod="openstack/rabbitmq-server-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.722881 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxrsx\" (UniqueName: \"kubernetes.io/projected/2cc53bf7-6e54-427a-9b63-88d694609c75-kube-api-access-cxrsx\") pod \"rabbitmq-server-0\" (UID: \"2cc53bf7-6e54-427a-9b63-88d694609c75\") " pod="openstack/rabbitmq-server-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.768675 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"2cc53bf7-6e54-427a-9b63-88d694609c75\") " pod="openstack/rabbitmq-server-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.812162 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d4b42d3-04a4-40e3-803b-11c77024f612-config-data\") pod \"ceilometer-0\" (UID: \"3d4b42d3-04a4-40e3-803b-11c77024f612\") " pod="openstack/ceilometer-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.812262 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d4b42d3-04a4-40e3-803b-11c77024f612-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3d4b42d3-04a4-40e3-803b-11c77024f612\") " pod="openstack/ceilometer-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.812292 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3d4b42d3-04a4-40e3-803b-11c77024f612-run-httpd\") pod \"ceilometer-0\" (UID: \"3d4b42d3-04a4-40e3-803b-11c77024f612\") " pod="openstack/ceilometer-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.812362 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d4b42d3-04a4-40e3-803b-11c77024f612-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3d4b42d3-04a4-40e3-803b-11c77024f612\") " pod="openstack/ceilometer-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.812400 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3d4b42d3-04a4-40e3-803b-11c77024f612-log-httpd\") pod \"ceilometer-0\" (UID: \"3d4b42d3-04a4-40e3-803b-11c77024f612\") " pod="openstack/ceilometer-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.812428 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3d4b42d3-04a4-40e3-803b-11c77024f612-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3d4b42d3-04a4-40e3-803b-11c77024f612\") " pod="openstack/ceilometer-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.812469 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d4b42d3-04a4-40e3-803b-11c77024f612-scripts\") pod \"ceilometer-0\" (UID: \"3d4b42d3-04a4-40e3-803b-11c77024f612\") " pod="openstack/ceilometer-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.812564 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvzk9\" (UniqueName: \"kubernetes.io/projected/3d4b42d3-04a4-40e3-803b-11c77024f612-kube-api-access-pvzk9\") pod \"ceilometer-0\" (UID: \"3d4b42d3-04a4-40e3-803b-11c77024f612\") " pod="openstack/ceilometer-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.896877 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.914445 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d4b42d3-04a4-40e3-803b-11c77024f612-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3d4b42d3-04a4-40e3-803b-11c77024f612\") " pod="openstack/ceilometer-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.914518 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3d4b42d3-04a4-40e3-803b-11c77024f612-log-httpd\") pod \"ceilometer-0\" (UID: \"3d4b42d3-04a4-40e3-803b-11c77024f612\") " pod="openstack/ceilometer-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.914553 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3d4b42d3-04a4-40e3-803b-11c77024f612-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3d4b42d3-04a4-40e3-803b-11c77024f612\") " pod="openstack/ceilometer-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.914598 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d4b42d3-04a4-40e3-803b-11c77024f612-scripts\") pod \"ceilometer-0\" (UID: \"3d4b42d3-04a4-40e3-803b-11c77024f612\") " pod="openstack/ceilometer-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.914681 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvzk9\" (UniqueName: \"kubernetes.io/projected/3d4b42d3-04a4-40e3-803b-11c77024f612-kube-api-access-pvzk9\") pod \"ceilometer-0\" (UID: \"3d4b42d3-04a4-40e3-803b-11c77024f612\") " pod="openstack/ceilometer-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.914762 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d4b42d3-04a4-40e3-803b-11c77024f612-config-data\") pod \"ceilometer-0\" (UID: \"3d4b42d3-04a4-40e3-803b-11c77024f612\") " pod="openstack/ceilometer-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.914826 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d4b42d3-04a4-40e3-803b-11c77024f612-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3d4b42d3-04a4-40e3-803b-11c77024f612\") " pod="openstack/ceilometer-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.914851 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3d4b42d3-04a4-40e3-803b-11c77024f612-run-httpd\") pod \"ceilometer-0\" (UID: \"3d4b42d3-04a4-40e3-803b-11c77024f612\") " pod="openstack/ceilometer-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.915390 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3d4b42d3-04a4-40e3-803b-11c77024f612-run-httpd\") pod \"ceilometer-0\" (UID: \"3d4b42d3-04a4-40e3-803b-11c77024f612\") " pod="openstack/ceilometer-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.916759 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3d4b42d3-04a4-40e3-803b-11c77024f612-log-httpd\") pod \"ceilometer-0\" (UID: \"3d4b42d3-04a4-40e3-803b-11c77024f612\") " pod="openstack/ceilometer-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.921056 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d4b42d3-04a4-40e3-803b-11c77024f612-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3d4b42d3-04a4-40e3-803b-11c77024f612\") " pod="openstack/ceilometer-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.921974 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d4b42d3-04a4-40e3-803b-11c77024f612-config-data\") pod \"ceilometer-0\" (UID: \"3d4b42d3-04a4-40e3-803b-11c77024f612\") " pod="openstack/ceilometer-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.923415 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3d4b42d3-04a4-40e3-803b-11c77024f612-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3d4b42d3-04a4-40e3-803b-11c77024f612\") " pod="openstack/ceilometer-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.924958 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d4b42d3-04a4-40e3-803b-11c77024f612-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3d4b42d3-04a4-40e3-803b-11c77024f612\") " pod="openstack/ceilometer-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.926881 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d4b42d3-04a4-40e3-803b-11c77024f612-scripts\") pod \"ceilometer-0\" (UID: \"3d4b42d3-04a4-40e3-803b-11c77024f612\") " pod="openstack/ceilometer-0" Dec 06 06:22:07 crc kubenswrapper[4809]: I1206 06:22:07.938965 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvzk9\" (UniqueName: \"kubernetes.io/projected/3d4b42d3-04a4-40e3-803b-11c77024f612-kube-api-access-pvzk9\") pod \"ceilometer-0\" (UID: \"3d4b42d3-04a4-40e3-803b-11c77024f612\") " pod="openstack/ceilometer-0" Dec 06 06:22:08 crc kubenswrapper[4809]: I1206 06:22:08.006814 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:22:08 crc kubenswrapper[4809]: I1206 06:22:08.212680 4809 generic.go:334] "Generic (PLEG): container finished" podID="1ceabc13-38c5-4425-8d85-c3afa8935243" containerID="162e28f89b66909f6341a5b02785bb02066b04eb4d2c67d23537c8ea8505ae27" exitCode=0 Dec 06 06:22:08 crc kubenswrapper[4809]: I1206 06:22:08.212718 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1ceabc13-38c5-4425-8d85-c3afa8935243","Type":"ContainerDied","Data":"162e28f89b66909f6341a5b02785bb02066b04eb4d2c67d23537c8ea8505ae27"} Dec 06 06:22:09 crc kubenswrapper[4809]: I1206 06:22:09.171493 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="eb1bd4de-6456-4817-a85d-ab0176153f2c" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.126:5671: i/o timeout" Dec 06 06:22:09 crc kubenswrapper[4809]: I1206 06:22:09.406950 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="475f8cc7-1795-455e-a448-734b3bc278a8" path="/var/lib/kubelet/pods/475f8cc7-1795-455e-a448-734b3bc278a8/volumes" Dec 06 06:22:09 crc kubenswrapper[4809]: I1206 06:22:09.408404 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb1bd4de-6456-4817-a85d-ab0176153f2c" path="/var/lib/kubelet/pods/eb1bd4de-6456-4817-a85d-ab0176153f2c/volumes" Dec 06 06:22:14 crc kubenswrapper[4809]: I1206 06:22:14.609618 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="1ceabc13-38c5-4425-8d85-c3afa8935243" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.127:5671: connect: connection refused" Dec 06 06:22:14 crc kubenswrapper[4809]: I1206 06:22:14.770157 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-j5rkj"] Dec 06 06:22:14 crc kubenswrapper[4809]: I1206 06:22:14.772507 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-j5rkj" Dec 06 06:22:14 crc kubenswrapper[4809]: I1206 06:22:14.775496 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 06 06:22:14 crc kubenswrapper[4809]: I1206 06:22:14.793278 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-j5rkj"] Dec 06 06:22:14 crc kubenswrapper[4809]: I1206 06:22:14.865628 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8f337bd4-4533-4599-8cb8-7bedb8e2c420-ovsdbserver-sb\") pod \"dnsmasq-dns-7d84b4d45c-j5rkj\" (UID: \"8f337bd4-4533-4599-8cb8-7bedb8e2c420\") " pod="openstack/dnsmasq-dns-7d84b4d45c-j5rkj" Dec 06 06:22:14 crc kubenswrapper[4809]: I1206 06:22:14.865696 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8f337bd4-4533-4599-8cb8-7bedb8e2c420-dns-swift-storage-0\") pod \"dnsmasq-dns-7d84b4d45c-j5rkj\" (UID: \"8f337bd4-4533-4599-8cb8-7bedb8e2c420\") " pod="openstack/dnsmasq-dns-7d84b4d45c-j5rkj" Dec 06 06:22:14 crc kubenswrapper[4809]: I1206 06:22:14.865727 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/8f337bd4-4533-4599-8cb8-7bedb8e2c420-openstack-edpm-ipam\") pod \"dnsmasq-dns-7d84b4d45c-j5rkj\" (UID: \"8f337bd4-4533-4599-8cb8-7bedb8e2c420\") " pod="openstack/dnsmasq-dns-7d84b4d45c-j5rkj" Dec 06 06:22:14 crc kubenswrapper[4809]: I1206 06:22:14.865816 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f337bd4-4533-4599-8cb8-7bedb8e2c420-config\") pod \"dnsmasq-dns-7d84b4d45c-j5rkj\" (UID: \"8f337bd4-4533-4599-8cb8-7bedb8e2c420\") " pod="openstack/dnsmasq-dns-7d84b4d45c-j5rkj" Dec 06 06:22:14 crc kubenswrapper[4809]: I1206 06:22:14.865882 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8l88f\" (UniqueName: \"kubernetes.io/projected/8f337bd4-4533-4599-8cb8-7bedb8e2c420-kube-api-access-8l88f\") pod \"dnsmasq-dns-7d84b4d45c-j5rkj\" (UID: \"8f337bd4-4533-4599-8cb8-7bedb8e2c420\") " pod="openstack/dnsmasq-dns-7d84b4d45c-j5rkj" Dec 06 06:22:14 crc kubenswrapper[4809]: I1206 06:22:14.865903 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8f337bd4-4533-4599-8cb8-7bedb8e2c420-dns-svc\") pod \"dnsmasq-dns-7d84b4d45c-j5rkj\" (UID: \"8f337bd4-4533-4599-8cb8-7bedb8e2c420\") " pod="openstack/dnsmasq-dns-7d84b4d45c-j5rkj" Dec 06 06:22:14 crc kubenswrapper[4809]: I1206 06:22:14.865920 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8f337bd4-4533-4599-8cb8-7bedb8e2c420-ovsdbserver-nb\") pod \"dnsmasq-dns-7d84b4d45c-j5rkj\" (UID: \"8f337bd4-4533-4599-8cb8-7bedb8e2c420\") " pod="openstack/dnsmasq-dns-7d84b4d45c-j5rkj" Dec 06 06:22:14 crc kubenswrapper[4809]: I1206 06:22:14.968725 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8f337bd4-4533-4599-8cb8-7bedb8e2c420-ovsdbserver-sb\") pod \"dnsmasq-dns-7d84b4d45c-j5rkj\" (UID: \"8f337bd4-4533-4599-8cb8-7bedb8e2c420\") " pod="openstack/dnsmasq-dns-7d84b4d45c-j5rkj" Dec 06 06:22:14 crc kubenswrapper[4809]: I1206 06:22:14.968802 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8f337bd4-4533-4599-8cb8-7bedb8e2c420-dns-swift-storage-0\") pod \"dnsmasq-dns-7d84b4d45c-j5rkj\" (UID: \"8f337bd4-4533-4599-8cb8-7bedb8e2c420\") " pod="openstack/dnsmasq-dns-7d84b4d45c-j5rkj" Dec 06 06:22:14 crc kubenswrapper[4809]: I1206 06:22:14.968845 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/8f337bd4-4533-4599-8cb8-7bedb8e2c420-openstack-edpm-ipam\") pod \"dnsmasq-dns-7d84b4d45c-j5rkj\" (UID: \"8f337bd4-4533-4599-8cb8-7bedb8e2c420\") " pod="openstack/dnsmasq-dns-7d84b4d45c-j5rkj" Dec 06 06:22:14 crc kubenswrapper[4809]: I1206 06:22:14.969144 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f337bd4-4533-4599-8cb8-7bedb8e2c420-config\") pod \"dnsmasq-dns-7d84b4d45c-j5rkj\" (UID: \"8f337bd4-4533-4599-8cb8-7bedb8e2c420\") " pod="openstack/dnsmasq-dns-7d84b4d45c-j5rkj" Dec 06 06:22:14 crc kubenswrapper[4809]: I1206 06:22:14.969250 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8l88f\" (UniqueName: \"kubernetes.io/projected/8f337bd4-4533-4599-8cb8-7bedb8e2c420-kube-api-access-8l88f\") pod \"dnsmasq-dns-7d84b4d45c-j5rkj\" (UID: \"8f337bd4-4533-4599-8cb8-7bedb8e2c420\") " pod="openstack/dnsmasq-dns-7d84b4d45c-j5rkj" Dec 06 06:22:14 crc kubenswrapper[4809]: I1206 06:22:14.969283 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8f337bd4-4533-4599-8cb8-7bedb8e2c420-dns-svc\") pod \"dnsmasq-dns-7d84b4d45c-j5rkj\" (UID: \"8f337bd4-4533-4599-8cb8-7bedb8e2c420\") " pod="openstack/dnsmasq-dns-7d84b4d45c-j5rkj" Dec 06 06:22:14 crc kubenswrapper[4809]: I1206 06:22:14.969311 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8f337bd4-4533-4599-8cb8-7bedb8e2c420-ovsdbserver-nb\") pod \"dnsmasq-dns-7d84b4d45c-j5rkj\" (UID: \"8f337bd4-4533-4599-8cb8-7bedb8e2c420\") " pod="openstack/dnsmasq-dns-7d84b4d45c-j5rkj" Dec 06 06:22:14 crc kubenswrapper[4809]: I1206 06:22:14.970067 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8f337bd4-4533-4599-8cb8-7bedb8e2c420-ovsdbserver-sb\") pod \"dnsmasq-dns-7d84b4d45c-j5rkj\" (UID: \"8f337bd4-4533-4599-8cb8-7bedb8e2c420\") " pod="openstack/dnsmasq-dns-7d84b4d45c-j5rkj" Dec 06 06:22:14 crc kubenswrapper[4809]: I1206 06:22:14.970231 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f337bd4-4533-4599-8cb8-7bedb8e2c420-config\") pod \"dnsmasq-dns-7d84b4d45c-j5rkj\" (UID: \"8f337bd4-4533-4599-8cb8-7bedb8e2c420\") " pod="openstack/dnsmasq-dns-7d84b4d45c-j5rkj" Dec 06 06:22:14 crc kubenswrapper[4809]: I1206 06:22:14.970249 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8f337bd4-4533-4599-8cb8-7bedb8e2c420-dns-svc\") pod \"dnsmasq-dns-7d84b4d45c-j5rkj\" (UID: \"8f337bd4-4533-4599-8cb8-7bedb8e2c420\") " pod="openstack/dnsmasq-dns-7d84b4d45c-j5rkj" Dec 06 06:22:14 crc kubenswrapper[4809]: I1206 06:22:14.970403 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8f337bd4-4533-4599-8cb8-7bedb8e2c420-ovsdbserver-nb\") pod \"dnsmasq-dns-7d84b4d45c-j5rkj\" (UID: \"8f337bd4-4533-4599-8cb8-7bedb8e2c420\") " pod="openstack/dnsmasq-dns-7d84b4d45c-j5rkj" Dec 06 06:22:14 crc kubenswrapper[4809]: I1206 06:22:14.972562 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/8f337bd4-4533-4599-8cb8-7bedb8e2c420-openstack-edpm-ipam\") pod \"dnsmasq-dns-7d84b4d45c-j5rkj\" (UID: \"8f337bd4-4533-4599-8cb8-7bedb8e2c420\") " pod="openstack/dnsmasq-dns-7d84b4d45c-j5rkj" Dec 06 06:22:14 crc kubenswrapper[4809]: I1206 06:22:14.974014 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8f337bd4-4533-4599-8cb8-7bedb8e2c420-dns-swift-storage-0\") pod \"dnsmasq-dns-7d84b4d45c-j5rkj\" (UID: \"8f337bd4-4533-4599-8cb8-7bedb8e2c420\") " pod="openstack/dnsmasq-dns-7d84b4d45c-j5rkj" Dec 06 06:22:15 crc kubenswrapper[4809]: I1206 06:22:15.000868 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8l88f\" (UniqueName: \"kubernetes.io/projected/8f337bd4-4533-4599-8cb8-7bedb8e2c420-kube-api-access-8l88f\") pod \"dnsmasq-dns-7d84b4d45c-j5rkj\" (UID: \"8f337bd4-4533-4599-8cb8-7bedb8e2c420\") " pod="openstack/dnsmasq-dns-7d84b4d45c-j5rkj" Dec 06 06:22:15 crc kubenswrapper[4809]: I1206 06:22:15.100677 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-j5rkj" Dec 06 06:22:17 crc kubenswrapper[4809]: I1206 06:22:17.391194 4809 scope.go:117] "RemoveContainer" containerID="9bc2dc88b3b9f309ce0169aa510326dd77022dece4c69cab3141df00fe5c0106" Dec 06 06:22:17 crc kubenswrapper[4809]: E1206 06:22:17.392097 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:22:18 crc kubenswrapper[4809]: I1206 06:22:18.668251 4809 scope.go:117] "RemoveContainer" containerID="cd1fea316ddc0d8b812b68b8ae2836932648f88ecde0d2a0cb057eb24d8002a9" Dec 06 06:22:20 crc kubenswrapper[4809]: I1206 06:22:20.447907 4809 scope.go:117] "RemoveContainer" containerID="cd1fea316ddc0d8b812b68b8ae2836932648f88ecde0d2a0cb057eb24d8002a9" Dec 06 06:22:20 crc kubenswrapper[4809]: E1206 06:22:20.636958 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 06 06:22:20 crc kubenswrapper[4809]: E1206 06:22:20.637037 4809 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 06 06:22:20 crc kubenswrapper[4809]: E1206 06:22:20.637213 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s5vwn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-ng4qq_openstack(18fa982c-0572-4702-8c51-f0f7af7bf52f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 06:22:20 crc kubenswrapper[4809]: E1206 06:22:20.639162 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/heat-db-sync-ng4qq" podUID="18fa982c-0572-4702-8c51-f0f7af7bf52f" Dec 06 06:22:20 crc kubenswrapper[4809]: I1206 06:22:20.697462 4809 scope.go:117] "RemoveContainer" containerID="e4e8d2c19a1b6d0c77a22e69bd9bb836d9d3511590e0116c8d1a49be5bc196be" Dec 06 06:22:20 crc kubenswrapper[4809]: E1206 06:22:20.739639 4809 log.go:32] "RemoveContainer from runtime service failed" err="rpc error: code = Unknown desc = failed to delete container k8s_setup-container_rabbitmq-server-0_openstack_eb1bd4de-6456-4817-a85d-ab0176153f2c_0 in pod sandbox e04ab83414709672446678fb1f5632ec7fd7e0a434fdd4fe2bad2914e8502fa3: identifier is not a container" containerID="cd1fea316ddc0d8b812b68b8ae2836932648f88ecde0d2a0cb057eb24d8002a9" Dec 06 06:22:20 crc kubenswrapper[4809]: I1206 06:22:20.739718 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd1fea316ddc0d8b812b68b8ae2836932648f88ecde0d2a0cb057eb24d8002a9"} err="rpc error: code = Unknown desc = failed to delete container k8s_setup-container_rabbitmq-server-0_openstack_eb1bd4de-6456-4817-a85d-ab0176153f2c_0 in pod sandbox e04ab83414709672446678fb1f5632ec7fd7e0a434fdd4fe2bad2914e8502fa3: identifier is not a container" Dec 06 06:22:20 crc kubenswrapper[4809]: I1206 06:22:20.739744 4809 scope.go:117] "RemoveContainer" containerID="c005ef42f1cd2a1c17707563dbe907644a57297c90f9f56ad9a2e907a171a8bc" Dec 06 06:22:20 crc kubenswrapper[4809]: I1206 06:22:20.816075 4809 scope.go:117] "RemoveContainer" containerID="f353b221a72c907f42a2fc4d8be666369112112697a23d895465668987d9ac5b" Dec 06 06:22:20 crc kubenswrapper[4809]: I1206 06:22:20.910744 4809 scope.go:117] "RemoveContainer" containerID="bb98373f270bc6b48302c14e5bb82802f71cddde648d84ddca4feb136955cbbb" Dec 06 06:22:20 crc kubenswrapper[4809]: I1206 06:22:20.945240 4809 scope.go:117] "RemoveContainer" containerID="a90c2aa3e3fe73387e9b264aedea8b1f549bfaa7885df6302588dd069fb8db47" Dec 06 06:22:20 crc kubenswrapper[4809]: E1206 06:22:20.947502 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a90c2aa3e3fe73387e9b264aedea8b1f549bfaa7885df6302588dd069fb8db47\": container with ID starting with a90c2aa3e3fe73387e9b264aedea8b1f549bfaa7885df6302588dd069fb8db47 not found: ID does not exist" containerID="a90c2aa3e3fe73387e9b264aedea8b1f549bfaa7885df6302588dd069fb8db47" Dec 06 06:22:20 crc kubenswrapper[4809]: E1206 06:22:20.947534 4809 kuberuntime_gc.go:150] "Failed to remove container" err="failed to get container status \"a90c2aa3e3fe73387e9b264aedea8b1f549bfaa7885df6302588dd069fb8db47\": rpc error: code = NotFound desc = could not find container \"a90c2aa3e3fe73387e9b264aedea8b1f549bfaa7885df6302588dd069fb8db47\": container with ID starting with a90c2aa3e3fe73387e9b264aedea8b1f549bfaa7885df6302588dd069fb8db47 not found: ID does not exist" containerID="a90c2aa3e3fe73387e9b264aedea8b1f549bfaa7885df6302588dd069fb8db47" Dec 06 06:22:20 crc kubenswrapper[4809]: I1206 06:22:20.959664 4809 scope.go:117] "RemoveContainer" containerID="569d10abcb4d5547b8dce30fe53ec252da0e9edb9c96e787021ef0b8e65589d6" Dec 06 06:22:20 crc kubenswrapper[4809]: I1206 06:22:20.992637 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.018136 4809 scope.go:117] "RemoveContainer" containerID="5d8d717f089ed3b927d6066c8f54271cf050ddf80f80dd72782cd64fb182206b" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.136532 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1ceabc13-38c5-4425-8d85-c3afa8935243-rabbitmq-confd\") pod \"1ceabc13-38c5-4425-8d85-c3afa8935243\" (UID: \"1ceabc13-38c5-4425-8d85-c3afa8935243\") " Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.136921 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1ceabc13-38c5-4425-8d85-c3afa8935243-pod-info\") pod \"1ceabc13-38c5-4425-8d85-c3afa8935243\" (UID: \"1ceabc13-38c5-4425-8d85-c3afa8935243\") " Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.136967 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1ceabc13-38c5-4425-8d85-c3afa8935243-server-conf\") pod \"1ceabc13-38c5-4425-8d85-c3afa8935243\" (UID: \"1ceabc13-38c5-4425-8d85-c3afa8935243\") " Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.137084 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1ceabc13-38c5-4425-8d85-c3afa8935243-config-data\") pod \"1ceabc13-38c5-4425-8d85-c3afa8935243\" (UID: \"1ceabc13-38c5-4425-8d85-c3afa8935243\") " Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.137124 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"1ceabc13-38c5-4425-8d85-c3afa8935243\" (UID: \"1ceabc13-38c5-4425-8d85-c3afa8935243\") " Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.137147 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1ceabc13-38c5-4425-8d85-c3afa8935243-plugins-conf\") pod \"1ceabc13-38c5-4425-8d85-c3afa8935243\" (UID: \"1ceabc13-38c5-4425-8d85-c3afa8935243\") " Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.137279 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1ceabc13-38c5-4425-8d85-c3afa8935243-rabbitmq-erlang-cookie\") pod \"1ceabc13-38c5-4425-8d85-c3afa8935243\" (UID: \"1ceabc13-38c5-4425-8d85-c3afa8935243\") " Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.137315 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf4mc\" (UniqueName: \"kubernetes.io/projected/1ceabc13-38c5-4425-8d85-c3afa8935243-kube-api-access-gf4mc\") pod \"1ceabc13-38c5-4425-8d85-c3afa8935243\" (UID: \"1ceabc13-38c5-4425-8d85-c3afa8935243\") " Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.137353 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1ceabc13-38c5-4425-8d85-c3afa8935243-erlang-cookie-secret\") pod \"1ceabc13-38c5-4425-8d85-c3afa8935243\" (UID: \"1ceabc13-38c5-4425-8d85-c3afa8935243\") " Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.137391 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1ceabc13-38c5-4425-8d85-c3afa8935243-rabbitmq-tls\") pod \"1ceabc13-38c5-4425-8d85-c3afa8935243\" (UID: \"1ceabc13-38c5-4425-8d85-c3afa8935243\") " Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.137413 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1ceabc13-38c5-4425-8d85-c3afa8935243-rabbitmq-plugins\") pod \"1ceabc13-38c5-4425-8d85-c3afa8935243\" (UID: \"1ceabc13-38c5-4425-8d85-c3afa8935243\") " Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.138530 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ceabc13-38c5-4425-8d85-c3afa8935243-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "1ceabc13-38c5-4425-8d85-c3afa8935243" (UID: "1ceabc13-38c5-4425-8d85-c3afa8935243"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.139461 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ceabc13-38c5-4425-8d85-c3afa8935243-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "1ceabc13-38c5-4425-8d85-c3afa8935243" (UID: "1ceabc13-38c5-4425-8d85-c3afa8935243"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.139576 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ceabc13-38c5-4425-8d85-c3afa8935243-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "1ceabc13-38c5-4425-8d85-c3afa8935243" (UID: "1ceabc13-38c5-4425-8d85-c3afa8935243"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.145520 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "persistence") pod "1ceabc13-38c5-4425-8d85-c3afa8935243" (UID: "1ceabc13-38c5-4425-8d85-c3afa8935243"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.145609 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ceabc13-38c5-4425-8d85-c3afa8935243-kube-api-access-gf4mc" (OuterVolumeSpecName: "kube-api-access-gf4mc") pod "1ceabc13-38c5-4425-8d85-c3afa8935243" (UID: "1ceabc13-38c5-4425-8d85-c3afa8935243"). InnerVolumeSpecName "kube-api-access-gf4mc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.145668 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ceabc13-38c5-4425-8d85-c3afa8935243-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "1ceabc13-38c5-4425-8d85-c3afa8935243" (UID: "1ceabc13-38c5-4425-8d85-c3afa8935243"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.148394 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ceabc13-38c5-4425-8d85-c3afa8935243-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "1ceabc13-38c5-4425-8d85-c3afa8935243" (UID: "1ceabc13-38c5-4425-8d85-c3afa8935243"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.152052 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/1ceabc13-38c5-4425-8d85-c3afa8935243-pod-info" (OuterVolumeSpecName: "pod-info") pod "1ceabc13-38c5-4425-8d85-c3afa8935243" (UID: "1ceabc13-38c5-4425-8d85-c3afa8935243"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.176359 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ceabc13-38c5-4425-8d85-c3afa8935243-config-data" (OuterVolumeSpecName: "config-data") pod "1ceabc13-38c5-4425-8d85-c3afa8935243" (UID: "1ceabc13-38c5-4425-8d85-c3afa8935243"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.215614 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ceabc13-38c5-4425-8d85-c3afa8935243-server-conf" (OuterVolumeSpecName: "server-conf") pod "1ceabc13-38c5-4425-8d85-c3afa8935243" (UID: "1ceabc13-38c5-4425-8d85-c3afa8935243"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.240445 4809 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1ceabc13-38c5-4425-8d85-c3afa8935243-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.240477 4809 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1ceabc13-38c5-4425-8d85-c3afa8935243-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.240489 4809 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1ceabc13-38c5-4425-8d85-c3afa8935243-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.240500 4809 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1ceabc13-38c5-4425-8d85-c3afa8935243-pod-info\") on node \"crc\" DevicePath \"\"" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.240514 4809 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1ceabc13-38c5-4425-8d85-c3afa8935243-server-conf\") on node \"crc\" DevicePath \"\"" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.240525 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1ceabc13-38c5-4425-8d85-c3afa8935243-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.240561 4809 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.240575 4809 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1ceabc13-38c5-4425-8d85-c3afa8935243-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.240588 4809 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1ceabc13-38c5-4425-8d85-c3afa8935243-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.240603 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf4mc\" (UniqueName: \"kubernetes.io/projected/1ceabc13-38c5-4425-8d85-c3afa8935243-kube-api-access-gf4mc\") on node \"crc\" DevicePath \"\"" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.273581 4809 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.296945 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ceabc13-38c5-4425-8d85-c3afa8935243-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "1ceabc13-38c5-4425-8d85-c3afa8935243" (UID: "1ceabc13-38c5-4425-8d85-c3afa8935243"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.343701 4809 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1ceabc13-38c5-4425-8d85-c3afa8935243-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.343742 4809 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.429484 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1ceabc13-38c5-4425-8d85-c3afa8935243","Type":"ContainerDied","Data":"d8209ae66e9765531af7b6a5fecfd56f9b7c3444a84bdbc8c222f43b3642712e"} Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.429565 4809 scope.go:117] "RemoveContainer" containerID="162e28f89b66909f6341a5b02785bb02066b04eb4d2c67d23537c8ea8505ae27" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.429502 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.441066 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lkm6p" event={"ID":"81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf","Type":"ContainerStarted","Data":"a4c656c4e48f2889f600c3e5ad29ed503a3315ff146bb9bf3f02bb05468d2fbb"} Dec 06 06:22:21 crc kubenswrapper[4809]: E1206 06:22:21.442637 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-ng4qq" podUID="18fa982c-0572-4702-8c51-f0f7af7bf52f" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.479066 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-j5rkj"] Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.491482 4809 scope.go:117] "RemoveContainer" containerID="eb40e5f3076606ef761086938cb7e8b4aad3ddeaa2e7eb9a9934ed5936e19c64" Dec 06 06:22:21 crc kubenswrapper[4809]: W1206 06:22:21.495550 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f337bd4_4533_4599_8cb8_7bedb8e2c420.slice/crio-4e09cc8c2323be25bad6170beae1a37be0bf5c1c3ec7d35f63115dac2c9e5f53 WatchSource:0}: Error finding container 4e09cc8c2323be25bad6170beae1a37be0bf5c1c3ec7d35f63115dac2c9e5f53: Status 404 returned error can't find the container with id 4e09cc8c2323be25bad6170beae1a37be0bf5c1c3ec7d35f63115dac2c9e5f53 Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.542430 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.592536 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.617587 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.645026 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.661533 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 06:22:21 crc kubenswrapper[4809]: E1206 06:22:21.663606 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ceabc13-38c5-4425-8d85-c3afa8935243" containerName="setup-container" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.663628 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ceabc13-38c5-4425-8d85-c3afa8935243" containerName="setup-container" Dec 06 06:22:21 crc kubenswrapper[4809]: E1206 06:22:21.663684 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ceabc13-38c5-4425-8d85-c3afa8935243" containerName="rabbitmq" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.663694 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ceabc13-38c5-4425-8d85-c3afa8935243" containerName="rabbitmq" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.664118 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ceabc13-38c5-4425-8d85-c3afa8935243" containerName="rabbitmq" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.665855 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.671259 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.671644 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.671889 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.671958 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.672329 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.673167 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.674414 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-tqjgv" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.684441 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.765565 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkdd7\" (UniqueName: \"kubernetes.io/projected/20646b41-62ab-42e7-8b11-538605db0506-kube-api-access-lkdd7\") pod \"rabbitmq-cell1-server-0\" (UID: \"20646b41-62ab-42e7-8b11-538605db0506\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.765625 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/20646b41-62ab-42e7-8b11-538605db0506-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"20646b41-62ab-42e7-8b11-538605db0506\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.765651 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/20646b41-62ab-42e7-8b11-538605db0506-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"20646b41-62ab-42e7-8b11-538605db0506\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.765669 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/20646b41-62ab-42e7-8b11-538605db0506-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"20646b41-62ab-42e7-8b11-538605db0506\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.765692 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/20646b41-62ab-42e7-8b11-538605db0506-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"20646b41-62ab-42e7-8b11-538605db0506\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.765717 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/20646b41-62ab-42e7-8b11-538605db0506-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"20646b41-62ab-42e7-8b11-538605db0506\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.765887 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/20646b41-62ab-42e7-8b11-538605db0506-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"20646b41-62ab-42e7-8b11-538605db0506\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.765971 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/20646b41-62ab-42e7-8b11-538605db0506-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"20646b41-62ab-42e7-8b11-538605db0506\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.766035 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"20646b41-62ab-42e7-8b11-538605db0506\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.766064 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/20646b41-62ab-42e7-8b11-538605db0506-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"20646b41-62ab-42e7-8b11-538605db0506\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.766161 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/20646b41-62ab-42e7-8b11-538605db0506-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"20646b41-62ab-42e7-8b11-538605db0506\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.868286 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/20646b41-62ab-42e7-8b11-538605db0506-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"20646b41-62ab-42e7-8b11-538605db0506\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.868366 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/20646b41-62ab-42e7-8b11-538605db0506-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"20646b41-62ab-42e7-8b11-538605db0506\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.868434 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"20646b41-62ab-42e7-8b11-538605db0506\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.868470 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/20646b41-62ab-42e7-8b11-538605db0506-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"20646b41-62ab-42e7-8b11-538605db0506\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.868557 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/20646b41-62ab-42e7-8b11-538605db0506-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"20646b41-62ab-42e7-8b11-538605db0506\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.868607 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkdd7\" (UniqueName: \"kubernetes.io/projected/20646b41-62ab-42e7-8b11-538605db0506-kube-api-access-lkdd7\") pod \"rabbitmq-cell1-server-0\" (UID: \"20646b41-62ab-42e7-8b11-538605db0506\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.868641 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/20646b41-62ab-42e7-8b11-538605db0506-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"20646b41-62ab-42e7-8b11-538605db0506\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.868676 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/20646b41-62ab-42e7-8b11-538605db0506-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"20646b41-62ab-42e7-8b11-538605db0506\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.868700 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/20646b41-62ab-42e7-8b11-538605db0506-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"20646b41-62ab-42e7-8b11-538605db0506\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.868728 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/20646b41-62ab-42e7-8b11-538605db0506-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"20646b41-62ab-42e7-8b11-538605db0506\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.868735 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"20646b41-62ab-42e7-8b11-538605db0506\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.869463 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/20646b41-62ab-42e7-8b11-538605db0506-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"20646b41-62ab-42e7-8b11-538605db0506\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.870411 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/20646b41-62ab-42e7-8b11-538605db0506-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"20646b41-62ab-42e7-8b11-538605db0506\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.873032 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/20646b41-62ab-42e7-8b11-538605db0506-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"20646b41-62ab-42e7-8b11-538605db0506\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.873139 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/20646b41-62ab-42e7-8b11-538605db0506-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"20646b41-62ab-42e7-8b11-538605db0506\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.873515 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/20646b41-62ab-42e7-8b11-538605db0506-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"20646b41-62ab-42e7-8b11-538605db0506\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.868758 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/20646b41-62ab-42e7-8b11-538605db0506-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"20646b41-62ab-42e7-8b11-538605db0506\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.873599 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/20646b41-62ab-42e7-8b11-538605db0506-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"20646b41-62ab-42e7-8b11-538605db0506\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.876556 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/20646b41-62ab-42e7-8b11-538605db0506-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"20646b41-62ab-42e7-8b11-538605db0506\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.891742 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkdd7\" (UniqueName: \"kubernetes.io/projected/20646b41-62ab-42e7-8b11-538605db0506-kube-api-access-lkdd7\") pod \"rabbitmq-cell1-server-0\" (UID: \"20646b41-62ab-42e7-8b11-538605db0506\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.916804 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/20646b41-62ab-42e7-8b11-538605db0506-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"20646b41-62ab-42e7-8b11-538605db0506\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.917207 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/20646b41-62ab-42e7-8b11-538605db0506-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"20646b41-62ab-42e7-8b11-538605db0506\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:22:21 crc kubenswrapper[4809]: I1206 06:22:21.919679 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"20646b41-62ab-42e7-8b11-538605db0506\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:22:22 crc kubenswrapper[4809]: I1206 06:22:22.036280 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:22:22 crc kubenswrapper[4809]: I1206 06:22:22.510557 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2cc53bf7-6e54-427a-9b63-88d694609c75","Type":"ContainerStarted","Data":"92d4c7d11ca2c0965f5119f3fbf03b90dc83a26b883e2c96a0572eb4c4995b1f"} Dec 06 06:22:22 crc kubenswrapper[4809]: I1206 06:22:22.525157 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3d4b42d3-04a4-40e3-803b-11c77024f612","Type":"ContainerStarted","Data":"a8230327f933146f837ab8150b34c63856481bed9f0cfc001a394c94f516fded"} Dec 06 06:22:22 crc kubenswrapper[4809]: I1206 06:22:22.546214 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-j5rkj" event={"ID":"8f337bd4-4533-4599-8cb8-7bedb8e2c420","Type":"ContainerStarted","Data":"4e09cc8c2323be25bad6170beae1a37be0bf5c1c3ec7d35f63115dac2c9e5f53"} Dec 06 06:22:23 crc kubenswrapper[4809]: I1206 06:22:23.406472 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ceabc13-38c5-4425-8d85-c3afa8935243" path="/var/lib/kubelet/pods/1ceabc13-38c5-4425-8d85-c3afa8935243/volumes" Dec 06 06:22:23 crc kubenswrapper[4809]: W1206 06:22:23.546336 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod20646b41_62ab_42e7_8b11_538605db0506.slice/crio-8f6e3ec3ccd49d409cd74230095ea6acdd4d78b82912bae23fdabba494badccd WatchSource:0}: Error finding container 8f6e3ec3ccd49d409cd74230095ea6acdd4d78b82912bae23fdabba494badccd: Status 404 returned error can't find the container with id 8f6e3ec3ccd49d409cd74230095ea6acdd4d78b82912bae23fdabba494badccd Dec 06 06:22:24 crc kubenswrapper[4809]: I1206 06:22:24.123166 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 06:22:24 crc kubenswrapper[4809]: I1206 06:22:24.123518 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"20646b41-62ab-42e7-8b11-538605db0506","Type":"ContainerStarted","Data":"8f6e3ec3ccd49d409cd74230095ea6acdd4d78b82912bae23fdabba494badccd"} Dec 06 06:22:25 crc kubenswrapper[4809]: I1206 06:22:25.613590 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-j5rkj" event={"ID":"8f337bd4-4533-4599-8cb8-7bedb8e2c420","Type":"ContainerStarted","Data":"e5cb2803c48c66ea417847983d79cf128cf6ebf3a3d21ece49b50e6aba9b35d6"} Dec 06 06:22:27 crc kubenswrapper[4809]: I1206 06:22:27.652921 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"20646b41-62ab-42e7-8b11-538605db0506","Type":"ContainerStarted","Data":"7fdbe581e08cf2fa165721423a7f859c28b8d58546515e9d601549721fefa0af"} Dec 06 06:22:27 crc kubenswrapper[4809]: I1206 06:22:27.654742 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2cc53bf7-6e54-427a-9b63-88d694609c75","Type":"ContainerStarted","Data":"929ef76322eb9f4df027d760ed9c5f62890de92642910bf48ccd63f92cb62498"} Dec 06 06:22:27 crc kubenswrapper[4809]: I1206 06:22:27.656637 4809 generic.go:334] "Generic (PLEG): container finished" podID="8f337bd4-4533-4599-8cb8-7bedb8e2c420" containerID="e5cb2803c48c66ea417847983d79cf128cf6ebf3a3d21ece49b50e6aba9b35d6" exitCode=0 Dec 06 06:22:27 crc kubenswrapper[4809]: I1206 06:22:27.656681 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-j5rkj" event={"ID":"8f337bd4-4533-4599-8cb8-7bedb8e2c420","Type":"ContainerDied","Data":"e5cb2803c48c66ea417847983d79cf128cf6ebf3a3d21ece49b50e6aba9b35d6"} Dec 06 06:22:28 crc kubenswrapper[4809]: I1206 06:22:28.668096 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-j5rkj" event={"ID":"8f337bd4-4533-4599-8cb8-7bedb8e2c420","Type":"ContainerStarted","Data":"18cc26322bc53a4671779f1e298b36146b2936a6e5533b627269f5818312c95c"} Dec 06 06:22:28 crc kubenswrapper[4809]: I1206 06:22:28.726689 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7d84b4d45c-j5rkj" podStartSLOduration=14.726672775 podStartE2EDuration="14.726672775s" podCreationTimestamp="2025-12-06 06:22:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:22:28.721722691 +0000 UTC m=+1873.610705643" watchObservedRunningTime="2025-12-06 06:22:28.726672775 +0000 UTC m=+1873.615655717" Dec 06 06:22:29 crc kubenswrapper[4809]: I1206 06:22:29.392736 4809 scope.go:117] "RemoveContainer" containerID="9bc2dc88b3b9f309ce0169aa510326dd77022dece4c69cab3141df00fe5c0106" Dec 06 06:22:29 crc kubenswrapper[4809]: E1206 06:22:29.393264 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:22:29 crc kubenswrapper[4809]: I1206 06:22:29.677668 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7d84b4d45c-j5rkj" Dec 06 06:22:31 crc kubenswrapper[4809]: I1206 06:22:31.708013 4809 generic.go:334] "Generic (PLEG): container finished" podID="81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf" containerID="a4c656c4e48f2889f600c3e5ad29ed503a3315ff146bb9bf3f02bb05468d2fbb" exitCode=0 Dec 06 06:22:31 crc kubenswrapper[4809]: I1206 06:22:31.708075 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lkm6p" event={"ID":"81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf","Type":"ContainerDied","Data":"a4c656c4e48f2889f600c3e5ad29ed503a3315ff146bb9bf3f02bb05468d2fbb"} Dec 06 06:22:34 crc kubenswrapper[4809]: I1206 06:22:34.747249 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3d4b42d3-04a4-40e3-803b-11c77024f612","Type":"ContainerStarted","Data":"644e6da2ea964da6dc3103aea27cc874c7220fb372f243ee56009f3395b2d887"} Dec 06 06:22:34 crc kubenswrapper[4809]: I1206 06:22:34.750227 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lkm6p" event={"ID":"81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf","Type":"ContainerStarted","Data":"17be1db1527822a3f5d47f6d584bab2952bcd5836d7afc29db89604138a6ac96"} Dec 06 06:22:35 crc kubenswrapper[4809]: I1206 06:22:35.102675 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7d84b4d45c-j5rkj" Dec 06 06:22:35 crc kubenswrapper[4809]: I1206 06:22:35.127084 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lkm6p" podStartSLOduration=4.061705304 podStartE2EDuration="44.127046514s" podCreationTimestamp="2025-12-06 06:21:51 +0000 UTC" firstStartedPulling="2025-12-06 06:21:53.953361971 +0000 UTC m=+1838.842344913" lastFinishedPulling="2025-12-06 06:22:34.018703181 +0000 UTC m=+1878.907686123" observedRunningTime="2025-12-06 06:22:34.781750037 +0000 UTC m=+1879.670732979" watchObservedRunningTime="2025-12-06 06:22:35.127046514 +0000 UTC m=+1880.016029456" Dec 06 06:22:35 crc kubenswrapper[4809]: I1206 06:22:35.167125 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-2kbmw"] Dec 06 06:22:35 crc kubenswrapper[4809]: I1206 06:22:35.167358 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6b7bbf7cf9-2kbmw" podUID="624776a0-1d73-4785-9c78-45a4a30e2f80" containerName="dnsmasq-dns" containerID="cri-o://f03d7617420960dcd030c464e2ad37d2d398434079e59ba703a6bc97f5a36856" gracePeriod=10 Dec 06 06:22:35 crc kubenswrapper[4809]: I1206 06:22:35.317961 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6f6df4f56c-2frbl"] Dec 06 06:22:35 crc kubenswrapper[4809]: I1206 06:22:35.320455 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f6df4f56c-2frbl" Dec 06 06:22:35 crc kubenswrapper[4809]: I1206 06:22:35.344126 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f6df4f56c-2frbl"] Dec 06 06:22:35 crc kubenswrapper[4809]: I1206 06:22:35.428522 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2558b055-d8d9-4309-acb2-addf23e33dcc-dns-svc\") pod \"dnsmasq-dns-6f6df4f56c-2frbl\" (UID: \"2558b055-d8d9-4309-acb2-addf23e33dcc\") " pod="openstack/dnsmasq-dns-6f6df4f56c-2frbl" Dec 06 06:22:35 crc kubenswrapper[4809]: I1206 06:22:35.429036 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2558b055-d8d9-4309-acb2-addf23e33dcc-ovsdbserver-nb\") pod \"dnsmasq-dns-6f6df4f56c-2frbl\" (UID: \"2558b055-d8d9-4309-acb2-addf23e33dcc\") " pod="openstack/dnsmasq-dns-6f6df4f56c-2frbl" Dec 06 06:22:35 crc kubenswrapper[4809]: I1206 06:22:35.429162 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2558b055-d8d9-4309-acb2-addf23e33dcc-ovsdbserver-sb\") pod \"dnsmasq-dns-6f6df4f56c-2frbl\" (UID: \"2558b055-d8d9-4309-acb2-addf23e33dcc\") " pod="openstack/dnsmasq-dns-6f6df4f56c-2frbl" Dec 06 06:22:35 crc kubenswrapper[4809]: I1206 06:22:35.429304 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/2558b055-d8d9-4309-acb2-addf23e33dcc-openstack-edpm-ipam\") pod \"dnsmasq-dns-6f6df4f56c-2frbl\" (UID: \"2558b055-d8d9-4309-acb2-addf23e33dcc\") " pod="openstack/dnsmasq-dns-6f6df4f56c-2frbl" Dec 06 06:22:35 crc kubenswrapper[4809]: I1206 06:22:35.429471 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2558b055-d8d9-4309-acb2-addf23e33dcc-config\") pod \"dnsmasq-dns-6f6df4f56c-2frbl\" (UID: \"2558b055-d8d9-4309-acb2-addf23e33dcc\") " pod="openstack/dnsmasq-dns-6f6df4f56c-2frbl" Dec 06 06:22:35 crc kubenswrapper[4809]: I1206 06:22:35.429598 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbkvc\" (UniqueName: \"kubernetes.io/projected/2558b055-d8d9-4309-acb2-addf23e33dcc-kube-api-access-zbkvc\") pod \"dnsmasq-dns-6f6df4f56c-2frbl\" (UID: \"2558b055-d8d9-4309-acb2-addf23e33dcc\") " pod="openstack/dnsmasq-dns-6f6df4f56c-2frbl" Dec 06 06:22:35 crc kubenswrapper[4809]: I1206 06:22:35.429747 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2558b055-d8d9-4309-acb2-addf23e33dcc-dns-swift-storage-0\") pod \"dnsmasq-dns-6f6df4f56c-2frbl\" (UID: \"2558b055-d8d9-4309-acb2-addf23e33dcc\") " pod="openstack/dnsmasq-dns-6f6df4f56c-2frbl" Dec 06 06:22:35 crc kubenswrapper[4809]: I1206 06:22:35.532362 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2558b055-d8d9-4309-acb2-addf23e33dcc-ovsdbserver-sb\") pod \"dnsmasq-dns-6f6df4f56c-2frbl\" (UID: \"2558b055-d8d9-4309-acb2-addf23e33dcc\") " pod="openstack/dnsmasq-dns-6f6df4f56c-2frbl" Dec 06 06:22:35 crc kubenswrapper[4809]: I1206 06:22:35.532430 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/2558b055-d8d9-4309-acb2-addf23e33dcc-openstack-edpm-ipam\") pod \"dnsmasq-dns-6f6df4f56c-2frbl\" (UID: \"2558b055-d8d9-4309-acb2-addf23e33dcc\") " pod="openstack/dnsmasq-dns-6f6df4f56c-2frbl" Dec 06 06:22:35 crc kubenswrapper[4809]: I1206 06:22:35.533361 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/2558b055-d8d9-4309-acb2-addf23e33dcc-openstack-edpm-ipam\") pod \"dnsmasq-dns-6f6df4f56c-2frbl\" (UID: \"2558b055-d8d9-4309-acb2-addf23e33dcc\") " pod="openstack/dnsmasq-dns-6f6df4f56c-2frbl" Dec 06 06:22:35 crc kubenswrapper[4809]: I1206 06:22:35.533415 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2558b055-d8d9-4309-acb2-addf23e33dcc-config\") pod \"dnsmasq-dns-6f6df4f56c-2frbl\" (UID: \"2558b055-d8d9-4309-acb2-addf23e33dcc\") " pod="openstack/dnsmasq-dns-6f6df4f56c-2frbl" Dec 06 06:22:35 crc kubenswrapper[4809]: I1206 06:22:35.533524 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbkvc\" (UniqueName: \"kubernetes.io/projected/2558b055-d8d9-4309-acb2-addf23e33dcc-kube-api-access-zbkvc\") pod \"dnsmasq-dns-6f6df4f56c-2frbl\" (UID: \"2558b055-d8d9-4309-acb2-addf23e33dcc\") " pod="openstack/dnsmasq-dns-6f6df4f56c-2frbl" Dec 06 06:22:35 crc kubenswrapper[4809]: I1206 06:22:35.533921 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2558b055-d8d9-4309-acb2-addf23e33dcc-config\") pod \"dnsmasq-dns-6f6df4f56c-2frbl\" (UID: \"2558b055-d8d9-4309-acb2-addf23e33dcc\") " pod="openstack/dnsmasq-dns-6f6df4f56c-2frbl" Dec 06 06:22:35 crc kubenswrapper[4809]: I1206 06:22:35.534006 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2558b055-d8d9-4309-acb2-addf23e33dcc-ovsdbserver-sb\") pod \"dnsmasq-dns-6f6df4f56c-2frbl\" (UID: \"2558b055-d8d9-4309-acb2-addf23e33dcc\") " pod="openstack/dnsmasq-dns-6f6df4f56c-2frbl" Dec 06 06:22:35 crc kubenswrapper[4809]: I1206 06:22:35.534459 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2558b055-d8d9-4309-acb2-addf23e33dcc-dns-swift-storage-0\") pod \"dnsmasq-dns-6f6df4f56c-2frbl\" (UID: \"2558b055-d8d9-4309-acb2-addf23e33dcc\") " pod="openstack/dnsmasq-dns-6f6df4f56c-2frbl" Dec 06 06:22:35 crc kubenswrapper[4809]: I1206 06:22:35.534700 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2558b055-d8d9-4309-acb2-addf23e33dcc-dns-svc\") pod \"dnsmasq-dns-6f6df4f56c-2frbl\" (UID: \"2558b055-d8d9-4309-acb2-addf23e33dcc\") " pod="openstack/dnsmasq-dns-6f6df4f56c-2frbl" Dec 06 06:22:35 crc kubenswrapper[4809]: I1206 06:22:35.534923 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2558b055-d8d9-4309-acb2-addf23e33dcc-ovsdbserver-nb\") pod \"dnsmasq-dns-6f6df4f56c-2frbl\" (UID: \"2558b055-d8d9-4309-acb2-addf23e33dcc\") " pod="openstack/dnsmasq-dns-6f6df4f56c-2frbl" Dec 06 06:22:35 crc kubenswrapper[4809]: I1206 06:22:35.534964 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2558b055-d8d9-4309-acb2-addf23e33dcc-dns-swift-storage-0\") pod \"dnsmasq-dns-6f6df4f56c-2frbl\" (UID: \"2558b055-d8d9-4309-acb2-addf23e33dcc\") " pod="openstack/dnsmasq-dns-6f6df4f56c-2frbl" Dec 06 06:22:35 crc kubenswrapper[4809]: I1206 06:22:35.535821 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2558b055-d8d9-4309-acb2-addf23e33dcc-dns-svc\") pod \"dnsmasq-dns-6f6df4f56c-2frbl\" (UID: \"2558b055-d8d9-4309-acb2-addf23e33dcc\") " pod="openstack/dnsmasq-dns-6f6df4f56c-2frbl" Dec 06 06:22:35 crc kubenswrapper[4809]: I1206 06:22:35.536464 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2558b055-d8d9-4309-acb2-addf23e33dcc-ovsdbserver-nb\") pod \"dnsmasq-dns-6f6df4f56c-2frbl\" (UID: \"2558b055-d8d9-4309-acb2-addf23e33dcc\") " pod="openstack/dnsmasq-dns-6f6df4f56c-2frbl" Dec 06 06:22:35 crc kubenswrapper[4809]: I1206 06:22:35.577046 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbkvc\" (UniqueName: \"kubernetes.io/projected/2558b055-d8d9-4309-acb2-addf23e33dcc-kube-api-access-zbkvc\") pod \"dnsmasq-dns-6f6df4f56c-2frbl\" (UID: \"2558b055-d8d9-4309-acb2-addf23e33dcc\") " pod="openstack/dnsmasq-dns-6f6df4f56c-2frbl" Dec 06 06:22:35 crc kubenswrapper[4809]: I1206 06:22:35.704628 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f6df4f56c-2frbl" Dec 06 06:22:35 crc kubenswrapper[4809]: I1206 06:22:35.818948 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-ng4qq" event={"ID":"18fa982c-0572-4702-8c51-f0f7af7bf52f","Type":"ContainerStarted","Data":"d29e7e6ec1780e38eaf406f18e34fb3ad7982088153da6e08d1bbec4979270ac"} Dec 06 06:22:35 crc kubenswrapper[4809]: I1206 06:22:35.846908 4809 generic.go:334] "Generic (PLEG): container finished" podID="624776a0-1d73-4785-9c78-45a4a30e2f80" containerID="f03d7617420960dcd030c464e2ad37d2d398434079e59ba703a6bc97f5a36856" exitCode=0 Dec 06 06:22:35 crc kubenswrapper[4809]: I1206 06:22:35.846985 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-2kbmw" event={"ID":"624776a0-1d73-4785-9c78-45a4a30e2f80","Type":"ContainerDied","Data":"f03d7617420960dcd030c464e2ad37d2d398434079e59ba703a6bc97f5a36856"} Dec 06 06:22:35 crc kubenswrapper[4809]: I1206 06:22:35.862844 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3d4b42d3-04a4-40e3-803b-11c77024f612","Type":"ContainerStarted","Data":"10189cd18cf124194d1cd5f8103894926b7f3965e6baefbf9c437f773512728e"} Dec 06 06:22:35 crc kubenswrapper[4809]: I1206 06:22:35.864492 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-ng4qq" podStartSLOduration=2.87508627 podStartE2EDuration="49.864471388s" podCreationTimestamp="2025-12-06 06:21:46 +0000 UTC" firstStartedPulling="2025-12-06 06:21:47.580058372 +0000 UTC m=+1832.469041314" lastFinishedPulling="2025-12-06 06:22:34.56944349 +0000 UTC m=+1879.458426432" observedRunningTime="2025-12-06 06:22:35.863202905 +0000 UTC m=+1880.752185847" watchObservedRunningTime="2025-12-06 06:22:35.864471388 +0000 UTC m=+1880.753454330" Dec 06 06:22:35 crc kubenswrapper[4809]: I1206 06:22:35.945382 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-2kbmw" Dec 06 06:22:35 crc kubenswrapper[4809]: I1206 06:22:35.979203 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/624776a0-1d73-4785-9c78-45a4a30e2f80-dns-svc\") pod \"624776a0-1d73-4785-9c78-45a4a30e2f80\" (UID: \"624776a0-1d73-4785-9c78-45a4a30e2f80\") " Dec 06 06:22:35 crc kubenswrapper[4809]: I1206 06:22:35.979304 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/624776a0-1d73-4785-9c78-45a4a30e2f80-config\") pod \"624776a0-1d73-4785-9c78-45a4a30e2f80\" (UID: \"624776a0-1d73-4785-9c78-45a4a30e2f80\") " Dec 06 06:22:35 crc kubenswrapper[4809]: I1206 06:22:35.979336 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/624776a0-1d73-4785-9c78-45a4a30e2f80-dns-swift-storage-0\") pod \"624776a0-1d73-4785-9c78-45a4a30e2f80\" (UID: \"624776a0-1d73-4785-9c78-45a4a30e2f80\") " Dec 06 06:22:35 crc kubenswrapper[4809]: I1206 06:22:35.979511 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/624776a0-1d73-4785-9c78-45a4a30e2f80-ovsdbserver-nb\") pod \"624776a0-1d73-4785-9c78-45a4a30e2f80\" (UID: \"624776a0-1d73-4785-9c78-45a4a30e2f80\") " Dec 06 06:22:35 crc kubenswrapper[4809]: I1206 06:22:35.979542 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z6br6\" (UniqueName: \"kubernetes.io/projected/624776a0-1d73-4785-9c78-45a4a30e2f80-kube-api-access-z6br6\") pod \"624776a0-1d73-4785-9c78-45a4a30e2f80\" (UID: \"624776a0-1d73-4785-9c78-45a4a30e2f80\") " Dec 06 06:22:35 crc kubenswrapper[4809]: I1206 06:22:35.979572 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/624776a0-1d73-4785-9c78-45a4a30e2f80-ovsdbserver-sb\") pod \"624776a0-1d73-4785-9c78-45a4a30e2f80\" (UID: \"624776a0-1d73-4785-9c78-45a4a30e2f80\") " Dec 06 06:22:36 crc kubenswrapper[4809]: I1206 06:22:36.114173 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/624776a0-1d73-4785-9c78-45a4a30e2f80-kube-api-access-z6br6" (OuterVolumeSpecName: "kube-api-access-z6br6") pod "624776a0-1d73-4785-9c78-45a4a30e2f80" (UID: "624776a0-1d73-4785-9c78-45a4a30e2f80"). InnerVolumeSpecName "kube-api-access-z6br6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:22:36 crc kubenswrapper[4809]: I1206 06:22:36.156726 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/624776a0-1d73-4785-9c78-45a4a30e2f80-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "624776a0-1d73-4785-9c78-45a4a30e2f80" (UID: "624776a0-1d73-4785-9c78-45a4a30e2f80"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:22:36 crc kubenswrapper[4809]: I1206 06:22:36.198117 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z6br6\" (UniqueName: \"kubernetes.io/projected/624776a0-1d73-4785-9c78-45a4a30e2f80-kube-api-access-z6br6\") on node \"crc\" DevicePath \"\"" Dec 06 06:22:36 crc kubenswrapper[4809]: I1206 06:22:36.198150 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/624776a0-1d73-4785-9c78-45a4a30e2f80-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 06:22:36 crc kubenswrapper[4809]: I1206 06:22:36.214803 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/624776a0-1d73-4785-9c78-45a4a30e2f80-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "624776a0-1d73-4785-9c78-45a4a30e2f80" (UID: "624776a0-1d73-4785-9c78-45a4a30e2f80"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:22:36 crc kubenswrapper[4809]: I1206 06:22:36.273682 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/624776a0-1d73-4785-9c78-45a4a30e2f80-config" (OuterVolumeSpecName: "config") pod "624776a0-1d73-4785-9c78-45a4a30e2f80" (UID: "624776a0-1d73-4785-9c78-45a4a30e2f80"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:22:36 crc kubenswrapper[4809]: I1206 06:22:36.301579 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/624776a0-1d73-4785-9c78-45a4a30e2f80-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:22:36 crc kubenswrapper[4809]: I1206 06:22:36.301615 4809 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/624776a0-1d73-4785-9c78-45a4a30e2f80-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:22:36 crc kubenswrapper[4809]: I1206 06:22:36.386660 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/624776a0-1d73-4785-9c78-45a4a30e2f80-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "624776a0-1d73-4785-9c78-45a4a30e2f80" (UID: "624776a0-1d73-4785-9c78-45a4a30e2f80"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:22:36 crc kubenswrapper[4809]: I1206 06:22:36.403811 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/624776a0-1d73-4785-9c78-45a4a30e2f80-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 06:22:36 crc kubenswrapper[4809]: I1206 06:22:36.448695 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/624776a0-1d73-4785-9c78-45a4a30e2f80-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "624776a0-1d73-4785-9c78-45a4a30e2f80" (UID: "624776a0-1d73-4785-9c78-45a4a30e2f80"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:22:36 crc kubenswrapper[4809]: I1206 06:22:36.506278 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/624776a0-1d73-4785-9c78-45a4a30e2f80-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 06:22:36 crc kubenswrapper[4809]: I1206 06:22:36.565357 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f6df4f56c-2frbl"] Dec 06 06:22:36 crc kubenswrapper[4809]: I1206 06:22:36.877717 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3d4b42d3-04a4-40e3-803b-11c77024f612","Type":"ContainerStarted","Data":"493dde793f5dce6b38eb58b439b1d65c77ff6c4a3277b1a17ad6b0e5a2f1c252"} Dec 06 06:22:36 crc kubenswrapper[4809]: I1206 06:22:36.881124 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f6df4f56c-2frbl" event={"ID":"2558b055-d8d9-4309-acb2-addf23e33dcc","Type":"ContainerStarted","Data":"e7a5efb08d89d66931135fdd22cac9494df4930c67ec30953fee5cf53cc7a829"} Dec 06 06:22:36 crc kubenswrapper[4809]: I1206 06:22:36.885502 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-2kbmw" event={"ID":"624776a0-1d73-4785-9c78-45a4a30e2f80","Type":"ContainerDied","Data":"1859763d34b18ac613b6a7721b806521d6b0f3c3482f7c9318562059bb68864f"} Dec 06 06:22:36 crc kubenswrapper[4809]: I1206 06:22:36.885565 4809 scope.go:117] "RemoveContainer" containerID="f03d7617420960dcd030c464e2ad37d2d398434079e59ba703a6bc97f5a36856" Dec 06 06:22:36 crc kubenswrapper[4809]: I1206 06:22:36.885593 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-2kbmw" Dec 06 06:22:36 crc kubenswrapper[4809]: I1206 06:22:36.928221 4809 scope.go:117] "RemoveContainer" containerID="d29627bf8956afc444882ed97f05eb1f9679336e8432e00c64d3c772a8acd554" Dec 06 06:22:36 crc kubenswrapper[4809]: I1206 06:22:36.942600 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-2kbmw"] Dec 06 06:22:36 crc kubenswrapper[4809]: I1206 06:22:36.990314 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-2kbmw"] Dec 06 06:22:37 crc kubenswrapper[4809]: I1206 06:22:37.407539 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="624776a0-1d73-4785-9c78-45a4a30e2f80" path="/var/lib/kubelet/pods/624776a0-1d73-4785-9c78-45a4a30e2f80/volumes" Dec 06 06:22:37 crc kubenswrapper[4809]: I1206 06:22:37.912276 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f6df4f56c-2frbl" event={"ID":"2558b055-d8d9-4309-acb2-addf23e33dcc","Type":"ContainerDied","Data":"ef21e54f6e33d757af53b11b15e0bf97155ce470289478a21af213538222b410"} Dec 06 06:22:37 crc kubenswrapper[4809]: I1206 06:22:37.912049 4809 generic.go:334] "Generic (PLEG): container finished" podID="2558b055-d8d9-4309-acb2-addf23e33dcc" containerID="ef21e54f6e33d757af53b11b15e0bf97155ce470289478a21af213538222b410" exitCode=0 Dec 06 06:22:38 crc kubenswrapper[4809]: I1206 06:22:38.937316 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f6df4f56c-2frbl" event={"ID":"2558b055-d8d9-4309-acb2-addf23e33dcc","Type":"ContainerStarted","Data":"50c7d48c12c7c088186b55b94e4e6f79bea62854c6e8b859368a858f2c507289"} Dec 06 06:22:38 crc kubenswrapper[4809]: I1206 06:22:38.937702 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6f6df4f56c-2frbl" Dec 06 06:22:38 crc kubenswrapper[4809]: I1206 06:22:38.941230 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3d4b42d3-04a4-40e3-803b-11c77024f612","Type":"ContainerStarted","Data":"f7069d7aabb93d228c5974d5a56a8727514829ecffaab6e9caf693226d4fdde2"} Dec 06 06:22:38 crc kubenswrapper[4809]: I1206 06:22:38.941444 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 06:22:38 crc kubenswrapper[4809]: I1206 06:22:38.963346 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6f6df4f56c-2frbl" podStartSLOduration=3.963326265 podStartE2EDuration="3.963326265s" podCreationTimestamp="2025-12-06 06:22:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:22:38.957684003 +0000 UTC m=+1883.846666945" watchObservedRunningTime="2025-12-06 06:22:38.963326265 +0000 UTC m=+1883.852309207" Dec 06 06:22:38 crc kubenswrapper[4809]: I1206 06:22:38.992193 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=15.70100215 podStartE2EDuration="31.992169293s" podCreationTimestamp="2025-12-06 06:22:07 +0000 UTC" firstStartedPulling="2025-12-06 06:22:21.567266607 +0000 UTC m=+1866.456249549" lastFinishedPulling="2025-12-06 06:22:37.85843375 +0000 UTC m=+1882.747416692" observedRunningTime="2025-12-06 06:22:38.986252844 +0000 UTC m=+1883.875235796" watchObservedRunningTime="2025-12-06 06:22:38.992169293 +0000 UTC m=+1883.881152235" Dec 06 06:22:39 crc kubenswrapper[4809]: I1206 06:22:39.952469 4809 generic.go:334] "Generic (PLEG): container finished" podID="18fa982c-0572-4702-8c51-f0f7af7bf52f" containerID="d29e7e6ec1780e38eaf406f18e34fb3ad7982088153da6e08d1bbec4979270ac" exitCode=0 Dec 06 06:22:39 crc kubenswrapper[4809]: I1206 06:22:39.952651 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-ng4qq" event={"ID":"18fa982c-0572-4702-8c51-f0f7af7bf52f","Type":"ContainerDied","Data":"d29e7e6ec1780e38eaf406f18e34fb3ad7982088153da6e08d1bbec4979270ac"} Dec 06 06:22:41 crc kubenswrapper[4809]: I1206 06:22:41.365248 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-ng4qq" Dec 06 06:22:41 crc kubenswrapper[4809]: I1206 06:22:41.549772 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18fa982c-0572-4702-8c51-f0f7af7bf52f-combined-ca-bundle\") pod \"18fa982c-0572-4702-8c51-f0f7af7bf52f\" (UID: \"18fa982c-0572-4702-8c51-f0f7af7bf52f\") " Dec 06 06:22:41 crc kubenswrapper[4809]: I1206 06:22:41.550040 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s5vwn\" (UniqueName: \"kubernetes.io/projected/18fa982c-0572-4702-8c51-f0f7af7bf52f-kube-api-access-s5vwn\") pod \"18fa982c-0572-4702-8c51-f0f7af7bf52f\" (UID: \"18fa982c-0572-4702-8c51-f0f7af7bf52f\") " Dec 06 06:22:41 crc kubenswrapper[4809]: I1206 06:22:41.550083 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18fa982c-0572-4702-8c51-f0f7af7bf52f-config-data\") pod \"18fa982c-0572-4702-8c51-f0f7af7bf52f\" (UID: \"18fa982c-0572-4702-8c51-f0f7af7bf52f\") " Dec 06 06:22:41 crc kubenswrapper[4809]: I1206 06:22:41.561130 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18fa982c-0572-4702-8c51-f0f7af7bf52f-kube-api-access-s5vwn" (OuterVolumeSpecName: "kube-api-access-s5vwn") pod "18fa982c-0572-4702-8c51-f0f7af7bf52f" (UID: "18fa982c-0572-4702-8c51-f0f7af7bf52f"). InnerVolumeSpecName "kube-api-access-s5vwn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:22:41 crc kubenswrapper[4809]: I1206 06:22:41.585205 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18fa982c-0572-4702-8c51-f0f7af7bf52f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "18fa982c-0572-4702-8c51-f0f7af7bf52f" (UID: "18fa982c-0572-4702-8c51-f0f7af7bf52f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:22:41 crc kubenswrapper[4809]: I1206 06:22:41.645397 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18fa982c-0572-4702-8c51-f0f7af7bf52f-config-data" (OuterVolumeSpecName: "config-data") pod "18fa982c-0572-4702-8c51-f0f7af7bf52f" (UID: "18fa982c-0572-4702-8c51-f0f7af7bf52f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:22:41 crc kubenswrapper[4809]: I1206 06:22:41.652898 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s5vwn\" (UniqueName: \"kubernetes.io/projected/18fa982c-0572-4702-8c51-f0f7af7bf52f-kube-api-access-s5vwn\") on node \"crc\" DevicePath \"\"" Dec 06 06:22:41 crc kubenswrapper[4809]: I1206 06:22:41.653752 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18fa982c-0572-4702-8c51-f0f7af7bf52f-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:22:41 crc kubenswrapper[4809]: I1206 06:22:41.653774 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18fa982c-0572-4702-8c51-f0f7af7bf52f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:22:41 crc kubenswrapper[4809]: I1206 06:22:41.853866 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lkm6p" Dec 06 06:22:41 crc kubenswrapper[4809]: I1206 06:22:41.853942 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lkm6p" Dec 06 06:22:41 crc kubenswrapper[4809]: I1206 06:22:41.974601 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-ng4qq" event={"ID":"18fa982c-0572-4702-8c51-f0f7af7bf52f","Type":"ContainerDied","Data":"a2c904fd7e9d2a74328c28fce664dc5fb7678da34d29f5b08c18e40cfe4bdcc4"} Dec 06 06:22:41 crc kubenswrapper[4809]: I1206 06:22:41.974640 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a2c904fd7e9d2a74328c28fce664dc5fb7678da34d29f5b08c18e40cfe4bdcc4" Dec 06 06:22:41 crc kubenswrapper[4809]: I1206 06:22:41.974658 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-ng4qq" Dec 06 06:22:43 crc kubenswrapper[4809]: I1206 06:22:43.061211 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-7fcd5d8964-kw6sz"] Dec 06 06:22:43 crc kubenswrapper[4809]: E1206 06:22:43.062269 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="624776a0-1d73-4785-9c78-45a4a30e2f80" containerName="init" Dec 06 06:22:43 crc kubenswrapper[4809]: I1206 06:22:43.062289 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="624776a0-1d73-4785-9c78-45a4a30e2f80" containerName="init" Dec 06 06:22:43 crc kubenswrapper[4809]: E1206 06:22:43.062335 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="624776a0-1d73-4785-9c78-45a4a30e2f80" containerName="dnsmasq-dns" Dec 06 06:22:43 crc kubenswrapper[4809]: I1206 06:22:43.062344 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="624776a0-1d73-4785-9c78-45a4a30e2f80" containerName="dnsmasq-dns" Dec 06 06:22:43 crc kubenswrapper[4809]: E1206 06:22:43.062359 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18fa982c-0572-4702-8c51-f0f7af7bf52f" containerName="heat-db-sync" Dec 06 06:22:43 crc kubenswrapper[4809]: I1206 06:22:43.062369 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="18fa982c-0572-4702-8c51-f0f7af7bf52f" containerName="heat-db-sync" Dec 06 06:22:43 crc kubenswrapper[4809]: I1206 06:22:43.062665 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="18fa982c-0572-4702-8c51-f0f7af7bf52f" containerName="heat-db-sync" Dec 06 06:22:43 crc kubenswrapper[4809]: I1206 06:22:43.062700 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="624776a0-1d73-4785-9c78-45a4a30e2f80" containerName="dnsmasq-dns" Dec 06 06:22:43 crc kubenswrapper[4809]: I1206 06:22:43.063852 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7fcd5d8964-kw6sz" Dec 06 06:22:43 crc kubenswrapper[4809]: I1206 06:22:43.082457 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-7fcd5d8964-kw6sz"] Dec 06 06:22:43 crc kubenswrapper[4809]: I1206 06:22:43.158874 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-794dfc48dc-pkjpx"] Dec 06 06:22:43 crc kubenswrapper[4809]: I1206 06:22:43.160608 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-794dfc48dc-pkjpx" Dec 06 06:22:43 crc kubenswrapper[4809]: I1206 06:22:43.181837 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-794dfc48dc-pkjpx"] Dec 06 06:22:43 crc kubenswrapper[4809]: I1206 06:22:43.188614 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qg4c8\" (UniqueName: \"kubernetes.io/projected/a71b0d8f-059d-4801-a2e4-6dc8ec44c78a-kube-api-access-qg4c8\") pod \"heat-engine-7fcd5d8964-kw6sz\" (UID: \"a71b0d8f-059d-4801-a2e4-6dc8ec44c78a\") " pod="openstack/heat-engine-7fcd5d8964-kw6sz" Dec 06 06:22:43 crc kubenswrapper[4809]: I1206 06:22:43.188723 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a71b0d8f-059d-4801-a2e4-6dc8ec44c78a-config-data-custom\") pod \"heat-engine-7fcd5d8964-kw6sz\" (UID: \"a71b0d8f-059d-4801-a2e4-6dc8ec44c78a\") " pod="openstack/heat-engine-7fcd5d8964-kw6sz" Dec 06 06:22:43 crc kubenswrapper[4809]: I1206 06:22:43.188964 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a71b0d8f-059d-4801-a2e4-6dc8ec44c78a-combined-ca-bundle\") pod \"heat-engine-7fcd5d8964-kw6sz\" (UID: \"a71b0d8f-059d-4801-a2e4-6dc8ec44c78a\") " pod="openstack/heat-engine-7fcd5d8964-kw6sz" Dec 06 06:22:43 crc kubenswrapper[4809]: I1206 06:22:43.189067 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a71b0d8f-059d-4801-a2e4-6dc8ec44c78a-config-data\") pod \"heat-engine-7fcd5d8964-kw6sz\" (UID: \"a71b0d8f-059d-4801-a2e4-6dc8ec44c78a\") " pod="openstack/heat-engine-7fcd5d8964-kw6sz" Dec 06 06:22:43 crc kubenswrapper[4809]: I1206 06:22:43.290737 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b09a69de-0d5f-44f7-a3c9-2e6902da17b8-combined-ca-bundle\") pod \"heat-cfnapi-794dfc48dc-pkjpx\" (UID: \"b09a69de-0d5f-44f7-a3c9-2e6902da17b8\") " pod="openstack/heat-cfnapi-794dfc48dc-pkjpx" Dec 06 06:22:43 crc kubenswrapper[4809]: I1206 06:22:43.290807 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a71b0d8f-059d-4801-a2e4-6dc8ec44c78a-config-data-custom\") pod \"heat-engine-7fcd5d8964-kw6sz\" (UID: \"a71b0d8f-059d-4801-a2e4-6dc8ec44c78a\") " pod="openstack/heat-engine-7fcd5d8964-kw6sz" Dec 06 06:22:43 crc kubenswrapper[4809]: I1206 06:22:43.290878 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b09a69de-0d5f-44f7-a3c9-2e6902da17b8-config-data-custom\") pod \"heat-cfnapi-794dfc48dc-pkjpx\" (UID: \"b09a69de-0d5f-44f7-a3c9-2e6902da17b8\") " pod="openstack/heat-cfnapi-794dfc48dc-pkjpx" Dec 06 06:22:43 crc kubenswrapper[4809]: I1206 06:22:43.290958 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b09a69de-0d5f-44f7-a3c9-2e6902da17b8-internal-tls-certs\") pod \"heat-cfnapi-794dfc48dc-pkjpx\" (UID: \"b09a69de-0d5f-44f7-a3c9-2e6902da17b8\") " pod="openstack/heat-cfnapi-794dfc48dc-pkjpx" Dec 06 06:22:43 crc kubenswrapper[4809]: I1206 06:22:43.290996 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdcfx\" (UniqueName: \"kubernetes.io/projected/b09a69de-0d5f-44f7-a3c9-2e6902da17b8-kube-api-access-tdcfx\") pod \"heat-cfnapi-794dfc48dc-pkjpx\" (UID: \"b09a69de-0d5f-44f7-a3c9-2e6902da17b8\") " pod="openstack/heat-cfnapi-794dfc48dc-pkjpx" Dec 06 06:22:43 crc kubenswrapper[4809]: I1206 06:22:43.291046 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a71b0d8f-059d-4801-a2e4-6dc8ec44c78a-combined-ca-bundle\") pod \"heat-engine-7fcd5d8964-kw6sz\" (UID: \"a71b0d8f-059d-4801-a2e4-6dc8ec44c78a\") " pod="openstack/heat-engine-7fcd5d8964-kw6sz" Dec 06 06:22:43 crc kubenswrapper[4809]: I1206 06:22:43.291067 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b09a69de-0d5f-44f7-a3c9-2e6902da17b8-config-data\") pod \"heat-cfnapi-794dfc48dc-pkjpx\" (UID: \"b09a69de-0d5f-44f7-a3c9-2e6902da17b8\") " pod="openstack/heat-cfnapi-794dfc48dc-pkjpx" Dec 06 06:22:43 crc kubenswrapper[4809]: I1206 06:22:43.291233 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a71b0d8f-059d-4801-a2e4-6dc8ec44c78a-config-data\") pod \"heat-engine-7fcd5d8964-kw6sz\" (UID: \"a71b0d8f-059d-4801-a2e4-6dc8ec44c78a\") " pod="openstack/heat-engine-7fcd5d8964-kw6sz" Dec 06 06:22:43 crc kubenswrapper[4809]: I1206 06:22:43.291326 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b09a69de-0d5f-44f7-a3c9-2e6902da17b8-public-tls-certs\") pod \"heat-cfnapi-794dfc48dc-pkjpx\" (UID: \"b09a69de-0d5f-44f7-a3c9-2e6902da17b8\") " pod="openstack/heat-cfnapi-794dfc48dc-pkjpx" Dec 06 06:22:43 crc kubenswrapper[4809]: I1206 06:22:43.291534 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qg4c8\" (UniqueName: \"kubernetes.io/projected/a71b0d8f-059d-4801-a2e4-6dc8ec44c78a-kube-api-access-qg4c8\") pod \"heat-engine-7fcd5d8964-kw6sz\" (UID: \"a71b0d8f-059d-4801-a2e4-6dc8ec44c78a\") " pod="openstack/heat-engine-7fcd5d8964-kw6sz" Dec 06 06:22:43 crc kubenswrapper[4809]: I1206 06:22:43.301037 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a71b0d8f-059d-4801-a2e4-6dc8ec44c78a-config-data\") pod \"heat-engine-7fcd5d8964-kw6sz\" (UID: \"a71b0d8f-059d-4801-a2e4-6dc8ec44c78a\") " pod="openstack/heat-engine-7fcd5d8964-kw6sz" Dec 06 06:22:43 crc kubenswrapper[4809]: I1206 06:22:43.301552 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a71b0d8f-059d-4801-a2e4-6dc8ec44c78a-config-data-custom\") pod \"heat-engine-7fcd5d8964-kw6sz\" (UID: \"a71b0d8f-059d-4801-a2e4-6dc8ec44c78a\") " pod="openstack/heat-engine-7fcd5d8964-kw6sz" Dec 06 06:22:43 crc kubenswrapper[4809]: I1206 06:22:43.301714 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a71b0d8f-059d-4801-a2e4-6dc8ec44c78a-combined-ca-bundle\") pod \"heat-engine-7fcd5d8964-kw6sz\" (UID: \"a71b0d8f-059d-4801-a2e4-6dc8ec44c78a\") " pod="openstack/heat-engine-7fcd5d8964-kw6sz" Dec 06 06:22:43 crc kubenswrapper[4809]: I1206 06:22:43.313698 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qg4c8\" (UniqueName: \"kubernetes.io/projected/a71b0d8f-059d-4801-a2e4-6dc8ec44c78a-kube-api-access-qg4c8\") pod \"heat-engine-7fcd5d8964-kw6sz\" (UID: \"a71b0d8f-059d-4801-a2e4-6dc8ec44c78a\") " pod="openstack/heat-engine-7fcd5d8964-kw6sz" Dec 06 06:22:43 crc kubenswrapper[4809]: I1206 06:22:43.335173 4809 patch_prober.go:28] interesting pod/logging-loki-gateway-65498c4f8f-48rjq container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.53:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 06:22:43 crc kubenswrapper[4809]: I1206 06:22:43.335238 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-65498c4f8f-48rjq" podUID="72a3ed95-3e3d-4faf-88e4-ad1731902910" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.53:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 06:22:43 crc kubenswrapper[4809]: I1206 06:22:43.389567 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7fcd5d8964-kw6sz" Dec 06 06:22:43 crc kubenswrapper[4809]: I1206 06:22:43.390548 4809 scope.go:117] "RemoveContainer" containerID="9bc2dc88b3b9f309ce0169aa510326dd77022dece4c69cab3141df00fe5c0106" Dec 06 06:22:43 crc kubenswrapper[4809]: E1206 06:22:43.390783 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:22:43 crc kubenswrapper[4809]: I1206 06:22:43.393192 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b09a69de-0d5f-44f7-a3c9-2e6902da17b8-internal-tls-certs\") pod \"heat-cfnapi-794dfc48dc-pkjpx\" (UID: \"b09a69de-0d5f-44f7-a3c9-2e6902da17b8\") " pod="openstack/heat-cfnapi-794dfc48dc-pkjpx" Dec 06 06:22:43 crc kubenswrapper[4809]: I1206 06:22:43.393240 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdcfx\" (UniqueName: \"kubernetes.io/projected/b09a69de-0d5f-44f7-a3c9-2e6902da17b8-kube-api-access-tdcfx\") pod \"heat-cfnapi-794dfc48dc-pkjpx\" (UID: \"b09a69de-0d5f-44f7-a3c9-2e6902da17b8\") " pod="openstack/heat-cfnapi-794dfc48dc-pkjpx" Dec 06 06:22:43 crc kubenswrapper[4809]: I1206 06:22:43.393266 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b09a69de-0d5f-44f7-a3c9-2e6902da17b8-config-data\") pod \"heat-cfnapi-794dfc48dc-pkjpx\" (UID: \"b09a69de-0d5f-44f7-a3c9-2e6902da17b8\") " pod="openstack/heat-cfnapi-794dfc48dc-pkjpx" Dec 06 06:22:43 crc kubenswrapper[4809]: I1206 06:22:43.393297 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b09a69de-0d5f-44f7-a3c9-2e6902da17b8-public-tls-certs\") pod \"heat-cfnapi-794dfc48dc-pkjpx\" (UID: \"b09a69de-0d5f-44f7-a3c9-2e6902da17b8\") " pod="openstack/heat-cfnapi-794dfc48dc-pkjpx" Dec 06 06:22:43 crc kubenswrapper[4809]: I1206 06:22:43.393381 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b09a69de-0d5f-44f7-a3c9-2e6902da17b8-combined-ca-bundle\") pod \"heat-cfnapi-794dfc48dc-pkjpx\" (UID: \"b09a69de-0d5f-44f7-a3c9-2e6902da17b8\") " pod="openstack/heat-cfnapi-794dfc48dc-pkjpx" Dec 06 06:22:43 crc kubenswrapper[4809]: I1206 06:22:43.393459 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b09a69de-0d5f-44f7-a3c9-2e6902da17b8-config-data-custom\") pod \"heat-cfnapi-794dfc48dc-pkjpx\" (UID: \"b09a69de-0d5f-44f7-a3c9-2e6902da17b8\") " pod="openstack/heat-cfnapi-794dfc48dc-pkjpx" Dec 06 06:22:43 crc kubenswrapper[4809]: I1206 06:22:43.397231 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b09a69de-0d5f-44f7-a3c9-2e6902da17b8-config-data-custom\") pod \"heat-cfnapi-794dfc48dc-pkjpx\" (UID: \"b09a69de-0d5f-44f7-a3c9-2e6902da17b8\") " pod="openstack/heat-cfnapi-794dfc48dc-pkjpx" Dec 06 06:22:43 crc kubenswrapper[4809]: I1206 06:22:43.399087 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b09a69de-0d5f-44f7-a3c9-2e6902da17b8-combined-ca-bundle\") pod \"heat-cfnapi-794dfc48dc-pkjpx\" (UID: \"b09a69de-0d5f-44f7-a3c9-2e6902da17b8\") " pod="openstack/heat-cfnapi-794dfc48dc-pkjpx" Dec 06 06:22:43 crc kubenswrapper[4809]: I1206 06:22:43.399997 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b09a69de-0d5f-44f7-a3c9-2e6902da17b8-config-data\") pod \"heat-cfnapi-794dfc48dc-pkjpx\" (UID: \"b09a69de-0d5f-44f7-a3c9-2e6902da17b8\") " pod="openstack/heat-cfnapi-794dfc48dc-pkjpx" Dec 06 06:22:43 crc kubenswrapper[4809]: I1206 06:22:43.400670 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b09a69de-0d5f-44f7-a3c9-2e6902da17b8-internal-tls-certs\") pod \"heat-cfnapi-794dfc48dc-pkjpx\" (UID: \"b09a69de-0d5f-44f7-a3c9-2e6902da17b8\") " pod="openstack/heat-cfnapi-794dfc48dc-pkjpx" Dec 06 06:22:43 crc kubenswrapper[4809]: I1206 06:22:43.401983 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b09a69de-0d5f-44f7-a3c9-2e6902da17b8-public-tls-certs\") pod \"heat-cfnapi-794dfc48dc-pkjpx\" (UID: \"b09a69de-0d5f-44f7-a3c9-2e6902da17b8\") " pod="openstack/heat-cfnapi-794dfc48dc-pkjpx" Dec 06 06:22:43 crc kubenswrapper[4809]: I1206 06:22:43.414809 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdcfx\" (UniqueName: \"kubernetes.io/projected/b09a69de-0d5f-44f7-a3c9-2e6902da17b8-kube-api-access-tdcfx\") pod \"heat-cfnapi-794dfc48dc-pkjpx\" (UID: \"b09a69de-0d5f-44f7-a3c9-2e6902da17b8\") " pod="openstack/heat-cfnapi-794dfc48dc-pkjpx" Dec 06 06:22:43 crc kubenswrapper[4809]: I1206 06:22:43.480507 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-794dfc48dc-pkjpx" Dec 06 06:22:44 crc kubenswrapper[4809]: I1206 06:22:44.223379 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-7849fbcbcb-msd4w"] Dec 06 06:22:44 crc kubenswrapper[4809]: I1206 06:22:44.225674 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-7849fbcbcb-msd4w" Dec 06 06:22:44 crc kubenswrapper[4809]: I1206 06:22:44.273081 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-7849fbcbcb-msd4w"] Dec 06 06:22:44 crc kubenswrapper[4809]: I1206 06:22:44.428693 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dccd4eeb-4ba8-4200-8366-28d4ead2e476-public-tls-certs\") pod \"heat-api-7849fbcbcb-msd4w\" (UID: \"dccd4eeb-4ba8-4200-8366-28d4ead2e476\") " pod="openstack/heat-api-7849fbcbcb-msd4w" Dec 06 06:22:44 crc kubenswrapper[4809]: I1206 06:22:44.451841 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dccd4eeb-4ba8-4200-8366-28d4ead2e476-combined-ca-bundle\") pod \"heat-api-7849fbcbcb-msd4w\" (UID: \"dccd4eeb-4ba8-4200-8366-28d4ead2e476\") " pod="openstack/heat-api-7849fbcbcb-msd4w" Dec 06 06:22:44 crc kubenswrapper[4809]: I1206 06:22:44.451902 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dccd4eeb-4ba8-4200-8366-28d4ead2e476-config-data-custom\") pod \"heat-api-7849fbcbcb-msd4w\" (UID: \"dccd4eeb-4ba8-4200-8366-28d4ead2e476\") " pod="openstack/heat-api-7849fbcbcb-msd4w" Dec 06 06:22:44 crc kubenswrapper[4809]: I1206 06:22:44.451972 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dccd4eeb-4ba8-4200-8366-28d4ead2e476-internal-tls-certs\") pod \"heat-api-7849fbcbcb-msd4w\" (UID: \"dccd4eeb-4ba8-4200-8366-28d4ead2e476\") " pod="openstack/heat-api-7849fbcbcb-msd4w" Dec 06 06:22:44 crc kubenswrapper[4809]: I1206 06:22:44.451999 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dccd4eeb-4ba8-4200-8366-28d4ead2e476-config-data\") pod \"heat-api-7849fbcbcb-msd4w\" (UID: \"dccd4eeb-4ba8-4200-8366-28d4ead2e476\") " pod="openstack/heat-api-7849fbcbcb-msd4w" Dec 06 06:22:44 crc kubenswrapper[4809]: I1206 06:22:44.452042 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v26gb\" (UniqueName: \"kubernetes.io/projected/dccd4eeb-4ba8-4200-8366-28d4ead2e476-kube-api-access-v26gb\") pod \"heat-api-7849fbcbcb-msd4w\" (UID: \"dccd4eeb-4ba8-4200-8366-28d4ead2e476\") " pod="openstack/heat-api-7849fbcbcb-msd4w" Dec 06 06:22:44 crc kubenswrapper[4809]: I1206 06:22:44.554894 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dccd4eeb-4ba8-4200-8366-28d4ead2e476-combined-ca-bundle\") pod \"heat-api-7849fbcbcb-msd4w\" (UID: \"dccd4eeb-4ba8-4200-8366-28d4ead2e476\") " pod="openstack/heat-api-7849fbcbcb-msd4w" Dec 06 06:22:44 crc kubenswrapper[4809]: I1206 06:22:44.554958 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dccd4eeb-4ba8-4200-8366-28d4ead2e476-config-data-custom\") pod \"heat-api-7849fbcbcb-msd4w\" (UID: \"dccd4eeb-4ba8-4200-8366-28d4ead2e476\") " pod="openstack/heat-api-7849fbcbcb-msd4w" Dec 06 06:22:44 crc kubenswrapper[4809]: I1206 06:22:44.563340 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dccd4eeb-4ba8-4200-8366-28d4ead2e476-internal-tls-certs\") pod \"heat-api-7849fbcbcb-msd4w\" (UID: \"dccd4eeb-4ba8-4200-8366-28d4ead2e476\") " pod="openstack/heat-api-7849fbcbcb-msd4w" Dec 06 06:22:44 crc kubenswrapper[4809]: I1206 06:22:44.563377 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dccd4eeb-4ba8-4200-8366-28d4ead2e476-config-data\") pod \"heat-api-7849fbcbcb-msd4w\" (UID: \"dccd4eeb-4ba8-4200-8366-28d4ead2e476\") " pod="openstack/heat-api-7849fbcbcb-msd4w" Dec 06 06:22:44 crc kubenswrapper[4809]: I1206 06:22:44.572778 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v26gb\" (UniqueName: \"kubernetes.io/projected/dccd4eeb-4ba8-4200-8366-28d4ead2e476-kube-api-access-v26gb\") pod \"heat-api-7849fbcbcb-msd4w\" (UID: \"dccd4eeb-4ba8-4200-8366-28d4ead2e476\") " pod="openstack/heat-api-7849fbcbcb-msd4w" Dec 06 06:22:44 crc kubenswrapper[4809]: I1206 06:22:44.573346 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dccd4eeb-4ba8-4200-8366-28d4ead2e476-public-tls-certs\") pod \"heat-api-7849fbcbcb-msd4w\" (UID: \"dccd4eeb-4ba8-4200-8366-28d4ead2e476\") " pod="openstack/heat-api-7849fbcbcb-msd4w" Dec 06 06:22:44 crc kubenswrapper[4809]: I1206 06:22:44.621481 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dccd4eeb-4ba8-4200-8366-28d4ead2e476-config-data\") pod \"heat-api-7849fbcbcb-msd4w\" (UID: \"dccd4eeb-4ba8-4200-8366-28d4ead2e476\") " pod="openstack/heat-api-7849fbcbcb-msd4w" Dec 06 06:22:44 crc kubenswrapper[4809]: I1206 06:22:44.622607 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dccd4eeb-4ba8-4200-8366-28d4ead2e476-public-tls-certs\") pod \"heat-api-7849fbcbcb-msd4w\" (UID: \"dccd4eeb-4ba8-4200-8366-28d4ead2e476\") " pod="openstack/heat-api-7849fbcbcb-msd4w" Dec 06 06:22:44 crc kubenswrapper[4809]: I1206 06:22:44.623604 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dccd4eeb-4ba8-4200-8366-28d4ead2e476-combined-ca-bundle\") pod \"heat-api-7849fbcbcb-msd4w\" (UID: \"dccd4eeb-4ba8-4200-8366-28d4ead2e476\") " pod="openstack/heat-api-7849fbcbcb-msd4w" Dec 06 06:22:44 crc kubenswrapper[4809]: I1206 06:22:44.637833 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v26gb\" (UniqueName: \"kubernetes.io/projected/dccd4eeb-4ba8-4200-8366-28d4ead2e476-kube-api-access-v26gb\") pod \"heat-api-7849fbcbcb-msd4w\" (UID: \"dccd4eeb-4ba8-4200-8366-28d4ead2e476\") " pod="openstack/heat-api-7849fbcbcb-msd4w" Dec 06 06:22:44 crc kubenswrapper[4809]: I1206 06:22:44.641301 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dccd4eeb-4ba8-4200-8366-28d4ead2e476-config-data-custom\") pod \"heat-api-7849fbcbcb-msd4w\" (UID: \"dccd4eeb-4ba8-4200-8366-28d4ead2e476\") " pod="openstack/heat-api-7849fbcbcb-msd4w" Dec 06 06:22:44 crc kubenswrapper[4809]: I1206 06:22:44.644859 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dccd4eeb-4ba8-4200-8366-28d4ead2e476-internal-tls-certs\") pod \"heat-api-7849fbcbcb-msd4w\" (UID: \"dccd4eeb-4ba8-4200-8366-28d4ead2e476\") " pod="openstack/heat-api-7849fbcbcb-msd4w" Dec 06 06:22:44 crc kubenswrapper[4809]: I1206 06:22:44.839533 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-7849fbcbcb-msd4w" Dec 06 06:22:45 crc kubenswrapper[4809]: I1206 06:22:45.088397 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-7fcd5d8964-kw6sz"] Dec 06 06:22:45 crc kubenswrapper[4809]: W1206 06:22:45.096461 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda71b0d8f_059d_4801_a2e4_6dc8ec44c78a.slice/crio-f6a84853672d7d8584cde18cdb0aff480e869884766f518fbe94603f1ba07c0e WatchSource:0}: Error finding container f6a84853672d7d8584cde18cdb0aff480e869884766f518fbe94603f1ba07c0e: Status 404 returned error can't find the container with id f6a84853672d7d8584cde18cdb0aff480e869884766f518fbe94603f1ba07c0e Dec 06 06:22:45 crc kubenswrapper[4809]: I1206 06:22:45.101487 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-794dfc48dc-pkjpx"] Dec 06 06:22:45 crc kubenswrapper[4809]: I1206 06:22:45.279607 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lkm6p" podUID="81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf" containerName="registry-server" probeResult="failure" output=< Dec 06 06:22:45 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Dec 06 06:22:45 crc kubenswrapper[4809]: > Dec 06 06:22:45 crc kubenswrapper[4809]: I1206 06:22:45.350967 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-7849fbcbcb-msd4w"] Dec 06 06:22:45 crc kubenswrapper[4809]: I1206 06:22:45.706894 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6f6df4f56c-2frbl" Dec 06 06:22:45 crc kubenswrapper[4809]: I1206 06:22:45.807028 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-j5rkj"] Dec 06 06:22:45 crc kubenswrapper[4809]: I1206 06:22:45.807316 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7d84b4d45c-j5rkj" podUID="8f337bd4-4533-4599-8cb8-7bedb8e2c420" containerName="dnsmasq-dns" containerID="cri-o://18cc26322bc53a4671779f1e298b36146b2936a6e5533b627269f5818312c95c" gracePeriod=10 Dec 06 06:22:46 crc kubenswrapper[4809]: E1206 06:22:46.034558 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f337bd4_4533_4599_8cb8_7bedb8e2c420.slice/crio-18cc26322bc53a4671779f1e298b36146b2936a6e5533b627269f5818312c95c.scope\": RecentStats: unable to find data in memory cache]" Dec 06 06:22:46 crc kubenswrapper[4809]: I1206 06:22:46.037267 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7fcd5d8964-kw6sz" event={"ID":"a71b0d8f-059d-4801-a2e4-6dc8ec44c78a","Type":"ContainerStarted","Data":"f6a84853672d7d8584cde18cdb0aff480e869884766f518fbe94603f1ba07c0e"} Dec 06 06:22:46 crc kubenswrapper[4809]: I1206 06:22:46.040250 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-794dfc48dc-pkjpx" event={"ID":"b09a69de-0d5f-44f7-a3c9-2e6902da17b8","Type":"ContainerStarted","Data":"74fe81a4841eeb6c8335555ad670a2dc4c4b67963020103d1da3f4868665fe84"} Dec 06 06:22:47 crc kubenswrapper[4809]: I1206 06:22:47.060913 4809 generic.go:334] "Generic (PLEG): container finished" podID="8f337bd4-4533-4599-8cb8-7bedb8e2c420" containerID="18cc26322bc53a4671779f1e298b36146b2936a6e5533b627269f5818312c95c" exitCode=0 Dec 06 06:22:47 crc kubenswrapper[4809]: I1206 06:22:47.061265 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-j5rkj" event={"ID":"8f337bd4-4533-4599-8cb8-7bedb8e2c420","Type":"ContainerDied","Data":"18cc26322bc53a4671779f1e298b36146b2936a6e5533b627269f5818312c95c"} Dec 06 06:22:47 crc kubenswrapper[4809]: I1206 06:22:47.062826 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-7849fbcbcb-msd4w" event={"ID":"dccd4eeb-4ba8-4200-8366-28d4ead2e476","Type":"ContainerStarted","Data":"4f5c1c1b5bba24805b2dc938e5e1cbf56ce15e9f49ff2c715ca001cc536461e4"} Dec 06 06:22:47 crc kubenswrapper[4809]: I1206 06:22:47.798257 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-j5rkj" Dec 06 06:22:47 crc kubenswrapper[4809]: I1206 06:22:47.874105 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8f337bd4-4533-4599-8cb8-7bedb8e2c420-dns-svc\") pod \"8f337bd4-4533-4599-8cb8-7bedb8e2c420\" (UID: \"8f337bd4-4533-4599-8cb8-7bedb8e2c420\") " Dec 06 06:22:47 crc kubenswrapper[4809]: I1206 06:22:47.874290 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8l88f\" (UniqueName: \"kubernetes.io/projected/8f337bd4-4533-4599-8cb8-7bedb8e2c420-kube-api-access-8l88f\") pod \"8f337bd4-4533-4599-8cb8-7bedb8e2c420\" (UID: \"8f337bd4-4533-4599-8cb8-7bedb8e2c420\") " Dec 06 06:22:47 crc kubenswrapper[4809]: I1206 06:22:47.874400 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f337bd4-4533-4599-8cb8-7bedb8e2c420-config\") pod \"8f337bd4-4533-4599-8cb8-7bedb8e2c420\" (UID: \"8f337bd4-4533-4599-8cb8-7bedb8e2c420\") " Dec 06 06:22:47 crc kubenswrapper[4809]: I1206 06:22:47.874450 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/8f337bd4-4533-4599-8cb8-7bedb8e2c420-openstack-edpm-ipam\") pod \"8f337bd4-4533-4599-8cb8-7bedb8e2c420\" (UID: \"8f337bd4-4533-4599-8cb8-7bedb8e2c420\") " Dec 06 06:22:47 crc kubenswrapper[4809]: I1206 06:22:47.874503 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8f337bd4-4533-4599-8cb8-7bedb8e2c420-ovsdbserver-sb\") pod \"8f337bd4-4533-4599-8cb8-7bedb8e2c420\" (UID: \"8f337bd4-4533-4599-8cb8-7bedb8e2c420\") " Dec 06 06:22:47 crc kubenswrapper[4809]: I1206 06:22:47.874539 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8f337bd4-4533-4599-8cb8-7bedb8e2c420-ovsdbserver-nb\") pod \"8f337bd4-4533-4599-8cb8-7bedb8e2c420\" (UID: \"8f337bd4-4533-4599-8cb8-7bedb8e2c420\") " Dec 06 06:22:47 crc kubenswrapper[4809]: I1206 06:22:47.874581 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8f337bd4-4533-4599-8cb8-7bedb8e2c420-dns-swift-storage-0\") pod \"8f337bd4-4533-4599-8cb8-7bedb8e2c420\" (UID: \"8f337bd4-4533-4599-8cb8-7bedb8e2c420\") " Dec 06 06:22:47 crc kubenswrapper[4809]: I1206 06:22:47.963229 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f337bd4-4533-4599-8cb8-7bedb8e2c420-kube-api-access-8l88f" (OuterVolumeSpecName: "kube-api-access-8l88f") pod "8f337bd4-4533-4599-8cb8-7bedb8e2c420" (UID: "8f337bd4-4533-4599-8cb8-7bedb8e2c420"). InnerVolumeSpecName "kube-api-access-8l88f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:22:47 crc kubenswrapper[4809]: I1206 06:22:47.986343 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8l88f\" (UniqueName: \"kubernetes.io/projected/8f337bd4-4533-4599-8cb8-7bedb8e2c420-kube-api-access-8l88f\") on node \"crc\" DevicePath \"\"" Dec 06 06:22:47 crc kubenswrapper[4809]: I1206 06:22:47.996613 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f337bd4-4533-4599-8cb8-7bedb8e2c420-config" (OuterVolumeSpecName: "config") pod "8f337bd4-4533-4599-8cb8-7bedb8e2c420" (UID: "8f337bd4-4533-4599-8cb8-7bedb8e2c420"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:22:47 crc kubenswrapper[4809]: I1206 06:22:47.996959 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f337bd4-4533-4599-8cb8-7bedb8e2c420-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "8f337bd4-4533-4599-8cb8-7bedb8e2c420" (UID: "8f337bd4-4533-4599-8cb8-7bedb8e2c420"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:22:48 crc kubenswrapper[4809]: I1206 06:22:48.055004 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f337bd4-4533-4599-8cb8-7bedb8e2c420-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8f337bd4-4533-4599-8cb8-7bedb8e2c420" (UID: "8f337bd4-4533-4599-8cb8-7bedb8e2c420"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:22:48 crc kubenswrapper[4809]: I1206 06:22:48.088004 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f337bd4-4533-4599-8cb8-7bedb8e2c420-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:22:48 crc kubenswrapper[4809]: I1206 06:22:48.088033 4809 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/8f337bd4-4533-4599-8cb8-7bedb8e2c420-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 06 06:22:48 crc kubenswrapper[4809]: I1206 06:22:48.088045 4809 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8f337bd4-4533-4599-8cb8-7bedb8e2c420-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:22:48 crc kubenswrapper[4809]: I1206 06:22:48.101810 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-j5rkj" event={"ID":"8f337bd4-4533-4599-8cb8-7bedb8e2c420","Type":"ContainerDied","Data":"4e09cc8c2323be25bad6170beae1a37be0bf5c1c3ec7d35f63115dac2c9e5f53"} Dec 06 06:22:48 crc kubenswrapper[4809]: I1206 06:22:48.101869 4809 scope.go:117] "RemoveContainer" containerID="18cc26322bc53a4671779f1e298b36146b2936a6e5533b627269f5818312c95c" Dec 06 06:22:48 crc kubenswrapper[4809]: I1206 06:22:48.102068 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-j5rkj" Dec 06 06:22:48 crc kubenswrapper[4809]: I1206 06:22:48.134625 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7fcd5d8964-kw6sz" event={"ID":"a71b0d8f-059d-4801-a2e4-6dc8ec44c78a","Type":"ContainerStarted","Data":"f6e032690d7cd3f99c2c0ad00b16c6ce8cffa5e5a3a1b7eeee52870960ff912a"} Dec 06 06:22:48 crc kubenswrapper[4809]: I1206 06:22:48.136158 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-7fcd5d8964-kw6sz" Dec 06 06:22:48 crc kubenswrapper[4809]: I1206 06:22:48.138564 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f337bd4-4533-4599-8cb8-7bedb8e2c420-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8f337bd4-4533-4599-8cb8-7bedb8e2c420" (UID: "8f337bd4-4533-4599-8cb8-7bedb8e2c420"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:22:48 crc kubenswrapper[4809]: I1206 06:22:48.193769 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8f337bd4-4533-4599-8cb8-7bedb8e2c420-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 06:22:48 crc kubenswrapper[4809]: I1206 06:22:48.195708 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-7fcd5d8964-kw6sz" podStartSLOduration=5.195695763 podStartE2EDuration="5.195695763s" podCreationTimestamp="2025-12-06 06:22:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:22:48.195499967 +0000 UTC m=+1893.084482909" watchObservedRunningTime="2025-12-06 06:22:48.195695763 +0000 UTC m=+1893.084678705" Dec 06 06:22:48 crc kubenswrapper[4809]: I1206 06:22:48.214995 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f337bd4-4533-4599-8cb8-7bedb8e2c420-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8f337bd4-4533-4599-8cb8-7bedb8e2c420" (UID: "8f337bd4-4533-4599-8cb8-7bedb8e2c420"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:22:48 crc kubenswrapper[4809]: I1206 06:22:48.232675 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f337bd4-4533-4599-8cb8-7bedb8e2c420-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8f337bd4-4533-4599-8cb8-7bedb8e2c420" (UID: "8f337bd4-4533-4599-8cb8-7bedb8e2c420"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:22:48 crc kubenswrapper[4809]: I1206 06:22:48.295824 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8f337bd4-4533-4599-8cb8-7bedb8e2c420-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 06:22:48 crc kubenswrapper[4809]: I1206 06:22:48.295860 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8f337bd4-4533-4599-8cb8-7bedb8e2c420-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 06:22:48 crc kubenswrapper[4809]: I1206 06:22:48.457162 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-j5rkj"] Dec 06 06:22:48 crc kubenswrapper[4809]: I1206 06:22:48.475586 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-j5rkj"] Dec 06 06:22:49 crc kubenswrapper[4809]: I1206 06:22:49.401919 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f337bd4-4533-4599-8cb8-7bedb8e2c420" path="/var/lib/kubelet/pods/8f337bd4-4533-4599-8cb8-7bedb8e2c420/volumes" Dec 06 06:22:49 crc kubenswrapper[4809]: I1206 06:22:49.581130 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/metallb-operator-webhook-server-765b7d7f99-8dngt" podUID="60798ea0-75be-4a0c-a8c6-9fb431ba0e67" containerName="webhook-server" probeResult="failure" output="Get \"http://10.217.0.93:7472/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 06:22:52 crc kubenswrapper[4809]: I1206 06:22:52.307988 4809 scope.go:117] "RemoveContainer" containerID="e5cb2803c48c66ea417847983d79cf128cf6ebf3a3d21ece49b50e6aba9b35d6" Dec 06 06:22:52 crc kubenswrapper[4809]: I1206 06:22:52.912588 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lkm6p" podUID="81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf" containerName="registry-server" probeResult="failure" output=< Dec 06 06:22:52 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Dec 06 06:22:52 crc kubenswrapper[4809]: > Dec 06 06:22:53 crc kubenswrapper[4809]: I1206 06:22:53.193584 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-7849fbcbcb-msd4w" event={"ID":"dccd4eeb-4ba8-4200-8366-28d4ead2e476","Type":"ContainerStarted","Data":"a58040540b107c6b62bf7c0d52e6f0378dc8d73f9f1f1763d797bcd89cd00eff"} Dec 06 06:22:53 crc kubenswrapper[4809]: I1206 06:22:53.195687 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-7849fbcbcb-msd4w" Dec 06 06:22:53 crc kubenswrapper[4809]: I1206 06:22:53.198586 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-794dfc48dc-pkjpx" event={"ID":"b09a69de-0d5f-44f7-a3c9-2e6902da17b8","Type":"ContainerStarted","Data":"d427ab350eae8b51f0f08076ecb224a09506002454adcb3a5f8205662ad5bd80"} Dec 06 06:22:53 crc kubenswrapper[4809]: I1206 06:22:53.199447 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-794dfc48dc-pkjpx" Dec 06 06:22:53 crc kubenswrapper[4809]: I1206 06:22:53.229997 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-7849fbcbcb-msd4w" podStartSLOduration=3.463555956 podStartE2EDuration="9.229971814s" podCreationTimestamp="2025-12-06 06:22:44 +0000 UTC" firstStartedPulling="2025-12-06 06:22:47.002782902 +0000 UTC m=+1891.891765864" lastFinishedPulling="2025-12-06 06:22:52.76919878 +0000 UTC m=+1897.658181722" observedRunningTime="2025-12-06 06:22:53.211715271 +0000 UTC m=+1898.100698213" watchObservedRunningTime="2025-12-06 06:22:53.229971814 +0000 UTC m=+1898.118954756" Dec 06 06:22:53 crc kubenswrapper[4809]: I1206 06:22:53.249058 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-794dfc48dc-pkjpx" podStartSLOduration=2.597479559 podStartE2EDuration="10.249035278s" podCreationTimestamp="2025-12-06 06:22:43 +0000 UTC" firstStartedPulling="2025-12-06 06:22:45.09613088 +0000 UTC m=+1889.985113822" lastFinishedPulling="2025-12-06 06:22:52.747686599 +0000 UTC m=+1897.636669541" observedRunningTime="2025-12-06 06:22:53.235221705 +0000 UTC m=+1898.124204647" watchObservedRunningTime="2025-12-06 06:22:53.249035278 +0000 UTC m=+1898.138018220" Dec 06 06:22:56 crc kubenswrapper[4809]: I1206 06:22:56.388682 4809 scope.go:117] "RemoveContainer" containerID="9bc2dc88b3b9f309ce0169aa510326dd77022dece4c69cab3141df00fe5c0106" Dec 06 06:22:56 crc kubenswrapper[4809]: E1206 06:22:56.389469 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:23:00 crc kubenswrapper[4809]: I1206 06:23:00.145417 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-s5cs4"] Dec 06 06:23:00 crc kubenswrapper[4809]: E1206 06:23:00.150420 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f337bd4-4533-4599-8cb8-7bedb8e2c420" containerName="init" Dec 06 06:23:00 crc kubenswrapper[4809]: I1206 06:23:00.150461 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f337bd4-4533-4599-8cb8-7bedb8e2c420" containerName="init" Dec 06 06:23:00 crc kubenswrapper[4809]: E1206 06:23:00.150491 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f337bd4-4533-4599-8cb8-7bedb8e2c420" containerName="dnsmasq-dns" Dec 06 06:23:00 crc kubenswrapper[4809]: I1206 06:23:00.150501 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f337bd4-4533-4599-8cb8-7bedb8e2c420" containerName="dnsmasq-dns" Dec 06 06:23:00 crc kubenswrapper[4809]: I1206 06:23:00.150997 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f337bd4-4533-4599-8cb8-7bedb8e2c420" containerName="dnsmasq-dns" Dec 06 06:23:00 crc kubenswrapper[4809]: I1206 06:23:00.152079 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-s5cs4" Dec 06 06:23:00 crc kubenswrapper[4809]: I1206 06:23:00.160895 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 06:23:00 crc kubenswrapper[4809]: I1206 06:23:00.161744 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 06:23:00 crc kubenswrapper[4809]: I1206 06:23:00.165766 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 06:23:00 crc kubenswrapper[4809]: I1206 06:23:00.166644 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-bzlf5" Dec 06 06:23:00 crc kubenswrapper[4809]: I1206 06:23:00.180085 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-s5cs4"] Dec 06 06:23:00 crc kubenswrapper[4809]: I1206 06:23:00.199406 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nv47x\" (UniqueName: \"kubernetes.io/projected/55c8fbdf-1083-4893-9e7c-76f98c35d656-kube-api-access-nv47x\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-s5cs4\" (UID: \"55c8fbdf-1083-4893-9e7c-76f98c35d656\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-s5cs4" Dec 06 06:23:00 crc kubenswrapper[4809]: I1206 06:23:00.199537 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/55c8fbdf-1083-4893-9e7c-76f98c35d656-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-s5cs4\" (UID: \"55c8fbdf-1083-4893-9e7c-76f98c35d656\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-s5cs4" Dec 06 06:23:00 crc kubenswrapper[4809]: I1206 06:23:00.199578 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55c8fbdf-1083-4893-9e7c-76f98c35d656-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-s5cs4\" (UID: \"55c8fbdf-1083-4893-9e7c-76f98c35d656\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-s5cs4" Dec 06 06:23:00 crc kubenswrapper[4809]: I1206 06:23:00.199620 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/55c8fbdf-1083-4893-9e7c-76f98c35d656-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-s5cs4\" (UID: \"55c8fbdf-1083-4893-9e7c-76f98c35d656\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-s5cs4" Dec 06 06:23:00 crc kubenswrapper[4809]: I1206 06:23:00.302679 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nv47x\" (UniqueName: \"kubernetes.io/projected/55c8fbdf-1083-4893-9e7c-76f98c35d656-kube-api-access-nv47x\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-s5cs4\" (UID: \"55c8fbdf-1083-4893-9e7c-76f98c35d656\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-s5cs4" Dec 06 06:23:00 crc kubenswrapper[4809]: I1206 06:23:00.302835 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/55c8fbdf-1083-4893-9e7c-76f98c35d656-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-s5cs4\" (UID: \"55c8fbdf-1083-4893-9e7c-76f98c35d656\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-s5cs4" Dec 06 06:23:00 crc kubenswrapper[4809]: I1206 06:23:00.302886 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55c8fbdf-1083-4893-9e7c-76f98c35d656-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-s5cs4\" (UID: \"55c8fbdf-1083-4893-9e7c-76f98c35d656\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-s5cs4" Dec 06 06:23:00 crc kubenswrapper[4809]: I1206 06:23:00.302955 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/55c8fbdf-1083-4893-9e7c-76f98c35d656-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-s5cs4\" (UID: \"55c8fbdf-1083-4893-9e7c-76f98c35d656\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-s5cs4" Dec 06 06:23:00 crc kubenswrapper[4809]: I1206 06:23:00.304427 4809 generic.go:334] "Generic (PLEG): container finished" podID="2cc53bf7-6e54-427a-9b63-88d694609c75" containerID="929ef76322eb9f4df027d760ed9c5f62890de92642910bf48ccd63f92cb62498" exitCode=0 Dec 06 06:23:00 crc kubenswrapper[4809]: I1206 06:23:00.304485 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2cc53bf7-6e54-427a-9b63-88d694609c75","Type":"ContainerDied","Data":"929ef76322eb9f4df027d760ed9c5f62890de92642910bf48ccd63f92cb62498"} Dec 06 06:23:00 crc kubenswrapper[4809]: I1206 06:23:00.309495 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55c8fbdf-1083-4893-9e7c-76f98c35d656-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-s5cs4\" (UID: \"55c8fbdf-1083-4893-9e7c-76f98c35d656\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-s5cs4" Dec 06 06:23:00 crc kubenswrapper[4809]: I1206 06:23:00.313783 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/55c8fbdf-1083-4893-9e7c-76f98c35d656-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-s5cs4\" (UID: \"55c8fbdf-1083-4893-9e7c-76f98c35d656\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-s5cs4" Dec 06 06:23:00 crc kubenswrapper[4809]: I1206 06:23:00.327182 4809 generic.go:334] "Generic (PLEG): container finished" podID="20646b41-62ab-42e7-8b11-538605db0506" containerID="7fdbe581e08cf2fa165721423a7f859c28b8d58546515e9d601549721fefa0af" exitCode=0 Dec 06 06:23:00 crc kubenswrapper[4809]: I1206 06:23:00.327241 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"20646b41-62ab-42e7-8b11-538605db0506","Type":"ContainerDied","Data":"7fdbe581e08cf2fa165721423a7f859c28b8d58546515e9d601549721fefa0af"} Dec 06 06:23:00 crc kubenswrapper[4809]: I1206 06:23:00.330539 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/55c8fbdf-1083-4893-9e7c-76f98c35d656-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-s5cs4\" (UID: \"55c8fbdf-1083-4893-9e7c-76f98c35d656\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-s5cs4" Dec 06 06:23:00 crc kubenswrapper[4809]: I1206 06:23:00.363855 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nv47x\" (UniqueName: \"kubernetes.io/projected/55c8fbdf-1083-4893-9e7c-76f98c35d656-kube-api-access-nv47x\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-s5cs4\" (UID: \"55c8fbdf-1083-4893-9e7c-76f98c35d656\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-s5cs4" Dec 06 06:23:00 crc kubenswrapper[4809]: I1206 06:23:00.410341 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-794dfc48dc-pkjpx" Dec 06 06:23:00 crc kubenswrapper[4809]: I1206 06:23:00.469799 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-s5cs4" Dec 06 06:23:00 crc kubenswrapper[4809]: I1206 06:23:00.518444 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-9f9dd4fbb-4pcn7"] Dec 06 06:23:00 crc kubenswrapper[4809]: I1206 06:23:00.518719 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-9f9dd4fbb-4pcn7" podUID="e2c1d3c1-fa26-4525-bca8-3449d4535409" containerName="heat-cfnapi" containerID="cri-o://4502d90eeaeae3ae19d45c87f3b707c6d81d5e4600ae45736dbc55681e40ee56" gracePeriod=60 Dec 06 06:23:01 crc kubenswrapper[4809]: I1206 06:23:01.344557 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"20646b41-62ab-42e7-8b11-538605db0506","Type":"ContainerStarted","Data":"616bd2a9d85beffb1358b5c621d1af7825c0b5627032a1071f768893b14fe82a"} Dec 06 06:23:01 crc kubenswrapper[4809]: I1206 06:23:01.345409 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:23:01 crc kubenswrapper[4809]: I1206 06:23:01.348751 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2cc53bf7-6e54-427a-9b63-88d694609c75","Type":"ContainerStarted","Data":"23d907a27ca81415cebc88510019b85f6bfd814f82ce5827a9d6c44d31147632"} Dec 06 06:23:01 crc kubenswrapper[4809]: I1206 06:23:01.348932 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 06 06:23:01 crc kubenswrapper[4809]: I1206 06:23:01.375116 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=40.375090731 podStartE2EDuration="40.375090731s" podCreationTimestamp="2025-12-06 06:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:23:01.367962129 +0000 UTC m=+1906.256945091" watchObservedRunningTime="2025-12-06 06:23:01.375090731 +0000 UTC m=+1906.264073703" Dec 06 06:23:01 crc kubenswrapper[4809]: I1206 06:23:01.399030 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=54.399006017 podStartE2EDuration="54.399006017s" podCreationTimestamp="2025-12-06 06:22:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:23:01.396073637 +0000 UTC m=+1906.285056579" watchObservedRunningTime="2025-12-06 06:23:01.399006017 +0000 UTC m=+1906.287988959" Dec 06 06:23:01 crc kubenswrapper[4809]: I1206 06:23:01.615612 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-s5cs4"] Dec 06 06:23:01 crc kubenswrapper[4809]: I1206 06:23:01.913507 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lkm6p" Dec 06 06:23:01 crc kubenswrapper[4809]: I1206 06:23:01.984264 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lkm6p" Dec 06 06:23:02 crc kubenswrapper[4809]: I1206 06:23:02.083673 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-7849fbcbcb-msd4w" Dec 06 06:23:02 crc kubenswrapper[4809]: I1206 06:23:02.158399 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-bdfbd85-ntgxq"] Dec 06 06:23:02 crc kubenswrapper[4809]: I1206 06:23:02.158812 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-bdfbd85-ntgxq" podUID="4d76c911-21ec-4f29-9707-e8101ed3aed9" containerName="heat-api" containerID="cri-o://fc6a1d95e1293610b382f626e7c6ef59525c05b478a9c00131b4ad52b7b351a1" gracePeriod=60 Dec 06 06:23:02 crc kubenswrapper[4809]: I1206 06:23:02.177184 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lkm6p"] Dec 06 06:23:02 crc kubenswrapper[4809]: I1206 06:23:02.371477 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-s5cs4" event={"ID":"55c8fbdf-1083-4893-9e7c-76f98c35d656","Type":"ContainerStarted","Data":"3792d2fa039bd2f00f8b66a50a9eb27baf97fffcd2197fc49d1e204a3881e112"} Dec 06 06:23:03 crc kubenswrapper[4809]: I1206 06:23:03.394158 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lkm6p" podUID="81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf" containerName="registry-server" containerID="cri-o://17be1db1527822a3f5d47f6d584bab2952bcd5836d7afc29db89604138a6ac96" gracePeriod=2 Dec 06 06:23:03 crc kubenswrapper[4809]: I1206 06:23:03.598565 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-7fcd5d8964-kw6sz" Dec 06 06:23:03 crc kubenswrapper[4809]: I1206 06:23:03.666791 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-6d8ccd7c9f-nbxnr"] Dec 06 06:23:03 crc kubenswrapper[4809]: I1206 06:23:03.669417 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-6d8ccd7c9f-nbxnr" podUID="04d8a790-dedd-44c4-9b8f-093abb541e73" containerName="heat-engine" containerID="cri-o://3178492e382f3c8585308c4ee39f1caea0bb67c6f97da76aea6237e1669dc416" gracePeriod=60 Dec 06 06:23:03 crc kubenswrapper[4809]: I1206 06:23:03.798513 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-9f9dd4fbb-4pcn7" podUID="e2c1d3c1-fa26-4525-bca8-3449d4535409" containerName="heat-cfnapi" probeResult="failure" output="Get \"https://10.217.0.215:8000/healthcheck\": read tcp 10.217.0.2:43554->10.217.0.215:8000: read: connection reset by peer" Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.140712 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lkm6p" Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.232591 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgzqt\" (UniqueName: \"kubernetes.io/projected/81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf-kube-api-access-bgzqt\") pod \"81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf\" (UID: \"81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf\") " Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.232990 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf-utilities\") pod \"81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf\" (UID: \"81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf\") " Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.233115 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf-catalog-content\") pod \"81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf\" (UID: \"81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf\") " Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.240774 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf-utilities" (OuterVolumeSpecName: "utilities") pod "81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf" (UID: "81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.263541 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf-kube-api-access-bgzqt" (OuterVolumeSpecName: "kube-api-access-bgzqt") pod "81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf" (UID: "81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf"). InnerVolumeSpecName "kube-api-access-bgzqt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.339326 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.339382 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgzqt\" (UniqueName: \"kubernetes.io/projected/81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf-kube-api-access-bgzqt\") on node \"crc\" DevicePath \"\"" Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.445782 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-9f9dd4fbb-4pcn7" Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.446985 4809 generic.go:334] "Generic (PLEG): container finished" podID="81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf" containerID="17be1db1527822a3f5d47f6d584bab2952bcd5836d7afc29db89604138a6ac96" exitCode=0 Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.447087 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lkm6p" event={"ID":"81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf","Type":"ContainerDied","Data":"17be1db1527822a3f5d47f6d584bab2952bcd5836d7afc29db89604138a6ac96"} Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.447123 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lkm6p" event={"ID":"81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf","Type":"ContainerDied","Data":"b9a7d47a334a975be6b37523c4b63f20f2de3910a5e8f9094bc13442d7102cfe"} Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.447146 4809 scope.go:117] "RemoveContainer" containerID="17be1db1527822a3f5d47f6d584bab2952bcd5836d7afc29db89604138a6ac96" Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.447802 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lkm6p" Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.460051 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf" (UID: "81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.468788 4809 generic.go:334] "Generic (PLEG): container finished" podID="e2c1d3c1-fa26-4525-bca8-3449d4535409" containerID="4502d90eeaeae3ae19d45c87f3b707c6d81d5e4600ae45736dbc55681e40ee56" exitCode=0 Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.468835 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-9f9dd4fbb-4pcn7" event={"ID":"e2c1d3c1-fa26-4525-bca8-3449d4535409","Type":"ContainerDied","Data":"4502d90eeaeae3ae19d45c87f3b707c6d81d5e4600ae45736dbc55681e40ee56"} Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.468870 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-9f9dd4fbb-4pcn7" Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.523049 4809 scope.go:117] "RemoveContainer" containerID="a4c656c4e48f2889f600c3e5ad29ed503a3315ff146bb9bf3f02bb05468d2fbb" Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.544153 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2c1d3c1-fa26-4525-bca8-3449d4535409-internal-tls-certs\") pod \"e2c1d3c1-fa26-4525-bca8-3449d4535409\" (UID: \"e2c1d3c1-fa26-4525-bca8-3449d4535409\") " Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.544234 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2c1d3c1-fa26-4525-bca8-3449d4535409-public-tls-certs\") pod \"e2c1d3c1-fa26-4525-bca8-3449d4535409\" (UID: \"e2c1d3c1-fa26-4525-bca8-3449d4535409\") " Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.544308 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2c1d3c1-fa26-4525-bca8-3449d4535409-config-data\") pod \"e2c1d3c1-fa26-4525-bca8-3449d4535409\" (UID: \"e2c1d3c1-fa26-4525-bca8-3449d4535409\") " Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.544474 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e2c1d3c1-fa26-4525-bca8-3449d4535409-config-data-custom\") pod \"e2c1d3c1-fa26-4525-bca8-3449d4535409\" (UID: \"e2c1d3c1-fa26-4525-bca8-3449d4535409\") " Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.544548 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2c1d3c1-fa26-4525-bca8-3449d4535409-combined-ca-bundle\") pod \"e2c1d3c1-fa26-4525-bca8-3449d4535409\" (UID: \"e2c1d3c1-fa26-4525-bca8-3449d4535409\") " Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.544661 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grs95\" (UniqueName: \"kubernetes.io/projected/e2c1d3c1-fa26-4525-bca8-3449d4535409-kube-api-access-grs95\") pod \"e2c1d3c1-fa26-4525-bca8-3449d4535409\" (UID: \"e2c1d3c1-fa26-4525-bca8-3449d4535409\") " Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.545836 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.556129 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2c1d3c1-fa26-4525-bca8-3449d4535409-kube-api-access-grs95" (OuterVolumeSpecName: "kube-api-access-grs95") pod "e2c1d3c1-fa26-4525-bca8-3449d4535409" (UID: "e2c1d3c1-fa26-4525-bca8-3449d4535409"). InnerVolumeSpecName "kube-api-access-grs95". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.558857 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2c1d3c1-fa26-4525-bca8-3449d4535409-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e2c1d3c1-fa26-4525-bca8-3449d4535409" (UID: "e2c1d3c1-fa26-4525-bca8-3449d4535409"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.613566 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2c1d3c1-fa26-4525-bca8-3449d4535409-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e2c1d3c1-fa26-4525-bca8-3449d4535409" (UID: "e2c1d3c1-fa26-4525-bca8-3449d4535409"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.631517 4809 scope.go:117] "RemoveContainer" containerID="799aefafe4febe7c505514f10f778adba637b52eda146b500661551502382ec5" Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.648279 4809 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e2c1d3c1-fa26-4525-bca8-3449d4535409-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.649197 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2c1d3c1-fa26-4525-bca8-3449d4535409-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.649207 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grs95\" (UniqueName: \"kubernetes.io/projected/e2c1d3c1-fa26-4525-bca8-3449d4535409-kube-api-access-grs95\") on node \"crc\" DevicePath \"\"" Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.686673 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2c1d3c1-fa26-4525-bca8-3449d4535409-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "e2c1d3c1-fa26-4525-bca8-3449d4535409" (UID: "e2c1d3c1-fa26-4525-bca8-3449d4535409"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.716172 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2c1d3c1-fa26-4525-bca8-3449d4535409-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "e2c1d3c1-fa26-4525-bca8-3449d4535409" (UID: "e2c1d3c1-fa26-4525-bca8-3449d4535409"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.739117 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2c1d3c1-fa26-4525-bca8-3449d4535409-config-data" (OuterVolumeSpecName: "config-data") pod "e2c1d3c1-fa26-4525-bca8-3449d4535409" (UID: "e2c1d3c1-fa26-4525-bca8-3449d4535409"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.753811 4809 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2c1d3c1-fa26-4525-bca8-3449d4535409-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.753851 4809 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2c1d3c1-fa26-4525-bca8-3449d4535409-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.753863 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2c1d3c1-fa26-4525-bca8-3449d4535409-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.860705 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-9f9dd4fbb-4pcn7"] Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.884899 4809 scope.go:117] "RemoveContainer" containerID="17be1db1527822a3f5d47f6d584bab2952bcd5836d7afc29db89604138a6ac96" Dec 06 06:23:04 crc kubenswrapper[4809]: E1206 06:23:04.885576 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17be1db1527822a3f5d47f6d584bab2952bcd5836d7afc29db89604138a6ac96\": container with ID starting with 17be1db1527822a3f5d47f6d584bab2952bcd5836d7afc29db89604138a6ac96 not found: ID does not exist" containerID="17be1db1527822a3f5d47f6d584bab2952bcd5836d7afc29db89604138a6ac96" Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.885618 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17be1db1527822a3f5d47f6d584bab2952bcd5836d7afc29db89604138a6ac96"} err="failed to get container status \"17be1db1527822a3f5d47f6d584bab2952bcd5836d7afc29db89604138a6ac96\": rpc error: code = NotFound desc = could not find container \"17be1db1527822a3f5d47f6d584bab2952bcd5836d7afc29db89604138a6ac96\": container with ID starting with 17be1db1527822a3f5d47f6d584bab2952bcd5836d7afc29db89604138a6ac96 not found: ID does not exist" Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.885649 4809 scope.go:117] "RemoveContainer" containerID="a4c656c4e48f2889f600c3e5ad29ed503a3315ff146bb9bf3f02bb05468d2fbb" Dec 06 06:23:04 crc kubenswrapper[4809]: E1206 06:23:04.893226 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4c656c4e48f2889f600c3e5ad29ed503a3315ff146bb9bf3f02bb05468d2fbb\": container with ID starting with a4c656c4e48f2889f600c3e5ad29ed503a3315ff146bb9bf3f02bb05468d2fbb not found: ID does not exist" containerID="a4c656c4e48f2889f600c3e5ad29ed503a3315ff146bb9bf3f02bb05468d2fbb" Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.893464 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4c656c4e48f2889f600c3e5ad29ed503a3315ff146bb9bf3f02bb05468d2fbb"} err="failed to get container status \"a4c656c4e48f2889f600c3e5ad29ed503a3315ff146bb9bf3f02bb05468d2fbb\": rpc error: code = NotFound desc = could not find container \"a4c656c4e48f2889f600c3e5ad29ed503a3315ff146bb9bf3f02bb05468d2fbb\": container with ID starting with a4c656c4e48f2889f600c3e5ad29ed503a3315ff146bb9bf3f02bb05468d2fbb not found: ID does not exist" Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.893539 4809 scope.go:117] "RemoveContainer" containerID="799aefafe4febe7c505514f10f778adba637b52eda146b500661551502382ec5" Dec 06 06:23:04 crc kubenswrapper[4809]: E1206 06:23:04.905355 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"799aefafe4febe7c505514f10f778adba637b52eda146b500661551502382ec5\": container with ID starting with 799aefafe4febe7c505514f10f778adba637b52eda146b500661551502382ec5 not found: ID does not exist" containerID="799aefafe4febe7c505514f10f778adba637b52eda146b500661551502382ec5" Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.905730 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"799aefafe4febe7c505514f10f778adba637b52eda146b500661551502382ec5"} err="failed to get container status \"799aefafe4febe7c505514f10f778adba637b52eda146b500661551502382ec5\": rpc error: code = NotFound desc = could not find container \"799aefafe4febe7c505514f10f778adba637b52eda146b500661551502382ec5\": container with ID starting with 799aefafe4febe7c505514f10f778adba637b52eda146b500661551502382ec5 not found: ID does not exist" Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.905822 4809 scope.go:117] "RemoveContainer" containerID="4502d90eeaeae3ae19d45c87f3b707c6d81d5e4600ae45736dbc55681e40ee56" Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.955894 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-9f9dd4fbb-4pcn7"] Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.969429 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lkm6p"] Dec 06 06:23:04 crc kubenswrapper[4809]: I1206 06:23:04.981854 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lkm6p"] Dec 06 06:23:05 crc kubenswrapper[4809]: I1206 06:23:05.406562 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf" path="/var/lib/kubelet/pods/81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf/volumes" Dec 06 06:23:05 crc kubenswrapper[4809]: I1206 06:23:05.407730 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2c1d3c1-fa26-4525-bca8-3449d4535409" path="/var/lib/kubelet/pods/e2c1d3c1-fa26-4525-bca8-3449d4535409/volumes" Dec 06 06:23:06 crc kubenswrapper[4809]: I1206 06:23:06.520333 4809 generic.go:334] "Generic (PLEG): container finished" podID="4d76c911-21ec-4f29-9707-e8101ed3aed9" containerID="fc6a1d95e1293610b382f626e7c6ef59525c05b478a9c00131b4ad52b7b351a1" exitCode=0 Dec 06 06:23:06 crc kubenswrapper[4809]: I1206 06:23:06.520407 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-bdfbd85-ntgxq" event={"ID":"4d76c911-21ec-4f29-9707-e8101ed3aed9","Type":"ContainerDied","Data":"fc6a1d95e1293610b382f626e7c6ef59525c05b478a9c00131b4ad52b7b351a1"} Dec 06 06:23:06 crc kubenswrapper[4809]: I1206 06:23:06.935158 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-xmngf"] Dec 06 06:23:06 crc kubenswrapper[4809]: I1206 06:23:06.945738 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-xmngf"] Dec 06 06:23:07 crc kubenswrapper[4809]: I1206 06:23:07.104177 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-q7rkl"] Dec 06 06:23:07 crc kubenswrapper[4809]: E1206 06:23:07.104895 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf" containerName="extract-content" Dec 06 06:23:07 crc kubenswrapper[4809]: I1206 06:23:07.104912 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf" containerName="extract-content" Dec 06 06:23:07 crc kubenswrapper[4809]: E1206 06:23:07.104941 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf" containerName="registry-server" Dec 06 06:23:07 crc kubenswrapper[4809]: I1206 06:23:07.104947 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf" containerName="registry-server" Dec 06 06:23:07 crc kubenswrapper[4809]: E1206 06:23:07.104984 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf" containerName="extract-utilities" Dec 06 06:23:07 crc kubenswrapper[4809]: I1206 06:23:07.104993 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf" containerName="extract-utilities" Dec 06 06:23:07 crc kubenswrapper[4809]: E1206 06:23:07.105021 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2c1d3c1-fa26-4525-bca8-3449d4535409" containerName="heat-cfnapi" Dec 06 06:23:07 crc kubenswrapper[4809]: I1206 06:23:07.105028 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2c1d3c1-fa26-4525-bca8-3449d4535409" containerName="heat-cfnapi" Dec 06 06:23:07 crc kubenswrapper[4809]: I1206 06:23:07.105235 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="81ebbc50-ccbd-4eb8-aeb5-5fdd5b27afbf" containerName="registry-server" Dec 06 06:23:07 crc kubenswrapper[4809]: I1206 06:23:07.105247 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2c1d3c1-fa26-4525-bca8-3449d4535409" containerName="heat-cfnapi" Dec 06 06:23:07 crc kubenswrapper[4809]: I1206 06:23:07.106056 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-q7rkl" Dec 06 06:23:07 crc kubenswrapper[4809]: I1206 06:23:07.108817 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 06 06:23:07 crc kubenswrapper[4809]: I1206 06:23:07.110459 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-q7rkl"] Dec 06 06:23:07 crc kubenswrapper[4809]: I1206 06:23:07.120163 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-bdfbd85-ntgxq" podUID="4d76c911-21ec-4f29-9707-e8101ed3aed9" containerName="heat-api" probeResult="failure" output="Get \"https://10.217.0.216:8004/healthcheck\": dial tcp 10.217.0.216:8004: connect: connection refused" Dec 06 06:23:07 crc kubenswrapper[4809]: I1206 06:23:07.215671 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a75c2789-4b84-455c-9215-2c52308dbace-combined-ca-bundle\") pod \"aodh-db-sync-q7rkl\" (UID: \"a75c2789-4b84-455c-9215-2c52308dbace\") " pod="openstack/aodh-db-sync-q7rkl" Dec 06 06:23:07 crc kubenswrapper[4809]: I1206 06:23:07.215854 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a75c2789-4b84-455c-9215-2c52308dbace-scripts\") pod \"aodh-db-sync-q7rkl\" (UID: \"a75c2789-4b84-455c-9215-2c52308dbace\") " pod="openstack/aodh-db-sync-q7rkl" Dec 06 06:23:07 crc kubenswrapper[4809]: I1206 06:23:07.216244 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gh2v\" (UniqueName: \"kubernetes.io/projected/a75c2789-4b84-455c-9215-2c52308dbace-kube-api-access-5gh2v\") pod \"aodh-db-sync-q7rkl\" (UID: \"a75c2789-4b84-455c-9215-2c52308dbace\") " pod="openstack/aodh-db-sync-q7rkl" Dec 06 06:23:07 crc kubenswrapper[4809]: I1206 06:23:07.216403 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a75c2789-4b84-455c-9215-2c52308dbace-config-data\") pod \"aodh-db-sync-q7rkl\" (UID: \"a75c2789-4b84-455c-9215-2c52308dbace\") " pod="openstack/aodh-db-sync-q7rkl" Dec 06 06:23:07 crc kubenswrapper[4809]: I1206 06:23:07.318536 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a75c2789-4b84-455c-9215-2c52308dbace-combined-ca-bundle\") pod \"aodh-db-sync-q7rkl\" (UID: \"a75c2789-4b84-455c-9215-2c52308dbace\") " pod="openstack/aodh-db-sync-q7rkl" Dec 06 06:23:07 crc kubenswrapper[4809]: I1206 06:23:07.318647 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a75c2789-4b84-455c-9215-2c52308dbace-scripts\") pod \"aodh-db-sync-q7rkl\" (UID: \"a75c2789-4b84-455c-9215-2c52308dbace\") " pod="openstack/aodh-db-sync-q7rkl" Dec 06 06:23:07 crc kubenswrapper[4809]: I1206 06:23:07.318773 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gh2v\" (UniqueName: \"kubernetes.io/projected/a75c2789-4b84-455c-9215-2c52308dbace-kube-api-access-5gh2v\") pod \"aodh-db-sync-q7rkl\" (UID: \"a75c2789-4b84-455c-9215-2c52308dbace\") " pod="openstack/aodh-db-sync-q7rkl" Dec 06 06:23:07 crc kubenswrapper[4809]: I1206 06:23:07.318838 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a75c2789-4b84-455c-9215-2c52308dbace-config-data\") pod \"aodh-db-sync-q7rkl\" (UID: \"a75c2789-4b84-455c-9215-2c52308dbace\") " pod="openstack/aodh-db-sync-q7rkl" Dec 06 06:23:07 crc kubenswrapper[4809]: I1206 06:23:07.327706 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a75c2789-4b84-455c-9215-2c52308dbace-scripts\") pod \"aodh-db-sync-q7rkl\" (UID: \"a75c2789-4b84-455c-9215-2c52308dbace\") " pod="openstack/aodh-db-sync-q7rkl" Dec 06 06:23:07 crc kubenswrapper[4809]: I1206 06:23:07.346243 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a75c2789-4b84-455c-9215-2c52308dbace-combined-ca-bundle\") pod \"aodh-db-sync-q7rkl\" (UID: \"a75c2789-4b84-455c-9215-2c52308dbace\") " pod="openstack/aodh-db-sync-q7rkl" Dec 06 06:23:07 crc kubenswrapper[4809]: I1206 06:23:07.349434 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gh2v\" (UniqueName: \"kubernetes.io/projected/a75c2789-4b84-455c-9215-2c52308dbace-kube-api-access-5gh2v\") pod \"aodh-db-sync-q7rkl\" (UID: \"a75c2789-4b84-455c-9215-2c52308dbace\") " pod="openstack/aodh-db-sync-q7rkl" Dec 06 06:23:07 crc kubenswrapper[4809]: I1206 06:23:07.359635 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a75c2789-4b84-455c-9215-2c52308dbace-config-data\") pod \"aodh-db-sync-q7rkl\" (UID: \"a75c2789-4b84-455c-9215-2c52308dbace\") " pod="openstack/aodh-db-sync-q7rkl" Dec 06 06:23:07 crc kubenswrapper[4809]: I1206 06:23:07.426242 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68932bed-06eb-4c47-843c-a171ccbe23c3" path="/var/lib/kubelet/pods/68932bed-06eb-4c47-843c-a171ccbe23c3/volumes" Dec 06 06:23:07 crc kubenswrapper[4809]: I1206 06:23:07.449762 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-q7rkl" Dec 06 06:23:08 crc kubenswrapper[4809]: I1206 06:23:08.016208 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 06 06:23:08 crc kubenswrapper[4809]: I1206 06:23:08.389516 4809 scope.go:117] "RemoveContainer" containerID="9bc2dc88b3b9f309ce0169aa510326dd77022dece4c69cab3141df00fe5c0106" Dec 06 06:23:08 crc kubenswrapper[4809]: E1206 06:23:08.390032 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:23:08 crc kubenswrapper[4809]: I1206 06:23:08.726549 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-bdfbd85-ntgxq" Dec 06 06:23:08 crc kubenswrapper[4809]: I1206 06:23:08.756679 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d76c911-21ec-4f29-9707-e8101ed3aed9-combined-ca-bundle\") pod \"4d76c911-21ec-4f29-9707-e8101ed3aed9\" (UID: \"4d76c911-21ec-4f29-9707-e8101ed3aed9\") " Dec 06 06:23:08 crc kubenswrapper[4809]: I1206 06:23:08.756761 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d76c911-21ec-4f29-9707-e8101ed3aed9-config-data\") pod \"4d76c911-21ec-4f29-9707-e8101ed3aed9\" (UID: \"4d76c911-21ec-4f29-9707-e8101ed3aed9\") " Dec 06 06:23:08 crc kubenswrapper[4809]: I1206 06:23:08.756843 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4d76c911-21ec-4f29-9707-e8101ed3aed9-config-data-custom\") pod \"4d76c911-21ec-4f29-9707-e8101ed3aed9\" (UID: \"4d76c911-21ec-4f29-9707-e8101ed3aed9\") " Dec 06 06:23:08 crc kubenswrapper[4809]: I1206 06:23:08.756900 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d76c911-21ec-4f29-9707-e8101ed3aed9-public-tls-certs\") pod \"4d76c911-21ec-4f29-9707-e8101ed3aed9\" (UID: \"4d76c911-21ec-4f29-9707-e8101ed3aed9\") " Dec 06 06:23:08 crc kubenswrapper[4809]: I1206 06:23:08.757211 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5z7dq\" (UniqueName: \"kubernetes.io/projected/4d76c911-21ec-4f29-9707-e8101ed3aed9-kube-api-access-5z7dq\") pod \"4d76c911-21ec-4f29-9707-e8101ed3aed9\" (UID: \"4d76c911-21ec-4f29-9707-e8101ed3aed9\") " Dec 06 06:23:08 crc kubenswrapper[4809]: I1206 06:23:08.757257 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d76c911-21ec-4f29-9707-e8101ed3aed9-internal-tls-certs\") pod \"4d76c911-21ec-4f29-9707-e8101ed3aed9\" (UID: \"4d76c911-21ec-4f29-9707-e8101ed3aed9\") " Dec 06 06:23:08 crc kubenswrapper[4809]: I1206 06:23:08.780757 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d76c911-21ec-4f29-9707-e8101ed3aed9-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "4d76c911-21ec-4f29-9707-e8101ed3aed9" (UID: "4d76c911-21ec-4f29-9707-e8101ed3aed9"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:23:08 crc kubenswrapper[4809]: I1206 06:23:08.798550 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d76c911-21ec-4f29-9707-e8101ed3aed9-kube-api-access-5z7dq" (OuterVolumeSpecName: "kube-api-access-5z7dq") pod "4d76c911-21ec-4f29-9707-e8101ed3aed9" (UID: "4d76c911-21ec-4f29-9707-e8101ed3aed9"). InnerVolumeSpecName "kube-api-access-5z7dq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:23:08 crc kubenswrapper[4809]: I1206 06:23:08.827443 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d76c911-21ec-4f29-9707-e8101ed3aed9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4d76c911-21ec-4f29-9707-e8101ed3aed9" (UID: "4d76c911-21ec-4f29-9707-e8101ed3aed9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:23:08 crc kubenswrapper[4809]: I1206 06:23:08.846308 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d76c911-21ec-4f29-9707-e8101ed3aed9-config-data" (OuterVolumeSpecName: "config-data") pod "4d76c911-21ec-4f29-9707-e8101ed3aed9" (UID: "4d76c911-21ec-4f29-9707-e8101ed3aed9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:23:08 crc kubenswrapper[4809]: I1206 06:23:08.861442 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5z7dq\" (UniqueName: \"kubernetes.io/projected/4d76c911-21ec-4f29-9707-e8101ed3aed9-kube-api-access-5z7dq\") on node \"crc\" DevicePath \"\"" Dec 06 06:23:08 crc kubenswrapper[4809]: I1206 06:23:08.861480 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d76c911-21ec-4f29-9707-e8101ed3aed9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:23:08 crc kubenswrapper[4809]: I1206 06:23:08.861495 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d76c911-21ec-4f29-9707-e8101ed3aed9-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:23:08 crc kubenswrapper[4809]: I1206 06:23:08.861509 4809 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4d76c911-21ec-4f29-9707-e8101ed3aed9-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 06:23:08 crc kubenswrapper[4809]: I1206 06:23:08.888147 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d76c911-21ec-4f29-9707-e8101ed3aed9-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "4d76c911-21ec-4f29-9707-e8101ed3aed9" (UID: "4d76c911-21ec-4f29-9707-e8101ed3aed9"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:23:08 crc kubenswrapper[4809]: I1206 06:23:08.930057 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d76c911-21ec-4f29-9707-e8101ed3aed9-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "4d76c911-21ec-4f29-9707-e8101ed3aed9" (UID: "4d76c911-21ec-4f29-9707-e8101ed3aed9"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:23:08 crc kubenswrapper[4809]: I1206 06:23:08.967217 4809 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d76c911-21ec-4f29-9707-e8101ed3aed9-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 06:23:08 crc kubenswrapper[4809]: I1206 06:23:08.967271 4809 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d76c911-21ec-4f29-9707-e8101ed3aed9-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 06:23:08 crc kubenswrapper[4809]: I1206 06:23:08.984357 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-q7rkl"] Dec 06 06:23:09 crc kubenswrapper[4809]: I1206 06:23:09.558469 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-q7rkl" event={"ID":"a75c2789-4b84-455c-9215-2c52308dbace","Type":"ContainerStarted","Data":"1f976f35d3d6d7f7332353c9334d021b963c0e517df2b844162d23e349fd6976"} Dec 06 06:23:09 crc kubenswrapper[4809]: I1206 06:23:09.560759 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-bdfbd85-ntgxq" event={"ID":"4d76c911-21ec-4f29-9707-e8101ed3aed9","Type":"ContainerDied","Data":"e77f2c6fca8113852b1528ba7eddcfbf74cc021cd24983b020fb02b1080a99e9"} Dec 06 06:23:09 crc kubenswrapper[4809]: I1206 06:23:09.560795 4809 scope.go:117] "RemoveContainer" containerID="fc6a1d95e1293610b382f626e7c6ef59525c05b478a9c00131b4ad52b7b351a1" Dec 06 06:23:09 crc kubenswrapper[4809]: I1206 06:23:09.560956 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-bdfbd85-ntgxq" Dec 06 06:23:09 crc kubenswrapper[4809]: I1206 06:23:09.636025 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-bdfbd85-ntgxq"] Dec 06 06:23:09 crc kubenswrapper[4809]: I1206 06:23:09.649535 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-bdfbd85-ntgxq"] Dec 06 06:23:09 crc kubenswrapper[4809]: E1206 06:23:09.889192 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3178492e382f3c8585308c4ee39f1caea0bb67c6f97da76aea6237e1669dc416" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 06 06:23:09 crc kubenswrapper[4809]: E1206 06:23:09.892230 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3178492e382f3c8585308c4ee39f1caea0bb67c6f97da76aea6237e1669dc416" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 06 06:23:09 crc kubenswrapper[4809]: E1206 06:23:09.916359 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3178492e382f3c8585308c4ee39f1caea0bb67c6f97da76aea6237e1669dc416" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 06 06:23:09 crc kubenswrapper[4809]: E1206 06:23:09.916432 4809 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-6d8ccd7c9f-nbxnr" podUID="04d8a790-dedd-44c4-9b8f-093abb541e73" containerName="heat-engine" Dec 06 06:23:11 crc kubenswrapper[4809]: I1206 06:23:11.411789 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d76c911-21ec-4f29-9707-e8101ed3aed9" path="/var/lib/kubelet/pods/4d76c911-21ec-4f29-9707-e8101ed3aed9/volumes" Dec 06 06:23:12 crc kubenswrapper[4809]: I1206 06:23:12.041211 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:23:17 crc kubenswrapper[4809]: I1206 06:23:17.899853 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 06 06:23:19 crc kubenswrapper[4809]: E1206 06:23:19.884426 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3178492e382f3c8585308c4ee39f1caea0bb67c6f97da76aea6237e1669dc416" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 06 06:23:19 crc kubenswrapper[4809]: E1206 06:23:19.885767 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3178492e382f3c8585308c4ee39f1caea0bb67c6f97da76aea6237e1669dc416" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 06 06:23:19 crc kubenswrapper[4809]: E1206 06:23:19.886749 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3178492e382f3c8585308c4ee39f1caea0bb67c6f97da76aea6237e1669dc416" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 06 06:23:19 crc kubenswrapper[4809]: E1206 06:23:19.886810 4809 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-6d8ccd7c9f-nbxnr" podUID="04d8a790-dedd-44c4-9b8f-093abb541e73" containerName="heat-engine" Dec 06 06:23:21 crc kubenswrapper[4809]: I1206 06:23:21.105006 4809 scope.go:117] "RemoveContainer" containerID="d7f6127f423b0d477bde7856693ff15b1c9c16bf8a4fec6fd9c275df64f72120" Dec 06 06:23:22 crc kubenswrapper[4809]: I1206 06:23:22.389273 4809 scope.go:117] "RemoveContainer" containerID="9bc2dc88b3b9f309ce0169aa510326dd77022dece4c69cab3141df00fe5c0106" Dec 06 06:23:22 crc kubenswrapper[4809]: E1206 06:23:22.390765 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:23:25 crc kubenswrapper[4809]: I1206 06:23:25.772566 4809 generic.go:334] "Generic (PLEG): container finished" podID="04d8a790-dedd-44c4-9b8f-093abb541e73" containerID="3178492e382f3c8585308c4ee39f1caea0bb67c6f97da76aea6237e1669dc416" exitCode=0 Dec 06 06:23:25 crc kubenswrapper[4809]: I1206 06:23:25.772671 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-6d8ccd7c9f-nbxnr" event={"ID":"04d8a790-dedd-44c4-9b8f-093abb541e73","Type":"ContainerDied","Data":"3178492e382f3c8585308c4ee39f1caea0bb67c6f97da76aea6237e1669dc416"} Dec 06 06:23:26 crc kubenswrapper[4809]: I1206 06:23:26.765514 4809 scope.go:117] "RemoveContainer" containerID="30b14dfc79cb4d967b8537d39fe468fc1409477b840153c0872891dfe7ab084b" Dec 06 06:23:26 crc kubenswrapper[4809]: E1206 06:23:26.778399 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-aodh-api:current-tested" Dec 06 06:23:26 crc kubenswrapper[4809]: E1206 06:23:26.778453 4809 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-aodh-api:current-tested" Dec 06 06:23:26 crc kubenswrapper[4809]: E1206 06:23:26.778612 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:aodh-db-sync,Image:quay.rdoproject.org/podified-master-centos10/openstack-aodh-api:current-tested,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:AodhPassword,Value:,ValueFrom:&EnvVarSource{FieldRef:nil,ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:&SecretKeySelector{LocalObjectReference:LocalObjectReference{Name:osp-secret,},Key:AodhPassword,Optional:nil,},},},EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:aodh-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5gh2v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42402,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod aodh-db-sync-q7rkl_openstack(a75c2789-4b84-455c-9215-2c52308dbace): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 06:23:26 crc kubenswrapper[4809]: E1206 06:23:26.780208 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"aodh-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/aodh-db-sync-q7rkl" podUID="a75c2789-4b84-455c-9215-2c52308dbace" Dec 06 06:23:26 crc kubenswrapper[4809]: I1206 06:23:26.818164 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 06:23:27 crc kubenswrapper[4809]: I1206 06:23:27.004129 4809 scope.go:117] "RemoveContainer" containerID="74db0c60427064c172da346ab23eb30a4c9dae29ee5e6f33e04e8dee7cfde712" Dec 06 06:23:27 crc kubenswrapper[4809]: I1206 06:23:27.220302 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-6d8ccd7c9f-nbxnr" Dec 06 06:23:27 crc kubenswrapper[4809]: I1206 06:23:27.369252 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04d8a790-dedd-44c4-9b8f-093abb541e73-config-data\") pod \"04d8a790-dedd-44c4-9b8f-093abb541e73\" (UID: \"04d8a790-dedd-44c4-9b8f-093abb541e73\") " Dec 06 06:23:27 crc kubenswrapper[4809]: I1206 06:23:27.369595 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbw2p\" (UniqueName: \"kubernetes.io/projected/04d8a790-dedd-44c4-9b8f-093abb541e73-kube-api-access-bbw2p\") pod \"04d8a790-dedd-44c4-9b8f-093abb541e73\" (UID: \"04d8a790-dedd-44c4-9b8f-093abb541e73\") " Dec 06 06:23:27 crc kubenswrapper[4809]: I1206 06:23:27.369792 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04d8a790-dedd-44c4-9b8f-093abb541e73-combined-ca-bundle\") pod \"04d8a790-dedd-44c4-9b8f-093abb541e73\" (UID: \"04d8a790-dedd-44c4-9b8f-093abb541e73\") " Dec 06 06:23:27 crc kubenswrapper[4809]: I1206 06:23:27.369969 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/04d8a790-dedd-44c4-9b8f-093abb541e73-config-data-custom\") pod \"04d8a790-dedd-44c4-9b8f-093abb541e73\" (UID: \"04d8a790-dedd-44c4-9b8f-093abb541e73\") " Dec 06 06:23:27 crc kubenswrapper[4809]: I1206 06:23:27.376166 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04d8a790-dedd-44c4-9b8f-093abb541e73-kube-api-access-bbw2p" (OuterVolumeSpecName: "kube-api-access-bbw2p") pod "04d8a790-dedd-44c4-9b8f-093abb541e73" (UID: "04d8a790-dedd-44c4-9b8f-093abb541e73"). InnerVolumeSpecName "kube-api-access-bbw2p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:23:27 crc kubenswrapper[4809]: I1206 06:23:27.377493 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04d8a790-dedd-44c4-9b8f-093abb541e73-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "04d8a790-dedd-44c4-9b8f-093abb541e73" (UID: "04d8a790-dedd-44c4-9b8f-093abb541e73"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:23:27 crc kubenswrapper[4809]: I1206 06:23:27.411305 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04d8a790-dedd-44c4-9b8f-093abb541e73-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "04d8a790-dedd-44c4-9b8f-093abb541e73" (UID: "04d8a790-dedd-44c4-9b8f-093abb541e73"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:23:27 crc kubenswrapper[4809]: I1206 06:23:27.459352 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04d8a790-dedd-44c4-9b8f-093abb541e73-config-data" (OuterVolumeSpecName: "config-data") pod "04d8a790-dedd-44c4-9b8f-093abb541e73" (UID: "04d8a790-dedd-44c4-9b8f-093abb541e73"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:23:27 crc kubenswrapper[4809]: I1206 06:23:27.475833 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04d8a790-dedd-44c4-9b8f-093abb541e73-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:23:27 crc kubenswrapper[4809]: I1206 06:23:27.475867 4809 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/04d8a790-dedd-44c4-9b8f-093abb541e73-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 06:23:27 crc kubenswrapper[4809]: I1206 06:23:27.475878 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04d8a790-dedd-44c4-9b8f-093abb541e73-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:23:27 crc kubenswrapper[4809]: I1206 06:23:27.475887 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbw2p\" (UniqueName: \"kubernetes.io/projected/04d8a790-dedd-44c4-9b8f-093abb541e73-kube-api-access-bbw2p\") on node \"crc\" DevicePath \"\"" Dec 06 06:23:27 crc kubenswrapper[4809]: I1206 06:23:27.816380 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-6d8ccd7c9f-nbxnr" event={"ID":"04d8a790-dedd-44c4-9b8f-093abb541e73","Type":"ContainerDied","Data":"c8a90f915d63b2ce34446edce039cffc57b5a7aa443396df4dc86065c69b389d"} Dec 06 06:23:27 crc kubenswrapper[4809]: I1206 06:23:27.816408 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-6d8ccd7c9f-nbxnr" Dec 06 06:23:27 crc kubenswrapper[4809]: I1206 06:23:27.816435 4809 scope.go:117] "RemoveContainer" containerID="3178492e382f3c8585308c4ee39f1caea0bb67c6f97da76aea6237e1669dc416" Dec 06 06:23:27 crc kubenswrapper[4809]: I1206 06:23:27.822892 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-s5cs4" event={"ID":"55c8fbdf-1083-4893-9e7c-76f98c35d656","Type":"ContainerStarted","Data":"177ece144dfeac325702d8bf1015457ca54bca119cb1e84240e05234dcd1b9e9"} Dec 06 06:23:27 crc kubenswrapper[4809]: E1206 06:23:27.824591 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"aodh-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-aodh-api:current-tested\\\"\"" pod="openstack/aodh-db-sync-q7rkl" podUID="a75c2789-4b84-455c-9215-2c52308dbace" Dec 06 06:23:27 crc kubenswrapper[4809]: I1206 06:23:27.846826 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-s5cs4" podStartSLOduration=2.657684723 podStartE2EDuration="27.846807697s" podCreationTimestamp="2025-12-06 06:23:00 +0000 UTC" firstStartedPulling="2025-12-06 06:23:01.62629912 +0000 UTC m=+1906.515282062" lastFinishedPulling="2025-12-06 06:23:26.815422094 +0000 UTC m=+1931.704405036" observedRunningTime="2025-12-06 06:23:27.840086295 +0000 UTC m=+1932.729069247" watchObservedRunningTime="2025-12-06 06:23:27.846807697 +0000 UTC m=+1932.735790639" Dec 06 06:23:27 crc kubenswrapper[4809]: I1206 06:23:27.896455 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-6d8ccd7c9f-nbxnr"] Dec 06 06:23:27 crc kubenswrapper[4809]: I1206 06:23:27.908187 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-6d8ccd7c9f-nbxnr"] Dec 06 06:23:29 crc kubenswrapper[4809]: I1206 06:23:29.404391 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04d8a790-dedd-44c4-9b8f-093abb541e73" path="/var/lib/kubelet/pods/04d8a790-dedd-44c4-9b8f-093abb541e73/volumes" Dec 06 06:23:35 crc kubenswrapper[4809]: I1206 06:23:35.404539 4809 scope.go:117] "RemoveContainer" containerID="9bc2dc88b3b9f309ce0169aa510326dd77022dece4c69cab3141df00fe5c0106" Dec 06 06:23:37 crc kubenswrapper[4809]: I1206 06:23:37.958908 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" event={"ID":"cbf4cf62-024e-4703-a8b8-9aecda9cd26a","Type":"ContainerStarted","Data":"54ba3d8af898b6b995ee9173b17fafab2449b8d6a7b61c005c29dabf843d35fe"} Dec 06 06:23:40 crc kubenswrapper[4809]: I1206 06:23:40.939187 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-z95cs" podUID="a3b6197b-05d4-49b2-9a67-56b675046a1d" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 06:23:44 crc kubenswrapper[4809]: I1206 06:23:44.001864 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 06 06:23:44 crc kubenswrapper[4809]: I1206 06:23:44.067857 4809 generic.go:334] "Generic (PLEG): container finished" podID="55c8fbdf-1083-4893-9e7c-76f98c35d656" containerID="177ece144dfeac325702d8bf1015457ca54bca119cb1e84240e05234dcd1b9e9" exitCode=0 Dec 06 06:23:44 crc kubenswrapper[4809]: I1206 06:23:44.067912 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-s5cs4" event={"ID":"55c8fbdf-1083-4893-9e7c-76f98c35d656","Type":"ContainerDied","Data":"177ece144dfeac325702d8bf1015457ca54bca119cb1e84240e05234dcd1b9e9"} Dec 06 06:23:45 crc kubenswrapper[4809]: I1206 06:23:45.079597 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-q7rkl" event={"ID":"a75c2789-4b84-455c-9215-2c52308dbace","Type":"ContainerStarted","Data":"afcce747e881895f738981fccf8e65088a80ac900f8d433121f09c6ca21f80da"} Dec 06 06:23:45 crc kubenswrapper[4809]: I1206 06:23:45.116120 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-q7rkl" podStartSLOduration=3.105021951 podStartE2EDuration="38.116097072s" podCreationTimestamp="2025-12-06 06:23:07 +0000 UTC" firstStartedPulling="2025-12-06 06:23:08.987089544 +0000 UTC m=+1913.876072486" lastFinishedPulling="2025-12-06 06:23:43.998164655 +0000 UTC m=+1948.887147607" observedRunningTime="2025-12-06 06:23:45.093528794 +0000 UTC m=+1949.982511746" watchObservedRunningTime="2025-12-06 06:23:45.116097072 +0000 UTC m=+1950.005080014" Dec 06 06:23:45 crc kubenswrapper[4809]: I1206 06:23:45.781601 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-s5cs4" Dec 06 06:23:45 crc kubenswrapper[4809]: I1206 06:23:45.941080 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nv47x\" (UniqueName: \"kubernetes.io/projected/55c8fbdf-1083-4893-9e7c-76f98c35d656-kube-api-access-nv47x\") pod \"55c8fbdf-1083-4893-9e7c-76f98c35d656\" (UID: \"55c8fbdf-1083-4893-9e7c-76f98c35d656\") " Dec 06 06:23:45 crc kubenswrapper[4809]: I1206 06:23:45.941202 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/55c8fbdf-1083-4893-9e7c-76f98c35d656-ssh-key\") pod \"55c8fbdf-1083-4893-9e7c-76f98c35d656\" (UID: \"55c8fbdf-1083-4893-9e7c-76f98c35d656\") " Dec 06 06:23:45 crc kubenswrapper[4809]: I1206 06:23:45.941297 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55c8fbdf-1083-4893-9e7c-76f98c35d656-repo-setup-combined-ca-bundle\") pod \"55c8fbdf-1083-4893-9e7c-76f98c35d656\" (UID: \"55c8fbdf-1083-4893-9e7c-76f98c35d656\") " Dec 06 06:23:45 crc kubenswrapper[4809]: I1206 06:23:45.941327 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/55c8fbdf-1083-4893-9e7c-76f98c35d656-inventory\") pod \"55c8fbdf-1083-4893-9e7c-76f98c35d656\" (UID: \"55c8fbdf-1083-4893-9e7c-76f98c35d656\") " Dec 06 06:23:45 crc kubenswrapper[4809]: I1206 06:23:45.949531 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55c8fbdf-1083-4893-9e7c-76f98c35d656-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "55c8fbdf-1083-4893-9e7c-76f98c35d656" (UID: "55c8fbdf-1083-4893-9e7c-76f98c35d656"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:23:45 crc kubenswrapper[4809]: I1206 06:23:45.961275 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55c8fbdf-1083-4893-9e7c-76f98c35d656-kube-api-access-nv47x" (OuterVolumeSpecName: "kube-api-access-nv47x") pod "55c8fbdf-1083-4893-9e7c-76f98c35d656" (UID: "55c8fbdf-1083-4893-9e7c-76f98c35d656"). InnerVolumeSpecName "kube-api-access-nv47x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:23:45 crc kubenswrapper[4809]: I1206 06:23:45.990847 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55c8fbdf-1083-4893-9e7c-76f98c35d656-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "55c8fbdf-1083-4893-9e7c-76f98c35d656" (UID: "55c8fbdf-1083-4893-9e7c-76f98c35d656"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:23:45 crc kubenswrapper[4809]: I1206 06:23:45.996127 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55c8fbdf-1083-4893-9e7c-76f98c35d656-inventory" (OuterVolumeSpecName: "inventory") pod "55c8fbdf-1083-4893-9e7c-76f98c35d656" (UID: "55c8fbdf-1083-4893-9e7c-76f98c35d656"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:23:46 crc kubenswrapper[4809]: I1206 06:23:46.043901 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nv47x\" (UniqueName: \"kubernetes.io/projected/55c8fbdf-1083-4893-9e7c-76f98c35d656-kube-api-access-nv47x\") on node \"crc\" DevicePath \"\"" Dec 06 06:23:46 crc kubenswrapper[4809]: I1206 06:23:46.043963 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/55c8fbdf-1083-4893-9e7c-76f98c35d656-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 06:23:46 crc kubenswrapper[4809]: I1206 06:23:46.043976 4809 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55c8fbdf-1083-4893-9e7c-76f98c35d656-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:23:46 crc kubenswrapper[4809]: I1206 06:23:46.043986 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/55c8fbdf-1083-4893-9e7c-76f98c35d656-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 06:23:46 crc kubenswrapper[4809]: I1206 06:23:46.090772 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-s5cs4" event={"ID":"55c8fbdf-1083-4893-9e7c-76f98c35d656","Type":"ContainerDied","Data":"3792d2fa039bd2f00f8b66a50a9eb27baf97fffcd2197fc49d1e204a3881e112"} Dec 06 06:23:46 crc kubenswrapper[4809]: I1206 06:23:46.090809 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3792d2fa039bd2f00f8b66a50a9eb27baf97fffcd2197fc49d1e204a3881e112" Dec 06 06:23:46 crc kubenswrapper[4809]: I1206 06:23:46.090863 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-s5cs4" Dec 06 06:23:46 crc kubenswrapper[4809]: I1206 06:23:46.184347 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-5rgp6"] Dec 06 06:23:46 crc kubenswrapper[4809]: E1206 06:23:46.184978 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d76c911-21ec-4f29-9707-e8101ed3aed9" containerName="heat-api" Dec 06 06:23:46 crc kubenswrapper[4809]: I1206 06:23:46.185000 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d76c911-21ec-4f29-9707-e8101ed3aed9" containerName="heat-api" Dec 06 06:23:46 crc kubenswrapper[4809]: E1206 06:23:46.185076 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55c8fbdf-1083-4893-9e7c-76f98c35d656" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 06 06:23:46 crc kubenswrapper[4809]: I1206 06:23:46.185086 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="55c8fbdf-1083-4893-9e7c-76f98c35d656" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 06 06:23:46 crc kubenswrapper[4809]: E1206 06:23:46.185104 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04d8a790-dedd-44c4-9b8f-093abb541e73" containerName="heat-engine" Dec 06 06:23:46 crc kubenswrapper[4809]: I1206 06:23:46.185113 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="04d8a790-dedd-44c4-9b8f-093abb541e73" containerName="heat-engine" Dec 06 06:23:46 crc kubenswrapper[4809]: I1206 06:23:46.185468 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="55c8fbdf-1083-4893-9e7c-76f98c35d656" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 06 06:23:46 crc kubenswrapper[4809]: I1206 06:23:46.185511 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="04d8a790-dedd-44c4-9b8f-093abb541e73" containerName="heat-engine" Dec 06 06:23:46 crc kubenswrapper[4809]: I1206 06:23:46.185533 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d76c911-21ec-4f29-9707-e8101ed3aed9" containerName="heat-api" Dec 06 06:23:46 crc kubenswrapper[4809]: I1206 06:23:46.188109 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5rgp6" Dec 06 06:23:46 crc kubenswrapper[4809]: I1206 06:23:46.201129 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-5rgp6"] Dec 06 06:23:46 crc kubenswrapper[4809]: I1206 06:23:46.205570 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 06:23:46 crc kubenswrapper[4809]: I1206 06:23:46.205573 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 06:23:46 crc kubenswrapper[4809]: I1206 06:23:46.206205 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 06:23:46 crc kubenswrapper[4809]: I1206 06:23:46.206433 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-bzlf5" Dec 06 06:23:46 crc kubenswrapper[4809]: I1206 06:23:46.353206 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/362c4175-4f34-4343-9c19-f06e53132f6b-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-5rgp6\" (UID: \"362c4175-4f34-4343-9c19-f06e53132f6b\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5rgp6" Dec 06 06:23:46 crc kubenswrapper[4809]: I1206 06:23:46.353639 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/362c4175-4f34-4343-9c19-f06e53132f6b-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-5rgp6\" (UID: \"362c4175-4f34-4343-9c19-f06e53132f6b\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5rgp6" Dec 06 06:23:46 crc kubenswrapper[4809]: I1206 06:23:46.354384 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mzss\" (UniqueName: \"kubernetes.io/projected/362c4175-4f34-4343-9c19-f06e53132f6b-kube-api-access-4mzss\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-5rgp6\" (UID: \"362c4175-4f34-4343-9c19-f06e53132f6b\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5rgp6" Dec 06 06:23:46 crc kubenswrapper[4809]: I1206 06:23:46.455749 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mzss\" (UniqueName: \"kubernetes.io/projected/362c4175-4f34-4343-9c19-f06e53132f6b-kube-api-access-4mzss\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-5rgp6\" (UID: \"362c4175-4f34-4343-9c19-f06e53132f6b\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5rgp6" Dec 06 06:23:46 crc kubenswrapper[4809]: I1206 06:23:46.455879 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/362c4175-4f34-4343-9c19-f06e53132f6b-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-5rgp6\" (UID: \"362c4175-4f34-4343-9c19-f06e53132f6b\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5rgp6" Dec 06 06:23:46 crc kubenswrapper[4809]: I1206 06:23:46.455915 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/362c4175-4f34-4343-9c19-f06e53132f6b-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-5rgp6\" (UID: \"362c4175-4f34-4343-9c19-f06e53132f6b\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5rgp6" Dec 06 06:23:46 crc kubenswrapper[4809]: I1206 06:23:46.464362 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/362c4175-4f34-4343-9c19-f06e53132f6b-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-5rgp6\" (UID: \"362c4175-4f34-4343-9c19-f06e53132f6b\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5rgp6" Dec 06 06:23:46 crc kubenswrapper[4809]: I1206 06:23:46.465630 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/362c4175-4f34-4343-9c19-f06e53132f6b-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-5rgp6\" (UID: \"362c4175-4f34-4343-9c19-f06e53132f6b\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5rgp6" Dec 06 06:23:46 crc kubenswrapper[4809]: I1206 06:23:46.492033 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mzss\" (UniqueName: \"kubernetes.io/projected/362c4175-4f34-4343-9c19-f06e53132f6b-kube-api-access-4mzss\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-5rgp6\" (UID: \"362c4175-4f34-4343-9c19-f06e53132f6b\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5rgp6" Dec 06 06:23:46 crc kubenswrapper[4809]: I1206 06:23:46.511897 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5rgp6" Dec 06 06:23:47 crc kubenswrapper[4809]: W1206 06:23:47.109550 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod362c4175_4f34_4343_9c19_f06e53132f6b.slice/crio-3f41cdc6ef91364cdbd96e16919d5a34e4de824166175a26b25e3e80e1ac9ae2 WatchSource:0}: Error finding container 3f41cdc6ef91364cdbd96e16919d5a34e4de824166175a26b25e3e80e1ac9ae2: Status 404 returned error can't find the container with id 3f41cdc6ef91364cdbd96e16919d5a34e4de824166175a26b25e3e80e1ac9ae2 Dec 06 06:23:47 crc kubenswrapper[4809]: I1206 06:23:47.110572 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-5rgp6"] Dec 06 06:23:48 crc kubenswrapper[4809]: I1206 06:23:48.121336 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5rgp6" event={"ID":"362c4175-4f34-4343-9c19-f06e53132f6b","Type":"ContainerStarted","Data":"3f41cdc6ef91364cdbd96e16919d5a34e4de824166175a26b25e3e80e1ac9ae2"} Dec 06 06:23:49 crc kubenswrapper[4809]: I1206 06:23:49.138336 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5rgp6" event={"ID":"362c4175-4f34-4343-9c19-f06e53132f6b","Type":"ContainerStarted","Data":"e5ff99d3d9819eaa3a8e9b9cb1f7698e7efc9ab26a66f93c5984f427e857534e"} Dec 06 06:23:49 crc kubenswrapper[4809]: I1206 06:23:49.142725 4809 generic.go:334] "Generic (PLEG): container finished" podID="a75c2789-4b84-455c-9215-2c52308dbace" containerID="afcce747e881895f738981fccf8e65088a80ac900f8d433121f09c6ca21f80da" exitCode=0 Dec 06 06:23:49 crc kubenswrapper[4809]: I1206 06:23:49.142781 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-q7rkl" event={"ID":"a75c2789-4b84-455c-9215-2c52308dbace","Type":"ContainerDied","Data":"afcce747e881895f738981fccf8e65088a80ac900f8d433121f09c6ca21f80da"} Dec 06 06:23:49 crc kubenswrapper[4809]: I1206 06:23:49.169247 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5rgp6" podStartSLOduration=2.018478294 podStartE2EDuration="3.169221337s" podCreationTimestamp="2025-12-06 06:23:46 +0000 UTC" firstStartedPulling="2025-12-06 06:23:47.121087898 +0000 UTC m=+1952.010070850" lastFinishedPulling="2025-12-06 06:23:48.271830951 +0000 UTC m=+1953.160813893" observedRunningTime="2025-12-06 06:23:49.158270361 +0000 UTC m=+1954.047253313" watchObservedRunningTime="2025-12-06 06:23:49.169221337 +0000 UTC m=+1954.058204289" Dec 06 06:23:50 crc kubenswrapper[4809]: I1206 06:23:50.577417 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-q7rkl" Dec 06 06:23:50 crc kubenswrapper[4809]: I1206 06:23:50.684075 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5gh2v\" (UniqueName: \"kubernetes.io/projected/a75c2789-4b84-455c-9215-2c52308dbace-kube-api-access-5gh2v\") pod \"a75c2789-4b84-455c-9215-2c52308dbace\" (UID: \"a75c2789-4b84-455c-9215-2c52308dbace\") " Dec 06 06:23:50 crc kubenswrapper[4809]: I1206 06:23:50.684383 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a75c2789-4b84-455c-9215-2c52308dbace-config-data\") pod \"a75c2789-4b84-455c-9215-2c52308dbace\" (UID: \"a75c2789-4b84-455c-9215-2c52308dbace\") " Dec 06 06:23:50 crc kubenswrapper[4809]: I1206 06:23:50.684639 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a75c2789-4b84-455c-9215-2c52308dbace-scripts\") pod \"a75c2789-4b84-455c-9215-2c52308dbace\" (UID: \"a75c2789-4b84-455c-9215-2c52308dbace\") " Dec 06 06:23:50 crc kubenswrapper[4809]: I1206 06:23:50.684804 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a75c2789-4b84-455c-9215-2c52308dbace-combined-ca-bundle\") pod \"a75c2789-4b84-455c-9215-2c52308dbace\" (UID: \"a75c2789-4b84-455c-9215-2c52308dbace\") " Dec 06 06:23:50 crc kubenswrapper[4809]: I1206 06:23:50.691282 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a75c2789-4b84-455c-9215-2c52308dbace-kube-api-access-5gh2v" (OuterVolumeSpecName: "kube-api-access-5gh2v") pod "a75c2789-4b84-455c-9215-2c52308dbace" (UID: "a75c2789-4b84-455c-9215-2c52308dbace"). InnerVolumeSpecName "kube-api-access-5gh2v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:23:50 crc kubenswrapper[4809]: I1206 06:23:50.691749 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a75c2789-4b84-455c-9215-2c52308dbace-scripts" (OuterVolumeSpecName: "scripts") pod "a75c2789-4b84-455c-9215-2c52308dbace" (UID: "a75c2789-4b84-455c-9215-2c52308dbace"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:23:50 crc kubenswrapper[4809]: I1206 06:23:50.725666 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a75c2789-4b84-455c-9215-2c52308dbace-config-data" (OuterVolumeSpecName: "config-data") pod "a75c2789-4b84-455c-9215-2c52308dbace" (UID: "a75c2789-4b84-455c-9215-2c52308dbace"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:23:50 crc kubenswrapper[4809]: I1206 06:23:50.729152 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a75c2789-4b84-455c-9215-2c52308dbace-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a75c2789-4b84-455c-9215-2c52308dbace" (UID: "a75c2789-4b84-455c-9215-2c52308dbace"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:23:50 crc kubenswrapper[4809]: I1206 06:23:50.787958 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5gh2v\" (UniqueName: \"kubernetes.io/projected/a75c2789-4b84-455c-9215-2c52308dbace-kube-api-access-5gh2v\") on node \"crc\" DevicePath \"\"" Dec 06 06:23:50 crc kubenswrapper[4809]: I1206 06:23:50.788330 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a75c2789-4b84-455c-9215-2c52308dbace-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:23:50 crc kubenswrapper[4809]: I1206 06:23:50.788344 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a75c2789-4b84-455c-9215-2c52308dbace-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:23:50 crc kubenswrapper[4809]: I1206 06:23:50.788357 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a75c2789-4b84-455c-9215-2c52308dbace-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:23:51 crc kubenswrapper[4809]: I1206 06:23:51.165402 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-q7rkl" event={"ID":"a75c2789-4b84-455c-9215-2c52308dbace","Type":"ContainerDied","Data":"1f976f35d3d6d7f7332353c9334d021b963c0e517df2b844162d23e349fd6976"} Dec 06 06:23:51 crc kubenswrapper[4809]: I1206 06:23:51.165439 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1f976f35d3d6d7f7332353c9334d021b963c0e517df2b844162d23e349fd6976" Dec 06 06:23:51 crc kubenswrapper[4809]: I1206 06:23:51.165468 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-q7rkl" Dec 06 06:23:52 crc kubenswrapper[4809]: I1206 06:23:52.188791 4809 generic.go:334] "Generic (PLEG): container finished" podID="362c4175-4f34-4343-9c19-f06e53132f6b" containerID="e5ff99d3d9819eaa3a8e9b9cb1f7698e7efc9ab26a66f93c5984f427e857534e" exitCode=0 Dec 06 06:23:52 crc kubenswrapper[4809]: I1206 06:23:52.188839 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5rgp6" event={"ID":"362c4175-4f34-4343-9c19-f06e53132f6b","Type":"ContainerDied","Data":"e5ff99d3d9819eaa3a8e9b9cb1f7698e7efc9ab26a66f93c5984f427e857534e"} Dec 06 06:23:52 crc kubenswrapper[4809]: I1206 06:23:52.296627 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 06 06:23:52 crc kubenswrapper[4809]: I1206 06:23:52.296978 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="133d9795-adf6-44a2-908b-9790b2b6ed60" containerName="aodh-api" containerID="cri-o://5564afd45ab40683a86fabb94556e733086104d51ea1565ca54f162edeebc0e0" gracePeriod=30 Dec 06 06:23:52 crc kubenswrapper[4809]: I1206 06:23:52.297128 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="133d9795-adf6-44a2-908b-9790b2b6ed60" containerName="aodh-evaluator" containerID="cri-o://05acd4adbf641b564c0376628972d4b4613caec5cf5354ad2ac88856db4f45a8" gracePeriod=30 Dec 06 06:23:52 crc kubenswrapper[4809]: I1206 06:23:52.297170 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="133d9795-adf6-44a2-908b-9790b2b6ed60" containerName="aodh-listener" containerID="cri-o://8121ca8bcd2847fcc69f3adfca90a523aa1dfd3a8953789f71ff030b1f35b5ed" gracePeriod=30 Dec 06 06:23:52 crc kubenswrapper[4809]: I1206 06:23:52.297266 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="133d9795-adf6-44a2-908b-9790b2b6ed60" containerName="aodh-notifier" containerID="cri-o://ce59f044d20e36aae4146ae186fa291d71e4671b5e2b8253ce5c2d95beb99058" gracePeriod=30 Dec 06 06:23:53 crc kubenswrapper[4809]: I1206 06:23:53.210258 4809 generic.go:334] "Generic (PLEG): container finished" podID="133d9795-adf6-44a2-908b-9790b2b6ed60" containerID="05acd4adbf641b564c0376628972d4b4613caec5cf5354ad2ac88856db4f45a8" exitCode=0 Dec 06 06:23:53 crc kubenswrapper[4809]: I1206 06:23:53.210300 4809 generic.go:334] "Generic (PLEG): container finished" podID="133d9795-adf6-44a2-908b-9790b2b6ed60" containerID="5564afd45ab40683a86fabb94556e733086104d51ea1565ca54f162edeebc0e0" exitCode=0 Dec 06 06:23:53 crc kubenswrapper[4809]: I1206 06:23:53.210368 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"133d9795-adf6-44a2-908b-9790b2b6ed60","Type":"ContainerDied","Data":"05acd4adbf641b564c0376628972d4b4613caec5cf5354ad2ac88856db4f45a8"} Dec 06 06:23:53 crc kubenswrapper[4809]: I1206 06:23:53.210436 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"133d9795-adf6-44a2-908b-9790b2b6ed60","Type":"ContainerDied","Data":"5564afd45ab40683a86fabb94556e733086104d51ea1565ca54f162edeebc0e0"} Dec 06 06:23:54 crc kubenswrapper[4809]: I1206 06:23:54.034592 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5rgp6" Dec 06 06:23:54 crc kubenswrapper[4809]: I1206 06:23:54.084465 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/362c4175-4f34-4343-9c19-f06e53132f6b-ssh-key\") pod \"362c4175-4f34-4343-9c19-f06e53132f6b\" (UID: \"362c4175-4f34-4343-9c19-f06e53132f6b\") " Dec 06 06:23:54 crc kubenswrapper[4809]: I1206 06:23:54.084667 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4mzss\" (UniqueName: \"kubernetes.io/projected/362c4175-4f34-4343-9c19-f06e53132f6b-kube-api-access-4mzss\") pod \"362c4175-4f34-4343-9c19-f06e53132f6b\" (UID: \"362c4175-4f34-4343-9c19-f06e53132f6b\") " Dec 06 06:23:54 crc kubenswrapper[4809]: I1206 06:23:54.084854 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/362c4175-4f34-4343-9c19-f06e53132f6b-inventory\") pod \"362c4175-4f34-4343-9c19-f06e53132f6b\" (UID: \"362c4175-4f34-4343-9c19-f06e53132f6b\") " Dec 06 06:23:54 crc kubenswrapper[4809]: I1206 06:23:54.094433 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/362c4175-4f34-4343-9c19-f06e53132f6b-kube-api-access-4mzss" (OuterVolumeSpecName: "kube-api-access-4mzss") pod "362c4175-4f34-4343-9c19-f06e53132f6b" (UID: "362c4175-4f34-4343-9c19-f06e53132f6b"). InnerVolumeSpecName "kube-api-access-4mzss". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:23:54 crc kubenswrapper[4809]: I1206 06:23:54.115495 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/362c4175-4f34-4343-9c19-f06e53132f6b-inventory" (OuterVolumeSpecName: "inventory") pod "362c4175-4f34-4343-9c19-f06e53132f6b" (UID: "362c4175-4f34-4343-9c19-f06e53132f6b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:23:54 crc kubenswrapper[4809]: I1206 06:23:54.120037 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/362c4175-4f34-4343-9c19-f06e53132f6b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "362c4175-4f34-4343-9c19-f06e53132f6b" (UID: "362c4175-4f34-4343-9c19-f06e53132f6b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:23:54 crc kubenswrapper[4809]: I1206 06:23:54.188537 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/362c4175-4f34-4343-9c19-f06e53132f6b-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 06:23:54 crc kubenswrapper[4809]: I1206 06:23:54.188571 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/362c4175-4f34-4343-9c19-f06e53132f6b-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 06:23:54 crc kubenswrapper[4809]: I1206 06:23:54.188585 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4mzss\" (UniqueName: \"kubernetes.io/projected/362c4175-4f34-4343-9c19-f06e53132f6b-kube-api-access-4mzss\") on node \"crc\" DevicePath \"\"" Dec 06 06:23:54 crc kubenswrapper[4809]: I1206 06:23:54.226532 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5rgp6" event={"ID":"362c4175-4f34-4343-9c19-f06e53132f6b","Type":"ContainerDied","Data":"3f41cdc6ef91364cdbd96e16919d5a34e4de824166175a26b25e3e80e1ac9ae2"} Dec 06 06:23:54 crc kubenswrapper[4809]: I1206 06:23:54.226578 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f41cdc6ef91364cdbd96e16919d5a34e4de824166175a26b25e3e80e1ac9ae2" Dec 06 06:23:54 crc kubenswrapper[4809]: I1206 06:23:54.226609 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5rgp6" Dec 06 06:23:54 crc kubenswrapper[4809]: I1206 06:23:54.339148 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4c8nk"] Dec 06 06:23:54 crc kubenswrapper[4809]: E1206 06:23:54.339730 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="362c4175-4f34-4343-9c19-f06e53132f6b" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 06 06:23:54 crc kubenswrapper[4809]: I1206 06:23:54.339746 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="362c4175-4f34-4343-9c19-f06e53132f6b" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 06 06:23:54 crc kubenswrapper[4809]: E1206 06:23:54.339756 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a75c2789-4b84-455c-9215-2c52308dbace" containerName="aodh-db-sync" Dec 06 06:23:54 crc kubenswrapper[4809]: I1206 06:23:54.339762 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a75c2789-4b84-455c-9215-2c52308dbace" containerName="aodh-db-sync" Dec 06 06:23:54 crc kubenswrapper[4809]: I1206 06:23:54.340019 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="a75c2789-4b84-455c-9215-2c52308dbace" containerName="aodh-db-sync" Dec 06 06:23:54 crc kubenswrapper[4809]: I1206 06:23:54.340037 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="362c4175-4f34-4343-9c19-f06e53132f6b" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 06 06:23:54 crc kubenswrapper[4809]: I1206 06:23:54.340864 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4c8nk" Dec 06 06:23:54 crc kubenswrapper[4809]: I1206 06:23:54.345727 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 06:23:54 crc kubenswrapper[4809]: I1206 06:23:54.346060 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 06:23:54 crc kubenswrapper[4809]: I1206 06:23:54.346257 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-bzlf5" Dec 06 06:23:54 crc kubenswrapper[4809]: I1206 06:23:54.346413 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 06:23:54 crc kubenswrapper[4809]: I1206 06:23:54.348954 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4c8nk"] Dec 06 06:23:54 crc kubenswrapper[4809]: I1206 06:23:54.496065 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/13f6bc16-e535-42af-953e-b6b3827b35d1-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4c8nk\" (UID: \"13f6bc16-e535-42af-953e-b6b3827b35d1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4c8nk" Dec 06 06:23:54 crc kubenswrapper[4809]: I1206 06:23:54.496145 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13f6bc16-e535-42af-953e-b6b3827b35d1-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4c8nk\" (UID: \"13f6bc16-e535-42af-953e-b6b3827b35d1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4c8nk" Dec 06 06:23:54 crc kubenswrapper[4809]: I1206 06:23:54.496386 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29z5f\" (UniqueName: \"kubernetes.io/projected/13f6bc16-e535-42af-953e-b6b3827b35d1-kube-api-access-29z5f\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4c8nk\" (UID: \"13f6bc16-e535-42af-953e-b6b3827b35d1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4c8nk" Dec 06 06:23:54 crc kubenswrapper[4809]: I1206 06:23:54.496416 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/13f6bc16-e535-42af-953e-b6b3827b35d1-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4c8nk\" (UID: \"13f6bc16-e535-42af-953e-b6b3827b35d1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4c8nk" Dec 06 06:23:54 crc kubenswrapper[4809]: I1206 06:23:54.598520 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/13f6bc16-e535-42af-953e-b6b3827b35d1-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4c8nk\" (UID: \"13f6bc16-e535-42af-953e-b6b3827b35d1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4c8nk" Dec 06 06:23:54 crc kubenswrapper[4809]: I1206 06:23:54.598644 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13f6bc16-e535-42af-953e-b6b3827b35d1-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4c8nk\" (UID: \"13f6bc16-e535-42af-953e-b6b3827b35d1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4c8nk" Dec 06 06:23:54 crc kubenswrapper[4809]: I1206 06:23:54.598880 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29z5f\" (UniqueName: \"kubernetes.io/projected/13f6bc16-e535-42af-953e-b6b3827b35d1-kube-api-access-29z5f\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4c8nk\" (UID: \"13f6bc16-e535-42af-953e-b6b3827b35d1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4c8nk" Dec 06 06:23:54 crc kubenswrapper[4809]: I1206 06:23:54.598928 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/13f6bc16-e535-42af-953e-b6b3827b35d1-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4c8nk\" (UID: \"13f6bc16-e535-42af-953e-b6b3827b35d1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4c8nk" Dec 06 06:23:54 crc kubenswrapper[4809]: I1206 06:23:54.602878 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13f6bc16-e535-42af-953e-b6b3827b35d1-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4c8nk\" (UID: \"13f6bc16-e535-42af-953e-b6b3827b35d1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4c8nk" Dec 06 06:23:54 crc kubenswrapper[4809]: I1206 06:23:54.605268 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/13f6bc16-e535-42af-953e-b6b3827b35d1-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4c8nk\" (UID: \"13f6bc16-e535-42af-953e-b6b3827b35d1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4c8nk" Dec 06 06:23:54 crc kubenswrapper[4809]: I1206 06:23:54.606780 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/13f6bc16-e535-42af-953e-b6b3827b35d1-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4c8nk\" (UID: \"13f6bc16-e535-42af-953e-b6b3827b35d1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4c8nk" Dec 06 06:23:54 crc kubenswrapper[4809]: I1206 06:23:54.619403 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29z5f\" (UniqueName: \"kubernetes.io/projected/13f6bc16-e535-42af-953e-b6b3827b35d1-kube-api-access-29z5f\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4c8nk\" (UID: \"13f6bc16-e535-42af-953e-b6b3827b35d1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4c8nk" Dec 06 06:23:54 crc kubenswrapper[4809]: I1206 06:23:54.687636 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4c8nk" Dec 06 06:23:55 crc kubenswrapper[4809]: I1206 06:23:55.315885 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4c8nk"] Dec 06 06:23:56 crc kubenswrapper[4809]: I1206 06:23:56.248544 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4c8nk" event={"ID":"13f6bc16-e535-42af-953e-b6b3827b35d1","Type":"ContainerStarted","Data":"8b40d9d44a2b8d0cc14bc3aa415bd4ad612f5e66a23652a4989d089ae2370f94"} Dec 06 06:23:58 crc kubenswrapper[4809]: I1206 06:23:58.275417 4809 generic.go:334] "Generic (PLEG): container finished" podID="133d9795-adf6-44a2-908b-9790b2b6ed60" containerID="8121ca8bcd2847fcc69f3adfca90a523aa1dfd3a8953789f71ff030b1f35b5ed" exitCode=0 Dec 06 06:23:58 crc kubenswrapper[4809]: I1206 06:23:58.275802 4809 generic.go:334] "Generic (PLEG): container finished" podID="133d9795-adf6-44a2-908b-9790b2b6ed60" containerID="ce59f044d20e36aae4146ae186fa291d71e4671b5e2b8253ce5c2d95beb99058" exitCode=0 Dec 06 06:23:58 crc kubenswrapper[4809]: I1206 06:23:58.275626 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"133d9795-adf6-44a2-908b-9790b2b6ed60","Type":"ContainerDied","Data":"8121ca8bcd2847fcc69f3adfca90a523aa1dfd3a8953789f71ff030b1f35b5ed"} Dec 06 06:23:58 crc kubenswrapper[4809]: I1206 06:23:58.275843 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"133d9795-adf6-44a2-908b-9790b2b6ed60","Type":"ContainerDied","Data":"ce59f044d20e36aae4146ae186fa291d71e4671b5e2b8253ce5c2d95beb99058"} Dec 06 06:23:58 crc kubenswrapper[4809]: I1206 06:23:58.798617 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 06 06:23:58 crc kubenswrapper[4809]: I1206 06:23:58.903987 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/133d9795-adf6-44a2-908b-9790b2b6ed60-config-data\") pod \"133d9795-adf6-44a2-908b-9790b2b6ed60\" (UID: \"133d9795-adf6-44a2-908b-9790b2b6ed60\") " Dec 06 06:23:58 crc kubenswrapper[4809]: I1206 06:23:58.904243 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/133d9795-adf6-44a2-908b-9790b2b6ed60-public-tls-certs\") pod \"133d9795-adf6-44a2-908b-9790b2b6ed60\" (UID: \"133d9795-adf6-44a2-908b-9790b2b6ed60\") " Dec 06 06:23:58 crc kubenswrapper[4809]: I1206 06:23:58.904312 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9ksb\" (UniqueName: \"kubernetes.io/projected/133d9795-adf6-44a2-908b-9790b2b6ed60-kube-api-access-j9ksb\") pod \"133d9795-adf6-44a2-908b-9790b2b6ed60\" (UID: \"133d9795-adf6-44a2-908b-9790b2b6ed60\") " Dec 06 06:23:58 crc kubenswrapper[4809]: I1206 06:23:58.904378 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/133d9795-adf6-44a2-908b-9790b2b6ed60-scripts\") pod \"133d9795-adf6-44a2-908b-9790b2b6ed60\" (UID: \"133d9795-adf6-44a2-908b-9790b2b6ed60\") " Dec 06 06:23:58 crc kubenswrapper[4809]: I1206 06:23:58.904518 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/133d9795-adf6-44a2-908b-9790b2b6ed60-combined-ca-bundle\") pod \"133d9795-adf6-44a2-908b-9790b2b6ed60\" (UID: \"133d9795-adf6-44a2-908b-9790b2b6ed60\") " Dec 06 06:23:58 crc kubenswrapper[4809]: I1206 06:23:58.904622 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/133d9795-adf6-44a2-908b-9790b2b6ed60-internal-tls-certs\") pod \"133d9795-adf6-44a2-908b-9790b2b6ed60\" (UID: \"133d9795-adf6-44a2-908b-9790b2b6ed60\") " Dec 06 06:23:58 crc kubenswrapper[4809]: I1206 06:23:58.909559 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/133d9795-adf6-44a2-908b-9790b2b6ed60-scripts" (OuterVolumeSpecName: "scripts") pod "133d9795-adf6-44a2-908b-9790b2b6ed60" (UID: "133d9795-adf6-44a2-908b-9790b2b6ed60"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:23:58 crc kubenswrapper[4809]: I1206 06:23:58.913465 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/133d9795-adf6-44a2-908b-9790b2b6ed60-kube-api-access-j9ksb" (OuterVolumeSpecName: "kube-api-access-j9ksb") pod "133d9795-adf6-44a2-908b-9790b2b6ed60" (UID: "133d9795-adf6-44a2-908b-9790b2b6ed60"). InnerVolumeSpecName "kube-api-access-j9ksb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:23:58 crc kubenswrapper[4809]: I1206 06:23:58.994217 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/133d9795-adf6-44a2-908b-9790b2b6ed60-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "133d9795-adf6-44a2-908b-9790b2b6ed60" (UID: "133d9795-adf6-44a2-908b-9790b2b6ed60"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:23:58 crc kubenswrapper[4809]: I1206 06:23:58.997018 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/133d9795-adf6-44a2-908b-9790b2b6ed60-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "133d9795-adf6-44a2-908b-9790b2b6ed60" (UID: "133d9795-adf6-44a2-908b-9790b2b6ed60"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.007161 4809 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/133d9795-adf6-44a2-908b-9790b2b6ed60-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.007188 4809 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/133d9795-adf6-44a2-908b-9790b2b6ed60-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.007198 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9ksb\" (UniqueName: \"kubernetes.io/projected/133d9795-adf6-44a2-908b-9790b2b6ed60-kube-api-access-j9ksb\") on node \"crc\" DevicePath \"\"" Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.007210 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/133d9795-adf6-44a2-908b-9790b2b6ed60-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.035195 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/133d9795-adf6-44a2-908b-9790b2b6ed60-config-data" (OuterVolumeSpecName: "config-data") pod "133d9795-adf6-44a2-908b-9790b2b6ed60" (UID: "133d9795-adf6-44a2-908b-9790b2b6ed60"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.109730 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/133d9795-adf6-44a2-908b-9790b2b6ed60-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.298339 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4c8nk" event={"ID":"13f6bc16-e535-42af-953e-b6b3827b35d1","Type":"ContainerStarted","Data":"17d61f6d547386398decc45eb8d3f8bb5e620442628bcf05d07c4256076808d4"} Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.305853 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"133d9795-adf6-44a2-908b-9790b2b6ed60","Type":"ContainerDied","Data":"c3ab1b43518b68e6638a3f9d4a602bbd3c914665a7180c79bc44368b20cd3de3"} Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.305888 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.306205 4809 scope.go:117] "RemoveContainer" containerID="8121ca8bcd2847fcc69f3adfca90a523aa1dfd3a8953789f71ff030b1f35b5ed" Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.342322 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4c8nk" podStartSLOduration=2.302412477 podStartE2EDuration="5.342291969s" podCreationTimestamp="2025-12-06 06:23:54 +0000 UTC" firstStartedPulling="2025-12-06 06:23:55.327314504 +0000 UTC m=+1960.216297456" lastFinishedPulling="2025-12-06 06:23:58.367193996 +0000 UTC m=+1963.256176948" observedRunningTime="2025-12-06 06:23:59.318032414 +0000 UTC m=+1964.207015356" watchObservedRunningTime="2025-12-06 06:23:59.342291969 +0000 UTC m=+1964.231274911" Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.349483 4809 scope.go:117] "RemoveContainer" containerID="ce59f044d20e36aae4146ae186fa291d71e4671b5e2b8253ce5c2d95beb99058" Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.390400 4809 scope.go:117] "RemoveContainer" containerID="05acd4adbf641b564c0376628972d4b4613caec5cf5354ad2ac88856db4f45a8" Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.429213 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/133d9795-adf6-44a2-908b-9790b2b6ed60-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "133d9795-adf6-44a2-908b-9790b2b6ed60" (UID: "133d9795-adf6-44a2-908b-9790b2b6ed60"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.533907 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/133d9795-adf6-44a2-908b-9790b2b6ed60-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.551597 4809 scope.go:117] "RemoveContainer" containerID="5564afd45ab40683a86fabb94556e733086104d51ea1565ca54f162edeebc0e0" Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.640338 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.654016 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.671578 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 06 06:23:59 crc kubenswrapper[4809]: E1206 06:23:59.672446 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="133d9795-adf6-44a2-908b-9790b2b6ed60" containerName="aodh-notifier" Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.672517 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="133d9795-adf6-44a2-908b-9790b2b6ed60" containerName="aodh-notifier" Dec 06 06:23:59 crc kubenswrapper[4809]: E1206 06:23:59.672587 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="133d9795-adf6-44a2-908b-9790b2b6ed60" containerName="aodh-listener" Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.672638 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="133d9795-adf6-44a2-908b-9790b2b6ed60" containerName="aodh-listener" Dec 06 06:23:59 crc kubenswrapper[4809]: E1206 06:23:59.672701 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="133d9795-adf6-44a2-908b-9790b2b6ed60" containerName="aodh-evaluator" Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.672749 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="133d9795-adf6-44a2-908b-9790b2b6ed60" containerName="aodh-evaluator" Dec 06 06:23:59 crc kubenswrapper[4809]: E1206 06:23:59.672815 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="133d9795-adf6-44a2-908b-9790b2b6ed60" containerName="aodh-api" Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.672864 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="133d9795-adf6-44a2-908b-9790b2b6ed60" containerName="aodh-api" Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.673253 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="133d9795-adf6-44a2-908b-9790b2b6ed60" containerName="aodh-listener" Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.673334 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="133d9795-adf6-44a2-908b-9790b2b6ed60" containerName="aodh-notifier" Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.673427 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="133d9795-adf6-44a2-908b-9790b2b6ed60" containerName="aodh-evaluator" Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.673489 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="133d9795-adf6-44a2-908b-9790b2b6ed60" containerName="aodh-api" Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.683662 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.687551 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.687641 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.687694 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-f84tp" Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.687972 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.688184 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.690425 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.840209 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e16d0f3-9bd2-4887-84dd-3b8fb35d6451-public-tls-certs\") pod \"aodh-0\" (UID: \"7e16d0f3-9bd2-4887-84dd-3b8fb35d6451\") " pod="openstack/aodh-0" Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.840584 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2slh\" (UniqueName: \"kubernetes.io/projected/7e16d0f3-9bd2-4887-84dd-3b8fb35d6451-kube-api-access-b2slh\") pod \"aodh-0\" (UID: \"7e16d0f3-9bd2-4887-84dd-3b8fb35d6451\") " pod="openstack/aodh-0" Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.840661 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e16d0f3-9bd2-4887-84dd-3b8fb35d6451-combined-ca-bundle\") pod \"aodh-0\" (UID: \"7e16d0f3-9bd2-4887-84dd-3b8fb35d6451\") " pod="openstack/aodh-0" Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.840989 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e16d0f3-9bd2-4887-84dd-3b8fb35d6451-internal-tls-certs\") pod \"aodh-0\" (UID: \"7e16d0f3-9bd2-4887-84dd-3b8fb35d6451\") " pod="openstack/aodh-0" Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.841046 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e16d0f3-9bd2-4887-84dd-3b8fb35d6451-scripts\") pod \"aodh-0\" (UID: \"7e16d0f3-9bd2-4887-84dd-3b8fb35d6451\") " pod="openstack/aodh-0" Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.841100 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e16d0f3-9bd2-4887-84dd-3b8fb35d6451-config-data\") pod \"aodh-0\" (UID: \"7e16d0f3-9bd2-4887-84dd-3b8fb35d6451\") " pod="openstack/aodh-0" Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.943324 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e16d0f3-9bd2-4887-84dd-3b8fb35d6451-scripts\") pod \"aodh-0\" (UID: \"7e16d0f3-9bd2-4887-84dd-3b8fb35d6451\") " pod="openstack/aodh-0" Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.943560 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e16d0f3-9bd2-4887-84dd-3b8fb35d6451-config-data\") pod \"aodh-0\" (UID: \"7e16d0f3-9bd2-4887-84dd-3b8fb35d6451\") " pod="openstack/aodh-0" Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.943667 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e16d0f3-9bd2-4887-84dd-3b8fb35d6451-public-tls-certs\") pod \"aodh-0\" (UID: \"7e16d0f3-9bd2-4887-84dd-3b8fb35d6451\") " pod="openstack/aodh-0" Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.943826 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2slh\" (UniqueName: \"kubernetes.io/projected/7e16d0f3-9bd2-4887-84dd-3b8fb35d6451-kube-api-access-b2slh\") pod \"aodh-0\" (UID: \"7e16d0f3-9bd2-4887-84dd-3b8fb35d6451\") " pod="openstack/aodh-0" Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.943911 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e16d0f3-9bd2-4887-84dd-3b8fb35d6451-combined-ca-bundle\") pod \"aodh-0\" (UID: \"7e16d0f3-9bd2-4887-84dd-3b8fb35d6451\") " pod="openstack/aodh-0" Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.944072 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e16d0f3-9bd2-4887-84dd-3b8fb35d6451-internal-tls-certs\") pod \"aodh-0\" (UID: \"7e16d0f3-9bd2-4887-84dd-3b8fb35d6451\") " pod="openstack/aodh-0" Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.952260 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e16d0f3-9bd2-4887-84dd-3b8fb35d6451-combined-ca-bundle\") pod \"aodh-0\" (UID: \"7e16d0f3-9bd2-4887-84dd-3b8fb35d6451\") " pod="openstack/aodh-0" Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.952470 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e16d0f3-9bd2-4887-84dd-3b8fb35d6451-config-data\") pod \"aodh-0\" (UID: \"7e16d0f3-9bd2-4887-84dd-3b8fb35d6451\") " pod="openstack/aodh-0" Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.952660 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e16d0f3-9bd2-4887-84dd-3b8fb35d6451-scripts\") pod \"aodh-0\" (UID: \"7e16d0f3-9bd2-4887-84dd-3b8fb35d6451\") " pod="openstack/aodh-0" Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.953442 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e16d0f3-9bd2-4887-84dd-3b8fb35d6451-public-tls-certs\") pod \"aodh-0\" (UID: \"7e16d0f3-9bd2-4887-84dd-3b8fb35d6451\") " pod="openstack/aodh-0" Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.954696 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e16d0f3-9bd2-4887-84dd-3b8fb35d6451-internal-tls-certs\") pod \"aodh-0\" (UID: \"7e16d0f3-9bd2-4887-84dd-3b8fb35d6451\") " pod="openstack/aodh-0" Dec 06 06:23:59 crc kubenswrapper[4809]: I1206 06:23:59.966661 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2slh\" (UniqueName: \"kubernetes.io/projected/7e16d0f3-9bd2-4887-84dd-3b8fb35d6451-kube-api-access-b2slh\") pod \"aodh-0\" (UID: \"7e16d0f3-9bd2-4887-84dd-3b8fb35d6451\") " pod="openstack/aodh-0" Dec 06 06:24:00 crc kubenswrapper[4809]: I1206 06:24:00.023556 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 06 06:24:00 crc kubenswrapper[4809]: W1206 06:24:00.579428 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7e16d0f3_9bd2_4887_84dd_3b8fb35d6451.slice/crio-ea271bca56c178064b00351dd46b2980a782df8a379408728bb8ddf09fa20284 WatchSource:0}: Error finding container ea271bca56c178064b00351dd46b2980a782df8a379408728bb8ddf09fa20284: Status 404 returned error can't find the container with id ea271bca56c178064b00351dd46b2980a782df8a379408728bb8ddf09fa20284 Dec 06 06:24:00 crc kubenswrapper[4809]: I1206 06:24:00.583908 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 06 06:24:01 crc kubenswrapper[4809]: I1206 06:24:01.332853 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"7e16d0f3-9bd2-4887-84dd-3b8fb35d6451","Type":"ContainerStarted","Data":"2b48b716e98941fb5ea940fd8c712b9e1c6438a99419abe2738ec3bd8563b4e7"} Dec 06 06:24:01 crc kubenswrapper[4809]: I1206 06:24:01.333255 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"7e16d0f3-9bd2-4887-84dd-3b8fb35d6451","Type":"ContainerStarted","Data":"ea271bca56c178064b00351dd46b2980a782df8a379408728bb8ddf09fa20284"} Dec 06 06:24:01 crc kubenswrapper[4809]: I1206 06:24:01.408393 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="133d9795-adf6-44a2-908b-9790b2b6ed60" path="/var/lib/kubelet/pods/133d9795-adf6-44a2-908b-9790b2b6ed60/volumes" Dec 06 06:24:07 crc kubenswrapper[4809]: I1206 06:24:07.947415 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/community-operators-4p5rc" podUID="6fa58271-edf3-4d3d-8782-83a959a691fe" containerName="registry-server" probeResult="failure" output=< Dec 06 06:24:07 crc kubenswrapper[4809]: timeout: health rpc did not complete within 1s Dec 06 06:24:07 crc kubenswrapper[4809]: > Dec 06 06:24:08 crc kubenswrapper[4809]: I1206 06:24:08.394516 4809 patch_prober.go:28] interesting pod/logging-loki-gateway-65498c4f8f-9s8qq container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.54:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 06:24:08 crc kubenswrapper[4809]: I1206 06:24:08.394844 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq" podUID="ef8f3e7e-57ea-42b8-a777-1778e5ed975b" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.54:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 06:24:08 crc kubenswrapper[4809]: I1206 06:24:08.405098 4809 patch_prober.go:28] interesting pod/route-controller-manager-6cd65596c4-9z8ww container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.60:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 06:24:08 crc kubenswrapper[4809]: I1206 06:24:08.405314 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6cd65596c4-9z8ww" podUID="8e47b3dc-260e-46a6-9ce2-2e4f4a8a5977" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.60:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 06:24:14 crc kubenswrapper[4809]: I1206 06:24:14.507649 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"7e16d0f3-9bd2-4887-84dd-3b8fb35d6451","Type":"ContainerStarted","Data":"c278a1550468de7ca92beaafe6c1e7634bfc0e49a2a6542e2f2f0201f02ebaf0"} Dec 06 06:24:20 crc kubenswrapper[4809]: I1206 06:24:20.600605 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"7e16d0f3-9bd2-4887-84dd-3b8fb35d6451","Type":"ContainerStarted","Data":"13309a48d171b7c76c4a36b69a6537cc7a9ce71f204896869620caca5133325d"} Dec 06 06:24:21 crc kubenswrapper[4809]: I1206 06:24:21.617289 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"7e16d0f3-9bd2-4887-84dd-3b8fb35d6451","Type":"ContainerStarted","Data":"2247cdaff9103c75cf372fb65460f7927ecf8a42ecb17407bae42b2c98a3fe29"} Dec 06 06:24:21 crc kubenswrapper[4809]: I1206 06:24:21.655324 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.695713686 podStartE2EDuration="22.65530022s" podCreationTimestamp="2025-12-06 06:23:59 +0000 UTC" firstStartedPulling="2025-12-06 06:24:00.582677891 +0000 UTC m=+1965.471660833" lastFinishedPulling="2025-12-06 06:24:20.542264425 +0000 UTC m=+1985.431247367" observedRunningTime="2025-12-06 06:24:21.646376749 +0000 UTC m=+1986.535359691" watchObservedRunningTime="2025-12-06 06:24:21.65530022 +0000 UTC m=+1986.544283172" Dec 06 06:24:27 crc kubenswrapper[4809]: I1206 06:24:27.408491 4809 scope.go:117] "RemoveContainer" containerID="6db290f4a2b99fba5d0365f55cdced03d05a875cf48e355dd49d6d4a42fbb0e4" Dec 06 06:24:27 crc kubenswrapper[4809]: I1206 06:24:27.552551 4809 scope.go:117] "RemoveContainer" containerID="1a542df6256ee56b5fbdc3ca23fb895140362d7cb61722388669e994937cd78d" Dec 06 06:24:27 crc kubenswrapper[4809]: I1206 06:24:27.599073 4809 scope.go:117] "RemoveContainer" containerID="8b94911ee4fe5e9d341dbaf959bfe7aba30b4a12c818bd27741e70e6a6818015" Dec 06 06:24:27 crc kubenswrapper[4809]: I1206 06:24:27.670971 4809 scope.go:117] "RemoveContainer" containerID="c4906c28036b8e2bd8d1e7626a169b8573e55ab8f1c5a01c0506facd86f74b6f" Dec 06 06:25:04 crc kubenswrapper[4809]: I1206 06:25:04.441698 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kth48"] Dec 06 06:25:04 crc kubenswrapper[4809]: I1206 06:25:04.449202 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kth48" Dec 06 06:25:04 crc kubenswrapper[4809]: I1206 06:25:04.459017 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kth48"] Dec 06 06:25:04 crc kubenswrapper[4809]: I1206 06:25:04.567798 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbxc6\" (UniqueName: \"kubernetes.io/projected/b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2-kube-api-access-mbxc6\") pod \"certified-operators-kth48\" (UID: \"b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2\") " pod="openshift-marketplace/certified-operators-kth48" Dec 06 06:25:04 crc kubenswrapper[4809]: I1206 06:25:04.568148 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2-catalog-content\") pod \"certified-operators-kth48\" (UID: \"b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2\") " pod="openshift-marketplace/certified-operators-kth48" Dec 06 06:25:04 crc kubenswrapper[4809]: I1206 06:25:04.568237 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2-utilities\") pod \"certified-operators-kth48\" (UID: \"b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2\") " pod="openshift-marketplace/certified-operators-kth48" Dec 06 06:25:04 crc kubenswrapper[4809]: I1206 06:25:04.670089 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2-catalog-content\") pod \"certified-operators-kth48\" (UID: \"b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2\") " pod="openshift-marketplace/certified-operators-kth48" Dec 06 06:25:04 crc kubenswrapper[4809]: I1206 06:25:04.670166 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2-utilities\") pod \"certified-operators-kth48\" (UID: \"b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2\") " pod="openshift-marketplace/certified-operators-kth48" Dec 06 06:25:04 crc kubenswrapper[4809]: I1206 06:25:04.670293 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbxc6\" (UniqueName: \"kubernetes.io/projected/b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2-kube-api-access-mbxc6\") pod \"certified-operators-kth48\" (UID: \"b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2\") " pod="openshift-marketplace/certified-operators-kth48" Dec 06 06:25:04 crc kubenswrapper[4809]: I1206 06:25:04.670676 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2-utilities\") pod \"certified-operators-kth48\" (UID: \"b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2\") " pod="openshift-marketplace/certified-operators-kth48" Dec 06 06:25:04 crc kubenswrapper[4809]: I1206 06:25:04.670774 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2-catalog-content\") pod \"certified-operators-kth48\" (UID: \"b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2\") " pod="openshift-marketplace/certified-operators-kth48" Dec 06 06:25:04 crc kubenswrapper[4809]: I1206 06:25:04.692831 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbxc6\" (UniqueName: \"kubernetes.io/projected/b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2-kube-api-access-mbxc6\") pod \"certified-operators-kth48\" (UID: \"b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2\") " pod="openshift-marketplace/certified-operators-kth48" Dec 06 06:25:04 crc kubenswrapper[4809]: I1206 06:25:04.793544 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kth48" Dec 06 06:25:05 crc kubenswrapper[4809]: I1206 06:25:05.349744 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kth48"] Dec 06 06:25:06 crc kubenswrapper[4809]: I1206 06:25:06.185668 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kth48" event={"ID":"b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2","Type":"ContainerStarted","Data":"78704de0806ce1c61dbeecd7680c97598a5596abdbd18f75c2545d04b0516a37"} Dec 06 06:25:07 crc kubenswrapper[4809]: I1206 06:25:07.197977 4809 generic.go:334] "Generic (PLEG): container finished" podID="b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2" containerID="5a761d93ef4023d6a8f1aff0ac7aacb4f222ad1ae8d7014a514fb541b94eb75c" exitCode=0 Dec 06 06:25:07 crc kubenswrapper[4809]: I1206 06:25:07.198052 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kth48" event={"ID":"b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2","Type":"ContainerDied","Data":"5a761d93ef4023d6a8f1aff0ac7aacb4f222ad1ae8d7014a514fb541b94eb75c"} Dec 06 06:25:09 crc kubenswrapper[4809]: I1206 06:25:09.227846 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kth48" event={"ID":"b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2","Type":"ContainerStarted","Data":"1c30484f54c1506391371b7860a5e6bd84e7ecb6d1a59a579217aed8e18fd7db"} Dec 06 06:25:12 crc kubenswrapper[4809]: I1206 06:25:12.268051 4809 generic.go:334] "Generic (PLEG): container finished" podID="b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2" containerID="1c30484f54c1506391371b7860a5e6bd84e7ecb6d1a59a579217aed8e18fd7db" exitCode=0 Dec 06 06:25:12 crc kubenswrapper[4809]: I1206 06:25:12.268128 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kth48" event={"ID":"b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2","Type":"ContainerDied","Data":"1c30484f54c1506391371b7860a5e6bd84e7ecb6d1a59a579217aed8e18fd7db"} Dec 06 06:25:12 crc kubenswrapper[4809]: I1206 06:25:12.271607 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 06:25:13 crc kubenswrapper[4809]: I1206 06:25:13.291376 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kth48" event={"ID":"b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2","Type":"ContainerStarted","Data":"b4e56e00c58894cde7141cbdf58ce75924f9471af818d37e164cefd772ddb6e0"} Dec 06 06:25:13 crc kubenswrapper[4809]: I1206 06:25:13.333386 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kth48" podStartSLOduration=3.857120811 podStartE2EDuration="9.333366198s" podCreationTimestamp="2025-12-06 06:25:04 +0000 UTC" firstStartedPulling="2025-12-06 06:25:07.201100947 +0000 UTC m=+2032.090083889" lastFinishedPulling="2025-12-06 06:25:12.677346344 +0000 UTC m=+2037.566329276" observedRunningTime="2025-12-06 06:25:13.326257916 +0000 UTC m=+2038.215240858" watchObservedRunningTime="2025-12-06 06:25:13.333366198 +0000 UTC m=+2038.222349140" Dec 06 06:25:14 crc kubenswrapper[4809]: I1206 06:25:14.794563 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kth48" Dec 06 06:25:14 crc kubenswrapper[4809]: I1206 06:25:14.794825 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kth48" Dec 06 06:25:14 crc kubenswrapper[4809]: I1206 06:25:14.878361 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kth48" Dec 06 06:25:24 crc kubenswrapper[4809]: I1206 06:25:24.843993 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kth48" Dec 06 06:25:24 crc kubenswrapper[4809]: I1206 06:25:24.897349 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kth48"] Dec 06 06:25:25 crc kubenswrapper[4809]: I1206 06:25:25.473421 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kth48" podUID="b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2" containerName="registry-server" containerID="cri-o://b4e56e00c58894cde7141cbdf58ce75924f9471af818d37e164cefd772ddb6e0" gracePeriod=2 Dec 06 06:25:26 crc kubenswrapper[4809]: I1206 06:25:26.488544 4809 generic.go:334] "Generic (PLEG): container finished" podID="b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2" containerID="b4e56e00c58894cde7141cbdf58ce75924f9471af818d37e164cefd772ddb6e0" exitCode=0 Dec 06 06:25:26 crc kubenswrapper[4809]: I1206 06:25:26.488630 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kth48" event={"ID":"b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2","Type":"ContainerDied","Data":"b4e56e00c58894cde7141cbdf58ce75924f9471af818d37e164cefd772ddb6e0"} Dec 06 06:25:26 crc kubenswrapper[4809]: I1206 06:25:26.488665 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kth48" event={"ID":"b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2","Type":"ContainerDied","Data":"78704de0806ce1c61dbeecd7680c97598a5596abdbd18f75c2545d04b0516a37"} Dec 06 06:25:26 crc kubenswrapper[4809]: I1206 06:25:26.488680 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="78704de0806ce1c61dbeecd7680c97598a5596abdbd18f75c2545d04b0516a37" Dec 06 06:25:26 crc kubenswrapper[4809]: I1206 06:25:26.604714 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kth48" Dec 06 06:25:26 crc kubenswrapper[4809]: I1206 06:25:26.745533 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mbxc6\" (UniqueName: \"kubernetes.io/projected/b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2-kube-api-access-mbxc6\") pod \"b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2\" (UID: \"b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2\") " Dec 06 06:25:26 crc kubenswrapper[4809]: I1206 06:25:26.746835 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2-catalog-content\") pod \"b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2\" (UID: \"b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2\") " Dec 06 06:25:26 crc kubenswrapper[4809]: I1206 06:25:26.747200 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2-utilities\") pod \"b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2\" (UID: \"b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2\") " Dec 06 06:25:26 crc kubenswrapper[4809]: I1206 06:25:26.748792 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2-utilities" (OuterVolumeSpecName: "utilities") pod "b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2" (UID: "b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:25:26 crc kubenswrapper[4809]: I1206 06:25:26.755367 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2-kube-api-access-mbxc6" (OuterVolumeSpecName: "kube-api-access-mbxc6") pod "b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2" (UID: "b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2"). InnerVolumeSpecName "kube-api-access-mbxc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:25:26 crc kubenswrapper[4809]: I1206 06:25:26.807712 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2" (UID: "b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:25:26 crc kubenswrapper[4809]: I1206 06:25:26.850258 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:25:26 crc kubenswrapper[4809]: I1206 06:25:26.850290 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:25:26 crc kubenswrapper[4809]: I1206 06:25:26.850302 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mbxc6\" (UniqueName: \"kubernetes.io/projected/b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2-kube-api-access-mbxc6\") on node \"crc\" DevicePath \"\"" Dec 06 06:25:27 crc kubenswrapper[4809]: I1206 06:25:27.499678 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kth48" Dec 06 06:25:27 crc kubenswrapper[4809]: I1206 06:25:27.536492 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kth48"] Dec 06 06:25:27 crc kubenswrapper[4809]: I1206 06:25:27.547904 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kth48"] Dec 06 06:25:27 crc kubenswrapper[4809]: I1206 06:25:27.842424 4809 scope.go:117] "RemoveContainer" containerID="2eaf3473d57f772f14f96c219113e3f7b1dc1bf47459cd506632caa4b1b48dc1" Dec 06 06:25:27 crc kubenswrapper[4809]: I1206 06:25:27.870147 4809 scope.go:117] "RemoveContainer" containerID="7acc7ef16f31e7c4b28b16816753e06fe96f5a8c7956bb11f29cce88ea77c9e9" Dec 06 06:25:27 crc kubenswrapper[4809]: I1206 06:25:27.899558 4809 scope.go:117] "RemoveContainer" containerID="81af9828c93bb80ee07a5d7ec2bdb431c342fcf222dafec261f249d39e43a7e6" Dec 06 06:25:27 crc kubenswrapper[4809]: I1206 06:25:27.926331 4809 scope.go:117] "RemoveContainer" containerID="214b66e1c2371010ad89f60894fff66ea2452e3abb99cdccc6dca247c7d5028e" Dec 06 06:25:29 crc kubenswrapper[4809]: I1206 06:25:29.406759 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2" path="/var/lib/kubelet/pods/b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2/volumes" Dec 06 06:25:41 crc kubenswrapper[4809]: I1206 06:25:41.057414 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-rksl7"] Dec 06 06:25:41 crc kubenswrapper[4809]: I1206 06:25:41.069342 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-rksl7"] Dec 06 06:25:41 crc kubenswrapper[4809]: I1206 06:25:41.404188 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc6d16d3-5a2f-4d15-ba47-a25e4608703b" path="/var/lib/kubelet/pods/cc6d16d3-5a2f-4d15-ba47-a25e4608703b/volumes" Dec 06 06:25:43 crc kubenswrapper[4809]: I1206 06:25:43.033517 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-qz5fx"] Dec 06 06:25:43 crc kubenswrapper[4809]: I1206 06:25:43.044900 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-qz5fx"] Dec 06 06:25:43 crc kubenswrapper[4809]: I1206 06:25:43.055808 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-c863-account-create-update-29vzq"] Dec 06 06:25:43 crc kubenswrapper[4809]: I1206 06:25:43.065415 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-8cd6q"] Dec 06 06:25:43 crc kubenswrapper[4809]: I1206 06:25:43.076090 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-99dc-account-create-update-qf55d"] Dec 06 06:25:43 crc kubenswrapper[4809]: I1206 06:25:43.086480 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-8cd6q"] Dec 06 06:25:43 crc kubenswrapper[4809]: I1206 06:25:43.098479 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-c863-account-create-update-29vzq"] Dec 06 06:25:43 crc kubenswrapper[4809]: I1206 06:25:43.110474 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-99dc-account-create-update-qf55d"] Dec 06 06:25:43 crc kubenswrapper[4809]: I1206 06:25:43.407860 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c342dde-c9c4-4ead-a4a5-6db383219a32" path="/var/lib/kubelet/pods/1c342dde-c9c4-4ead-a4a5-6db383219a32/volumes" Dec 06 06:25:43 crc kubenswrapper[4809]: I1206 06:25:43.410579 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2afacc78-69cc-4fd2-b857-60a53a29c650" path="/var/lib/kubelet/pods/2afacc78-69cc-4fd2-b857-60a53a29c650/volumes" Dec 06 06:25:43 crc kubenswrapper[4809]: I1206 06:25:43.412311 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69e67ebc-dcc4-4f25-9366-7d8a23256d58" path="/var/lib/kubelet/pods/69e67ebc-dcc4-4f25-9366-7d8a23256d58/volumes" Dec 06 06:25:43 crc kubenswrapper[4809]: I1206 06:25:43.413557 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b5b8648-fb4d-4b44-9e6d-0fc714e55add" path="/var/lib/kubelet/pods/6b5b8648-fb4d-4b44-9e6d-0fc714e55add/volumes" Dec 06 06:25:44 crc kubenswrapper[4809]: I1206 06:25:44.077307 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-e6aa-account-create-update-pz4lp"] Dec 06 06:25:44 crc kubenswrapper[4809]: I1206 06:25:44.089993 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-xrxv5"] Dec 06 06:25:44 crc kubenswrapper[4809]: I1206 06:25:44.100755 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-d99d-account-create-update-l7vkv"] Dec 06 06:25:44 crc kubenswrapper[4809]: I1206 06:25:44.110140 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-xrxv5"] Dec 06 06:25:44 crc kubenswrapper[4809]: I1206 06:25:44.119172 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-e6aa-account-create-update-pz4lp"] Dec 06 06:25:44 crc kubenswrapper[4809]: I1206 06:25:44.127676 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-d99d-account-create-update-l7vkv"] Dec 06 06:25:45 crc kubenswrapper[4809]: I1206 06:25:45.401688 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53bf9e4c-33f6-4069-b6f1-8f99be5dbe67" path="/var/lib/kubelet/pods/53bf9e4c-33f6-4069-b6f1-8f99be5dbe67/volumes" Dec 06 06:25:45 crc kubenswrapper[4809]: I1206 06:25:45.402480 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa09ddab-61ad-4c63-af1c-f974502055f1" path="/var/lib/kubelet/pods/aa09ddab-61ad-4c63-af1c-f974502055f1/volumes" Dec 06 06:25:45 crc kubenswrapper[4809]: I1206 06:25:45.403285 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd020248-5e26-4df3-8a57-420613ffc902" path="/var/lib/kubelet/pods/fd020248-5e26-4df3-8a57-420613ffc902/volumes" Dec 06 06:25:49 crc kubenswrapper[4809]: I1206 06:25:49.039707 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-cj7v2"] Dec 06 06:25:49 crc kubenswrapper[4809]: I1206 06:25:49.056053 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-e068-account-create-update-8bklf"] Dec 06 06:25:49 crc kubenswrapper[4809]: I1206 06:25:49.068910 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-cj7v2"] Dec 06 06:25:49 crc kubenswrapper[4809]: I1206 06:25:49.080815 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-e068-account-create-update-8bklf"] Dec 06 06:25:49 crc kubenswrapper[4809]: I1206 06:25:49.402762 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bdf1730f-cf76-4ea5-b89f-fee3ca2e5533" path="/var/lib/kubelet/pods/bdf1730f-cf76-4ea5-b89f-fee3ca2e5533/volumes" Dec 06 06:25:49 crc kubenswrapper[4809]: I1206 06:25:49.403730 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fdf6cf2e-69cc-4152-9e07-736e0fc4512d" path="/var/lib/kubelet/pods/fdf6cf2e-69cc-4152-9e07-736e0fc4512d/volumes" Dec 06 06:26:04 crc kubenswrapper[4809]: I1206 06:26:04.496861 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:26:04 crc kubenswrapper[4809]: I1206 06:26:04.497574 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:26:09 crc kubenswrapper[4809]: I1206 06:26:09.049748 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-8f52w"] Dec 06 06:26:09 crc kubenswrapper[4809]: I1206 06:26:09.060514 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-7ec9-account-create-update-llsx9"] Dec 06 06:26:09 crc kubenswrapper[4809]: I1206 06:26:09.072865 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-7ec9-account-create-update-llsx9"] Dec 06 06:26:09 crc kubenswrapper[4809]: I1206 06:26:09.084265 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-8f52w"] Dec 06 06:26:09 crc kubenswrapper[4809]: I1206 06:26:09.409159 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32aaf34e-00f0-4de8-9a07-ead18a18c8a9" path="/var/lib/kubelet/pods/32aaf34e-00f0-4de8-9a07-ead18a18c8a9/volumes" Dec 06 06:26:09 crc kubenswrapper[4809]: I1206 06:26:09.410537 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac97c219-a010-4070-84fe-ae41b3083627" path="/var/lib/kubelet/pods/ac97c219-a010-4070-84fe-ae41b3083627/volumes" Dec 06 06:26:10 crc kubenswrapper[4809]: I1206 06:26:10.046244 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-112c-account-create-update-kskg6"] Dec 06 06:26:10 crc kubenswrapper[4809]: I1206 06:26:10.063781 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-pc5wh"] Dec 06 06:26:10 crc kubenswrapper[4809]: I1206 06:26:10.075183 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-nr9js"] Dec 06 06:26:10 crc kubenswrapper[4809]: I1206 06:26:10.085750 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-8nl86"] Dec 06 06:26:10 crc kubenswrapper[4809]: I1206 06:26:10.097737 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-d911-account-create-update-wsk86"] Dec 06 06:26:10 crc kubenswrapper[4809]: I1206 06:26:10.111993 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-pc5wh"] Dec 06 06:26:10 crc kubenswrapper[4809]: I1206 06:26:10.125966 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-112c-account-create-update-kskg6"] Dec 06 06:26:10 crc kubenswrapper[4809]: I1206 06:26:10.136370 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-nr9js"] Dec 06 06:26:10 crc kubenswrapper[4809]: I1206 06:26:10.147696 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-8nl86"] Dec 06 06:26:10 crc kubenswrapper[4809]: I1206 06:26:10.158507 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-d911-account-create-update-wsk86"] Dec 06 06:26:10 crc kubenswrapper[4809]: I1206 06:26:10.168796 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-fbc7-account-create-update-95k9l"] Dec 06 06:26:10 crc kubenswrapper[4809]: I1206 06:26:10.179817 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-fbc7-account-create-update-95k9l"] Dec 06 06:26:11 crc kubenswrapper[4809]: I1206 06:26:11.403168 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51864703-13bc-484f-abb0-3a59bc5195ed" path="/var/lib/kubelet/pods/51864703-13bc-484f-abb0-3a59bc5195ed/volumes" Dec 06 06:26:11 crc kubenswrapper[4809]: I1206 06:26:11.404217 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5749399f-d0d9-40ba-ae30-448d7311a953" path="/var/lib/kubelet/pods/5749399f-d0d9-40ba-ae30-448d7311a953/volumes" Dec 06 06:26:11 crc kubenswrapper[4809]: I1206 06:26:11.405304 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b837741-4518-43ae-a8a4-befdbcb593b6" path="/var/lib/kubelet/pods/6b837741-4518-43ae-a8a4-befdbcb593b6/volumes" Dec 06 06:26:11 crc kubenswrapper[4809]: I1206 06:26:11.405979 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72b421c4-764d-464e-9e4b-5cbf6cdbd5d5" path="/var/lib/kubelet/pods/72b421c4-764d-464e-9e4b-5cbf6cdbd5d5/volumes" Dec 06 06:26:11 crc kubenswrapper[4809]: I1206 06:26:11.406556 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9138f866-bbfa-45f4-b3b8-a55b783e1c9c" path="/var/lib/kubelet/pods/9138f866-bbfa-45f4-b3b8-a55b783e1c9c/volumes" Dec 06 06:26:11 crc kubenswrapper[4809]: I1206 06:26:11.407735 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9eb5001b-9536-430d-8fae-cea0ecabd5ad" path="/var/lib/kubelet/pods/9eb5001b-9536-430d-8fae-cea0ecabd5ad/volumes" Dec 06 06:26:19 crc kubenswrapper[4809]: I1206 06:26:19.684412 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-bh98g"] Dec 06 06:26:19 crc kubenswrapper[4809]: I1206 06:26:19.696095 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-bh98g"] Dec 06 06:26:21 crc kubenswrapper[4809]: I1206 06:26:21.402344 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0c71ada-5bd2-467e-8d93-cbda68cafcb2" path="/var/lib/kubelet/pods/b0c71ada-5bd2-467e-8d93-cbda68cafcb2/volumes" Dec 06 06:26:27 crc kubenswrapper[4809]: I1206 06:26:27.992173 4809 scope.go:117] "RemoveContainer" containerID="afa603635846d9db85859e61469f8b9bd848799c589de4b7f49b27be5c74bd2e" Dec 06 06:26:28 crc kubenswrapper[4809]: I1206 06:26:28.024866 4809 scope.go:117] "RemoveContainer" containerID="5d6a3110494f958e2436dec41c34e634cee07ccb9796e45f2a9580c2a9604055" Dec 06 06:26:28 crc kubenswrapper[4809]: I1206 06:26:28.076065 4809 scope.go:117] "RemoveContainer" containerID="2138523ae14f02ce6e2b49b458eee3b5e22269e264269047d930c802c93186c8" Dec 06 06:26:28 crc kubenswrapper[4809]: I1206 06:26:28.139509 4809 scope.go:117] "RemoveContainer" containerID="642f19e6ec8cda9ad8da9265a06e7016d564c6600da0eeaaa72682a1f0da5669" Dec 06 06:26:28 crc kubenswrapper[4809]: I1206 06:26:28.212996 4809 scope.go:117] "RemoveContainer" containerID="955742661ebbcc1286b13ee089c7a6dad4bb1892566b14c31a44cbd5b6bb2a04" Dec 06 06:26:28 crc kubenswrapper[4809]: I1206 06:26:28.262314 4809 scope.go:117] "RemoveContainer" containerID="d8104ebb13ec1ef97f7b517163b0dfd873d1d121fae861dfbdb52a22e511eb38" Dec 06 06:26:28 crc kubenswrapper[4809]: I1206 06:26:28.323293 4809 scope.go:117] "RemoveContainer" containerID="c558b3048d490cfd96438f3b8b709314761686af8dce1c3fd2f186fe2ee23899" Dec 06 06:26:28 crc kubenswrapper[4809]: I1206 06:26:28.347509 4809 scope.go:117] "RemoveContainer" containerID="4f3773f4d9b40904f6c35e5546fd01aa5b09918b9be31362ce262e7fc4b010d7" Dec 06 06:26:28 crc kubenswrapper[4809]: I1206 06:26:28.377163 4809 scope.go:117] "RemoveContainer" containerID="e6e0eac6101c3a70a8f725eed6c9a714e4351da79309ead2e22db9074a7c17fd" Dec 06 06:26:28 crc kubenswrapper[4809]: I1206 06:26:28.406700 4809 scope.go:117] "RemoveContainer" containerID="978ea797bd9b2c048c05d45d0e936251240a37ac0ab20db1bb2a1d2bb5c5493d" Dec 06 06:26:28 crc kubenswrapper[4809]: I1206 06:26:28.431958 4809 scope.go:117] "RemoveContainer" containerID="e714667d5d6c52e4a0da5d11fdf1b19a24567dded8567c6053f2f5a1cf78cdc2" Dec 06 06:26:28 crc kubenswrapper[4809]: I1206 06:26:28.453174 4809 scope.go:117] "RemoveContainer" containerID="5518c89e0684238e8a60dbd595ff31fe19ce292ee3631f1c85a374d4bf4bec0a" Dec 06 06:26:28 crc kubenswrapper[4809]: I1206 06:26:28.477403 4809 scope.go:117] "RemoveContainer" containerID="3fd12289e7b2b235cdb57e742b514ec97b2d36c7e4fc9d295eba64e96f21c1a4" Dec 06 06:26:28 crc kubenswrapper[4809]: I1206 06:26:28.500505 4809 scope.go:117] "RemoveContainer" containerID="a4bac04ada57d950a028ed0c265dc011d7f9f458934984fd696ff4b0f3ae6177" Dec 06 06:26:28 crc kubenswrapper[4809]: I1206 06:26:28.526863 4809 scope.go:117] "RemoveContainer" containerID="54466e843436c80dbb0c46cb440bc4569458c697cf5eb5b3e0d61a937450d64e" Dec 06 06:26:28 crc kubenswrapper[4809]: I1206 06:26:28.551683 4809 scope.go:117] "RemoveContainer" containerID="4079a9916fc3a595c2f6cb41013ebe78e1024ab9a5322ccdeaa48049a5d43714" Dec 06 06:26:28 crc kubenswrapper[4809]: I1206 06:26:28.584382 4809 scope.go:117] "RemoveContainer" containerID="8b81316c32d7dad976ae169ecdfdc91bb6ef750ed7c1ba63de301ba0edd9128b" Dec 06 06:26:28 crc kubenswrapper[4809]: I1206 06:26:28.611872 4809 scope.go:117] "RemoveContainer" containerID="e586d2395ccc1ce18aea94ffe162c27ddadeed3797ac587c97054312fd868f8d" Dec 06 06:26:28 crc kubenswrapper[4809]: I1206 06:26:28.637966 4809 scope.go:117] "RemoveContainer" containerID="5e3d1b548c31544bd9756e5d158976cbae49fb8a786506f50f03d202850b668e" Dec 06 06:26:28 crc kubenswrapper[4809]: I1206 06:26:28.663488 4809 scope.go:117] "RemoveContainer" containerID="b69305cf8058c4689e05fd5338f34098ca60dc761ac06609ef9e3f17d5e6a45e" Dec 06 06:26:34 crc kubenswrapper[4809]: I1206 06:26:34.496526 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:26:34 crc kubenswrapper[4809]: I1206 06:26:34.498135 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:26:53 crc kubenswrapper[4809]: I1206 06:26:53.041055 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-mc9z8"] Dec 06 06:26:53 crc kubenswrapper[4809]: I1206 06:26:53.051720 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-mc9z8"] Dec 06 06:26:53 crc kubenswrapper[4809]: I1206 06:26:53.405185 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="413a30cf-4b57-44d7-b9bf-3d400bf4b897" path="/var/lib/kubelet/pods/413a30cf-4b57-44d7-b9bf-3d400bf4b897/volumes" Dec 06 06:26:57 crc kubenswrapper[4809]: I1206 06:26:57.035142 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-hsm9x"] Dec 06 06:26:57 crc kubenswrapper[4809]: I1206 06:26:57.047326 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-hsm9x"] Dec 06 06:26:57 crc kubenswrapper[4809]: I1206 06:26:57.404672 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08195221-42d1-4ce3-93a6-050e44339013" path="/var/lib/kubelet/pods/08195221-42d1-4ce3-93a6-050e44339013/volumes" Dec 06 06:27:04 crc kubenswrapper[4809]: I1206 06:27:04.495959 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:27:04 crc kubenswrapper[4809]: I1206 06:27:04.496398 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:27:04 crc kubenswrapper[4809]: I1206 06:27:04.496442 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-npms2" Dec 06 06:27:04 crc kubenswrapper[4809]: I1206 06:27:04.497466 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"54ba3d8af898b6b995ee9173b17fafab2449b8d6a7b61c005c29dabf843d35fe"} pod="openshift-machine-config-operator/machine-config-daemon-npms2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 06:27:04 crc kubenswrapper[4809]: I1206 06:27:04.497526 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" containerID="cri-o://54ba3d8af898b6b995ee9173b17fafab2449b8d6a7b61c005c29dabf843d35fe" gracePeriod=600 Dec 06 06:27:05 crc kubenswrapper[4809]: I1206 06:27:05.973841 4809 generic.go:334] "Generic (PLEG): container finished" podID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerID="54ba3d8af898b6b995ee9173b17fafab2449b8d6a7b61c005c29dabf843d35fe" exitCode=0 Dec 06 06:27:05 crc kubenswrapper[4809]: I1206 06:27:05.973972 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" event={"ID":"cbf4cf62-024e-4703-a8b8-9aecda9cd26a","Type":"ContainerDied","Data":"54ba3d8af898b6b995ee9173b17fafab2449b8d6a7b61c005c29dabf843d35fe"} Dec 06 06:27:05 crc kubenswrapper[4809]: I1206 06:27:05.974561 4809 scope.go:117] "RemoveContainer" containerID="9bc2dc88b3b9f309ce0169aa510326dd77022dece4c69cab3141df00fe5c0106" Dec 06 06:27:06 crc kubenswrapper[4809]: I1206 06:27:06.989157 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" event={"ID":"cbf4cf62-024e-4703-a8b8-9aecda9cd26a","Type":"ContainerStarted","Data":"3aa8c940468b111de41fd7720fd216eebfc0bb377ed933226628430658845c32"} Dec 06 06:27:07 crc kubenswrapper[4809]: I1206 06:27:07.038602 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-tc5gt"] Dec 06 06:27:07 crc kubenswrapper[4809]: I1206 06:27:07.052765 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-tc5gt"] Dec 06 06:27:07 crc kubenswrapper[4809]: I1206 06:27:07.407219 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c805cd51-5aaa-4c78-b98e-29a5a6ae44a9" path="/var/lib/kubelet/pods/c805cd51-5aaa-4c78-b98e-29a5a6ae44a9/volumes" Dec 06 06:27:15 crc kubenswrapper[4809]: I1206 06:27:15.040456 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-m98hv"] Dec 06 06:27:15 crc kubenswrapper[4809]: I1206 06:27:15.052635 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-m98hv"] Dec 06 06:27:15 crc kubenswrapper[4809]: I1206 06:27:15.068612 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-4gxwk"] Dec 06 06:27:15 crc kubenswrapper[4809]: I1206 06:27:15.081236 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-4gxwk"] Dec 06 06:27:17 crc kubenswrapper[4809]: I1206 06:27:17.345914 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="662e24e4-5dd6-4d18-822c-2863017f15af" path="/var/lib/kubelet/pods/662e24e4-5dd6-4d18-822c-2863017f15af/volumes" Dec 06 06:27:17 crc kubenswrapper[4809]: I1206 06:27:17.351813 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8336af68-51f9-4f3a-8c92-61dd1aba9067" path="/var/lib/kubelet/pods/8336af68-51f9-4f3a-8c92-61dd1aba9067/volumes" Dec 06 06:27:17 crc kubenswrapper[4809]: E1206 06:27:17.353138 4809 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.966s" Dec 06 06:27:29 crc kubenswrapper[4809]: I1206 06:27:29.226833 4809 scope.go:117] "RemoveContainer" containerID="12a159c9d2e59d6ab6c567a3ac994fcd15176cf11ff241ba517b9ef132bec4bc" Dec 06 06:27:29 crc kubenswrapper[4809]: I1206 06:27:29.287904 4809 scope.go:117] "RemoveContainer" containerID="7c24eacb26efb396e941d2db0e9a7e55a9a496af0f47192b9d1f6a49c739e882" Dec 06 06:27:29 crc kubenswrapper[4809]: I1206 06:27:29.395041 4809 scope.go:117] "RemoveContainer" containerID="5fa4ff9a4f10e7d52d4499104cc17b8069a880b80c0127d3a607c15b16f88cf1" Dec 06 06:27:29 crc kubenswrapper[4809]: I1206 06:27:29.427019 4809 scope.go:117] "RemoveContainer" containerID="cbb746c60ce6a9134f7e45e0a79d24227920a2c5bece823b7c07cef89c91e92a" Dec 06 06:27:29 crc kubenswrapper[4809]: I1206 06:27:29.547730 4809 scope.go:117] "RemoveContainer" containerID="096c45017712c6fe423f137af1e72ea25ccc70231ab4d70c49373ba0b823944c" Dec 06 06:27:30 crc kubenswrapper[4809]: I1206 06:27:30.049752 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-xs2c8"] Dec 06 06:27:30 crc kubenswrapper[4809]: I1206 06:27:30.059050 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-xs2c8"] Dec 06 06:27:30 crc kubenswrapper[4809]: I1206 06:27:30.264639 4809 generic.go:334] "Generic (PLEG): container finished" podID="13f6bc16-e535-42af-953e-b6b3827b35d1" containerID="17d61f6d547386398decc45eb8d3f8bb5e620442628bcf05d07c4256076808d4" exitCode=0 Dec 06 06:27:30 crc kubenswrapper[4809]: I1206 06:27:30.264691 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4c8nk" event={"ID":"13f6bc16-e535-42af-953e-b6b3827b35d1","Type":"ContainerDied","Data":"17d61f6d547386398decc45eb8d3f8bb5e620442628bcf05d07c4256076808d4"} Dec 06 06:27:31 crc kubenswrapper[4809]: I1206 06:27:31.407146 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c75cf833-5179-4e18-8f74-3faed72757ed" path="/var/lib/kubelet/pods/c75cf833-5179-4e18-8f74-3faed72757ed/volumes" Dec 06 06:27:31 crc kubenswrapper[4809]: I1206 06:27:31.814347 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4c8nk" Dec 06 06:27:31 crc kubenswrapper[4809]: I1206 06:27:31.988882 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13f6bc16-e535-42af-953e-b6b3827b35d1-bootstrap-combined-ca-bundle\") pod \"13f6bc16-e535-42af-953e-b6b3827b35d1\" (UID: \"13f6bc16-e535-42af-953e-b6b3827b35d1\") " Dec 06 06:27:31 crc kubenswrapper[4809]: I1206 06:27:31.989139 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/13f6bc16-e535-42af-953e-b6b3827b35d1-ssh-key\") pod \"13f6bc16-e535-42af-953e-b6b3827b35d1\" (UID: \"13f6bc16-e535-42af-953e-b6b3827b35d1\") " Dec 06 06:27:31 crc kubenswrapper[4809]: I1206 06:27:31.989408 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/13f6bc16-e535-42af-953e-b6b3827b35d1-inventory\") pod \"13f6bc16-e535-42af-953e-b6b3827b35d1\" (UID: \"13f6bc16-e535-42af-953e-b6b3827b35d1\") " Dec 06 06:27:31 crc kubenswrapper[4809]: I1206 06:27:31.990080 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-29z5f\" (UniqueName: \"kubernetes.io/projected/13f6bc16-e535-42af-953e-b6b3827b35d1-kube-api-access-29z5f\") pod \"13f6bc16-e535-42af-953e-b6b3827b35d1\" (UID: \"13f6bc16-e535-42af-953e-b6b3827b35d1\") " Dec 06 06:27:31 crc kubenswrapper[4809]: I1206 06:27:31.995754 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13f6bc16-e535-42af-953e-b6b3827b35d1-kube-api-access-29z5f" (OuterVolumeSpecName: "kube-api-access-29z5f") pod "13f6bc16-e535-42af-953e-b6b3827b35d1" (UID: "13f6bc16-e535-42af-953e-b6b3827b35d1"). InnerVolumeSpecName "kube-api-access-29z5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:27:31 crc kubenswrapper[4809]: I1206 06:27:31.995763 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13f6bc16-e535-42af-953e-b6b3827b35d1-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "13f6bc16-e535-42af-953e-b6b3827b35d1" (UID: "13f6bc16-e535-42af-953e-b6b3827b35d1"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:27:32 crc kubenswrapper[4809]: I1206 06:27:32.031232 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13f6bc16-e535-42af-953e-b6b3827b35d1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "13f6bc16-e535-42af-953e-b6b3827b35d1" (UID: "13f6bc16-e535-42af-953e-b6b3827b35d1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:27:32 crc kubenswrapper[4809]: I1206 06:27:32.042519 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13f6bc16-e535-42af-953e-b6b3827b35d1-inventory" (OuterVolumeSpecName: "inventory") pod "13f6bc16-e535-42af-953e-b6b3827b35d1" (UID: "13f6bc16-e535-42af-953e-b6b3827b35d1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:27:32 crc kubenswrapper[4809]: I1206 06:27:32.094361 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/13f6bc16-e535-42af-953e-b6b3827b35d1-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 06:27:32 crc kubenswrapper[4809]: I1206 06:27:32.094655 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-29z5f\" (UniqueName: \"kubernetes.io/projected/13f6bc16-e535-42af-953e-b6b3827b35d1-kube-api-access-29z5f\") on node \"crc\" DevicePath \"\"" Dec 06 06:27:32 crc kubenswrapper[4809]: I1206 06:27:32.094784 4809 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13f6bc16-e535-42af-953e-b6b3827b35d1-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:27:32 crc kubenswrapper[4809]: I1206 06:27:32.094913 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/13f6bc16-e535-42af-953e-b6b3827b35d1-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 06:27:32 crc kubenswrapper[4809]: I1206 06:27:32.304051 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4c8nk" event={"ID":"13f6bc16-e535-42af-953e-b6b3827b35d1","Type":"ContainerDied","Data":"8b40d9d44a2b8d0cc14bc3aa415bd4ad612f5e66a23652a4989d089ae2370f94"} Dec 06 06:27:32 crc kubenswrapper[4809]: I1206 06:27:32.304109 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8b40d9d44a2b8d0cc14bc3aa415bd4ad612f5e66a23652a4989d089ae2370f94" Dec 06 06:27:32 crc kubenswrapper[4809]: I1206 06:27:32.304197 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4c8nk" Dec 06 06:27:32 crc kubenswrapper[4809]: I1206 06:27:32.392704 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-xphmc"] Dec 06 06:27:32 crc kubenswrapper[4809]: E1206 06:27:32.393437 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2" containerName="registry-server" Dec 06 06:27:32 crc kubenswrapper[4809]: I1206 06:27:32.393616 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2" containerName="registry-server" Dec 06 06:27:32 crc kubenswrapper[4809]: E1206 06:27:32.393694 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13f6bc16-e535-42af-953e-b6b3827b35d1" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 06 06:27:32 crc kubenswrapper[4809]: I1206 06:27:32.393748 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="13f6bc16-e535-42af-953e-b6b3827b35d1" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 06 06:27:32 crc kubenswrapper[4809]: E1206 06:27:32.393898 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2" containerName="extract-content" Dec 06 06:27:32 crc kubenswrapper[4809]: I1206 06:27:32.393978 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2" containerName="extract-content" Dec 06 06:27:32 crc kubenswrapper[4809]: E1206 06:27:32.394048 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2" containerName="extract-utilities" Dec 06 06:27:32 crc kubenswrapper[4809]: I1206 06:27:32.394099 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2" containerName="extract-utilities" Dec 06 06:27:32 crc kubenswrapper[4809]: I1206 06:27:32.394373 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="13f6bc16-e535-42af-953e-b6b3827b35d1" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 06 06:27:32 crc kubenswrapper[4809]: I1206 06:27:32.394444 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b97ccfdd-2f15-4979-bc2a-9e7e1cef48f2" containerName="registry-server" Dec 06 06:27:32 crc kubenswrapper[4809]: I1206 06:27:32.395279 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-xphmc" Dec 06 06:27:32 crc kubenswrapper[4809]: I1206 06:27:32.402674 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 06:27:32 crc kubenswrapper[4809]: I1206 06:27:32.402705 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-bzlf5" Dec 06 06:27:32 crc kubenswrapper[4809]: I1206 06:27:32.403022 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 06:27:32 crc kubenswrapper[4809]: I1206 06:27:32.403293 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 06:27:32 crc kubenswrapper[4809]: I1206 06:27:32.422764 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-xphmc"] Dec 06 06:27:32 crc kubenswrapper[4809]: I1206 06:27:32.509315 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-xphmc\" (UID: \"cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-xphmc" Dec 06 06:27:32 crc kubenswrapper[4809]: I1206 06:27:32.509373 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqqhr\" (UniqueName: \"kubernetes.io/projected/cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad-kube-api-access-cqqhr\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-xphmc\" (UID: \"cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-xphmc" Dec 06 06:27:32 crc kubenswrapper[4809]: I1206 06:27:32.509712 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-xphmc\" (UID: \"cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-xphmc" Dec 06 06:27:32 crc kubenswrapper[4809]: I1206 06:27:32.611460 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-xphmc\" (UID: \"cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-xphmc" Dec 06 06:27:32 crc kubenswrapper[4809]: I1206 06:27:32.611777 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqqhr\" (UniqueName: \"kubernetes.io/projected/cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad-kube-api-access-cqqhr\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-xphmc\" (UID: \"cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-xphmc" Dec 06 06:27:32 crc kubenswrapper[4809]: I1206 06:27:32.612083 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-xphmc\" (UID: \"cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-xphmc" Dec 06 06:27:32 crc kubenswrapper[4809]: I1206 06:27:32.615320 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-xphmc\" (UID: \"cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-xphmc" Dec 06 06:27:32 crc kubenswrapper[4809]: I1206 06:27:32.615565 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-xphmc\" (UID: \"cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-xphmc" Dec 06 06:27:32 crc kubenswrapper[4809]: I1206 06:27:32.634744 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqqhr\" (UniqueName: \"kubernetes.io/projected/cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad-kube-api-access-cqqhr\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-xphmc\" (UID: \"cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-xphmc" Dec 06 06:27:32 crc kubenswrapper[4809]: I1206 06:27:32.718172 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-xphmc" Dec 06 06:27:33 crc kubenswrapper[4809]: I1206 06:27:33.332153 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-xphmc"] Dec 06 06:27:34 crc kubenswrapper[4809]: I1206 06:27:34.326230 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-xphmc" event={"ID":"cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad","Type":"ContainerStarted","Data":"44c912fe85b83153c4d53df5fbc41d9ef70e5eb4ae00cee10246ea4e6b4b121b"} Dec 06 06:27:39 crc kubenswrapper[4809]: I1206 06:27:39.579140 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/metallb-operator-webhook-server-765b7d7f99-8dngt" podUID="60798ea0-75be-4a0c-a8c6-9fb431ba0e67" containerName="webhook-server" probeResult="failure" output="Get \"http://10.217.0.93:7472/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 06:27:52 crc kubenswrapper[4809]: I1206 06:27:52.536841 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-xphmc" event={"ID":"cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad","Type":"ContainerStarted","Data":"5ab47ea2f039a313c7434082ac029fc16e227ff5d8d0f659dc4cc755ebfd3c27"} Dec 06 06:27:52 crc kubenswrapper[4809]: I1206 06:27:52.553164 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-xphmc" podStartSLOduration=3.306283513 podStartE2EDuration="20.55314754s" podCreationTimestamp="2025-12-06 06:27:32 +0000 UTC" firstStartedPulling="2025-12-06 06:27:33.342849937 +0000 UTC m=+2178.231832879" lastFinishedPulling="2025-12-06 06:27:50.589713964 +0000 UTC m=+2195.478696906" observedRunningTime="2025-12-06 06:27:52.551319191 +0000 UTC m=+2197.440302133" watchObservedRunningTime="2025-12-06 06:27:52.55314754 +0000 UTC m=+2197.442130482" Dec 06 06:28:21 crc kubenswrapper[4809]: I1206 06:28:21.041999 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-sd9ds"] Dec 06 06:28:21 crc kubenswrapper[4809]: I1206 06:28:21.051807 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-bmh2p"] Dec 06 06:28:21 crc kubenswrapper[4809]: I1206 06:28:21.061695 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-sd9ds"] Dec 06 06:28:21 crc kubenswrapper[4809]: I1206 06:28:21.071358 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-bmh2p"] Dec 06 06:28:21 crc kubenswrapper[4809]: I1206 06:28:21.405438 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6b76b1b-c52d-464c-9b94-1e67d459f3b3" path="/var/lib/kubelet/pods/c6b76b1b-c52d-464c-9b94-1e67d459f3b3/volumes" Dec 06 06:28:21 crc kubenswrapper[4809]: I1206 06:28:21.407041 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee554177-cc9d-4969-9085-e507976f0f28" path="/var/lib/kubelet/pods/ee554177-cc9d-4969-9085-e507976f0f28/volumes" Dec 06 06:28:22 crc kubenswrapper[4809]: I1206 06:28:22.038289 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-4733-account-create-update-z8gxx"] Dec 06 06:28:22 crc kubenswrapper[4809]: I1206 06:28:22.056174 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-8841-account-create-update-qhld2"] Dec 06 06:28:22 crc kubenswrapper[4809]: I1206 06:28:22.075955 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-fb6zj"] Dec 06 06:28:22 crc kubenswrapper[4809]: I1206 06:28:22.088177 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-4733-account-create-update-z8gxx"] Dec 06 06:28:22 crc kubenswrapper[4809]: I1206 06:28:22.098204 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-356a-account-create-update-6mtbv"] Dec 06 06:28:22 crc kubenswrapper[4809]: I1206 06:28:22.108465 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-356a-account-create-update-6mtbv"] Dec 06 06:28:22 crc kubenswrapper[4809]: I1206 06:28:22.118346 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-8841-account-create-update-qhld2"] Dec 06 06:28:22 crc kubenswrapper[4809]: I1206 06:28:22.130986 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-fb6zj"] Dec 06 06:28:23 crc kubenswrapper[4809]: I1206 06:28:23.402327 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="412c5e46-e0e4-4b8a-af43-616019f9b619" path="/var/lib/kubelet/pods/412c5e46-e0e4-4b8a-af43-616019f9b619/volumes" Dec 06 06:28:23 crc kubenswrapper[4809]: I1206 06:28:23.402977 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5984ffba-4d1d-4976-b97f-606e2f557b03" path="/var/lib/kubelet/pods/5984ffba-4d1d-4976-b97f-606e2f557b03/volumes" Dec 06 06:28:23 crc kubenswrapper[4809]: I1206 06:28:23.404384 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7494a97b-56a3-45e8-a189-48eea155d3b3" path="/var/lib/kubelet/pods/7494a97b-56a3-45e8-a189-48eea155d3b3/volumes" Dec 06 06:28:23 crc kubenswrapper[4809]: I1206 06:28:23.404994 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0866cc1-8f06-4514-9939-5a4cfadc72b3" path="/var/lib/kubelet/pods/a0866cc1-8f06-4514-9939-5a4cfadc72b3/volumes" Dec 06 06:28:29 crc kubenswrapper[4809]: I1206 06:28:29.722026 4809 scope.go:117] "RemoveContainer" containerID="f8ba6acd951144b6cbf9feccba3d35a19374d54fcead893ca299bd97363407f4" Dec 06 06:28:29 crc kubenswrapper[4809]: I1206 06:28:29.859280 4809 scope.go:117] "RemoveContainer" containerID="fa80f87dee5acf1e4481c0a5cb3a2046e006e209b4ecea87d99c9b30c9815b13" Dec 06 06:28:29 crc kubenswrapper[4809]: I1206 06:28:29.887910 4809 scope.go:117] "RemoveContainer" containerID="e77b147ecc98479207fbf89e2b41c04292a66b4172aced62be7f5170feb90686" Dec 06 06:28:30 crc kubenswrapper[4809]: I1206 06:28:30.198794 4809 scope.go:117] "RemoveContainer" containerID="b30146671b82ca2efc77dadcc395a46bc9f67d0a5d0f6683eecfa651518e4103" Dec 06 06:28:30 crc kubenswrapper[4809]: I1206 06:28:30.237304 4809 scope.go:117] "RemoveContainer" containerID="308715c3c15148bf3a30aa4c43e68ba387da20816fc7ecf94ae2b946ccd0e9ac" Dec 06 06:28:30 crc kubenswrapper[4809]: I1206 06:28:30.305109 4809 scope.go:117] "RemoveContainer" containerID="fd7f07a5ac3909a5a15b8e1fdc9686616ea3e3cc8fbe64e0651e6f5fd586ebfe" Dec 06 06:28:30 crc kubenswrapper[4809]: I1206 06:28:30.572385 4809 scope.go:117] "RemoveContainer" containerID="13167b3c813f9d5bc3d34d0d09928c7bb068ae3337cb629047c05d9accdf14ab" Dec 06 06:29:11 crc kubenswrapper[4809]: I1206 06:29:11.049517 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-vt99m"] Dec 06 06:29:11 crc kubenswrapper[4809]: I1206 06:29:11.061272 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-vt99m"] Dec 06 06:29:11 crc kubenswrapper[4809]: I1206 06:29:11.405677 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13e0242d-c42e-4aad-95a2-62de95d87e42" path="/var/lib/kubelet/pods/13e0242d-c42e-4aad-95a2-62de95d87e42/volumes" Dec 06 06:29:25 crc kubenswrapper[4809]: I1206 06:29:25.045640 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-133e-account-create-update-cpzqq"] Dec 06 06:29:25 crc kubenswrapper[4809]: I1206 06:29:25.059470 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-ld8fs"] Dec 06 06:29:25 crc kubenswrapper[4809]: I1206 06:29:25.068472 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-133e-account-create-update-cpzqq"] Dec 06 06:29:25 crc kubenswrapper[4809]: I1206 06:29:25.077968 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-ld8fs"] Dec 06 06:29:25 crc kubenswrapper[4809]: I1206 06:29:25.412659 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="353be8e6-45e2-4731-bdf0-62a5260adf96" path="/var/lib/kubelet/pods/353be8e6-45e2-4731-bdf0-62a5260adf96/volumes" Dec 06 06:29:25 crc kubenswrapper[4809]: I1206 06:29:25.413671 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62e3de39-6323-4837-b746-b2b51b7d8e6c" path="/var/lib/kubelet/pods/62e3de39-6323-4837-b746-b2b51b7d8e6c/volumes" Dec 06 06:29:30 crc kubenswrapper[4809]: I1206 06:29:30.771026 4809 scope.go:117] "RemoveContainer" containerID="6ff2f5a5901ee62a2d83e64767490215843e438ad6b3ed23aad29389d4960353" Dec 06 06:29:30 crc kubenswrapper[4809]: I1206 06:29:30.812714 4809 scope.go:117] "RemoveContainer" containerID="71c474112cf4eb39fccdbfb9f48003b4719be0290686f158fed9729a0af7ebad" Dec 06 06:29:30 crc kubenswrapper[4809]: I1206 06:29:30.858434 4809 scope.go:117] "RemoveContainer" containerID="2a58fd22495f3af98ef23bc24a54ae7cfc770faff52f1f4baae10ae5065326b0" Dec 06 06:29:34 crc kubenswrapper[4809]: I1206 06:29:34.496453 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:29:34 crc kubenswrapper[4809]: I1206 06:29:34.496902 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:29:39 crc kubenswrapper[4809]: I1206 06:29:39.040958 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-x8zvs"] Dec 06 06:29:39 crc kubenswrapper[4809]: I1206 06:29:39.053969 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-x8zvs"] Dec 06 06:29:39 crc kubenswrapper[4809]: I1206 06:29:39.408295 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dad7df76-c6ae-4d0c-aef7-63ca9efb77a7" path="/var/lib/kubelet/pods/dad7df76-c6ae-4d0c-aef7-63ca9efb77a7/volumes" Dec 06 06:29:41 crc kubenswrapper[4809]: I1206 06:29:41.032459 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8jvh2"] Dec 06 06:29:41 crc kubenswrapper[4809]: I1206 06:29:41.045535 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8jvh2"] Dec 06 06:29:41 crc kubenswrapper[4809]: I1206 06:29:41.408611 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4af32e7-96e5-4f86-b271-6b4efcace0dc" path="/var/lib/kubelet/pods/d4af32e7-96e5-4f86-b271-6b4efcace0dc/volumes" Dec 06 06:30:00 crc kubenswrapper[4809]: I1206 06:30:00.187462 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416710-sjkdm"] Dec 06 06:30:00 crc kubenswrapper[4809]: I1206 06:30:00.189583 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416710-sjkdm" Dec 06 06:30:00 crc kubenswrapper[4809]: I1206 06:30:00.191144 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 06:30:00 crc kubenswrapper[4809]: I1206 06:30:00.191689 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 06:30:00 crc kubenswrapper[4809]: I1206 06:30:00.200192 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416710-sjkdm"] Dec 06 06:30:00 crc kubenswrapper[4809]: I1206 06:30:00.258962 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3e97b08e-c54b-42c8-9e6f-893215182a8e-config-volume\") pod \"collect-profiles-29416710-sjkdm\" (UID: \"3e97b08e-c54b-42c8-9e6f-893215182a8e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416710-sjkdm" Dec 06 06:30:00 crc kubenswrapper[4809]: I1206 06:30:00.259144 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cd97\" (UniqueName: \"kubernetes.io/projected/3e97b08e-c54b-42c8-9e6f-893215182a8e-kube-api-access-2cd97\") pod \"collect-profiles-29416710-sjkdm\" (UID: \"3e97b08e-c54b-42c8-9e6f-893215182a8e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416710-sjkdm" Dec 06 06:30:00 crc kubenswrapper[4809]: I1206 06:30:00.259183 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3e97b08e-c54b-42c8-9e6f-893215182a8e-secret-volume\") pod \"collect-profiles-29416710-sjkdm\" (UID: \"3e97b08e-c54b-42c8-9e6f-893215182a8e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416710-sjkdm" Dec 06 06:30:00 crc kubenswrapper[4809]: I1206 06:30:00.361235 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3e97b08e-c54b-42c8-9e6f-893215182a8e-config-volume\") pod \"collect-profiles-29416710-sjkdm\" (UID: \"3e97b08e-c54b-42c8-9e6f-893215182a8e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416710-sjkdm" Dec 06 06:30:00 crc kubenswrapper[4809]: I1206 06:30:00.361398 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cd97\" (UniqueName: \"kubernetes.io/projected/3e97b08e-c54b-42c8-9e6f-893215182a8e-kube-api-access-2cd97\") pod \"collect-profiles-29416710-sjkdm\" (UID: \"3e97b08e-c54b-42c8-9e6f-893215182a8e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416710-sjkdm" Dec 06 06:30:00 crc kubenswrapper[4809]: I1206 06:30:00.361465 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3e97b08e-c54b-42c8-9e6f-893215182a8e-secret-volume\") pod \"collect-profiles-29416710-sjkdm\" (UID: \"3e97b08e-c54b-42c8-9e6f-893215182a8e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416710-sjkdm" Dec 06 06:30:00 crc kubenswrapper[4809]: I1206 06:30:00.362747 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3e97b08e-c54b-42c8-9e6f-893215182a8e-config-volume\") pod \"collect-profiles-29416710-sjkdm\" (UID: \"3e97b08e-c54b-42c8-9e6f-893215182a8e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416710-sjkdm" Dec 06 06:30:00 crc kubenswrapper[4809]: I1206 06:30:00.371067 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3e97b08e-c54b-42c8-9e6f-893215182a8e-secret-volume\") pod \"collect-profiles-29416710-sjkdm\" (UID: \"3e97b08e-c54b-42c8-9e6f-893215182a8e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416710-sjkdm" Dec 06 06:30:00 crc kubenswrapper[4809]: I1206 06:30:00.385924 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cd97\" (UniqueName: \"kubernetes.io/projected/3e97b08e-c54b-42c8-9e6f-893215182a8e-kube-api-access-2cd97\") pod \"collect-profiles-29416710-sjkdm\" (UID: \"3e97b08e-c54b-42c8-9e6f-893215182a8e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416710-sjkdm" Dec 06 06:30:00 crc kubenswrapper[4809]: I1206 06:30:00.515993 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416710-sjkdm" Dec 06 06:30:00 crc kubenswrapper[4809]: I1206 06:30:00.971975 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416710-sjkdm"] Dec 06 06:30:01 crc kubenswrapper[4809]: I1206 06:30:01.045109 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416710-sjkdm" event={"ID":"3e97b08e-c54b-42c8-9e6f-893215182a8e","Type":"ContainerStarted","Data":"cc75efd85d4bb7d13a738b4b14ed41a8f57a9112fc0dc3f9322fe164adefaa7c"} Dec 06 06:30:02 crc kubenswrapper[4809]: I1206 06:30:02.058112 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416710-sjkdm" event={"ID":"3e97b08e-c54b-42c8-9e6f-893215182a8e","Type":"ContainerStarted","Data":"40c2ed333d9348bfc04097dc0c2374eeac937cb4c29d26f625c8626cba492c73"} Dec 06 06:30:03 crc kubenswrapper[4809]: I1206 06:30:03.068631 4809 generic.go:334] "Generic (PLEG): container finished" podID="3e97b08e-c54b-42c8-9e6f-893215182a8e" containerID="40c2ed333d9348bfc04097dc0c2374eeac937cb4c29d26f625c8626cba492c73" exitCode=0 Dec 06 06:30:03 crc kubenswrapper[4809]: I1206 06:30:03.068722 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416710-sjkdm" event={"ID":"3e97b08e-c54b-42c8-9e6f-893215182a8e","Type":"ContainerDied","Data":"40c2ed333d9348bfc04097dc0c2374eeac937cb4c29d26f625c8626cba492c73"} Dec 06 06:30:04 crc kubenswrapper[4809]: I1206 06:30:04.496322 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:30:04 crc kubenswrapper[4809]: I1206 06:30:04.496657 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:30:04 crc kubenswrapper[4809]: I1206 06:30:04.515230 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416710-sjkdm" Dec 06 06:30:04 crc kubenswrapper[4809]: I1206 06:30:04.565186 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2cd97\" (UniqueName: \"kubernetes.io/projected/3e97b08e-c54b-42c8-9e6f-893215182a8e-kube-api-access-2cd97\") pod \"3e97b08e-c54b-42c8-9e6f-893215182a8e\" (UID: \"3e97b08e-c54b-42c8-9e6f-893215182a8e\") " Dec 06 06:30:04 crc kubenswrapper[4809]: I1206 06:30:04.565577 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3e97b08e-c54b-42c8-9e6f-893215182a8e-config-volume\") pod \"3e97b08e-c54b-42c8-9e6f-893215182a8e\" (UID: \"3e97b08e-c54b-42c8-9e6f-893215182a8e\") " Dec 06 06:30:04 crc kubenswrapper[4809]: I1206 06:30:04.565655 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3e97b08e-c54b-42c8-9e6f-893215182a8e-secret-volume\") pod \"3e97b08e-c54b-42c8-9e6f-893215182a8e\" (UID: \"3e97b08e-c54b-42c8-9e6f-893215182a8e\") " Dec 06 06:30:04 crc kubenswrapper[4809]: I1206 06:30:04.566209 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e97b08e-c54b-42c8-9e6f-893215182a8e-config-volume" (OuterVolumeSpecName: "config-volume") pod "3e97b08e-c54b-42c8-9e6f-893215182a8e" (UID: "3e97b08e-c54b-42c8-9e6f-893215182a8e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:30:04 crc kubenswrapper[4809]: I1206 06:30:04.567153 4809 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3e97b08e-c54b-42c8-9e6f-893215182a8e-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 06:30:04 crc kubenswrapper[4809]: I1206 06:30:04.572152 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e97b08e-c54b-42c8-9e6f-893215182a8e-kube-api-access-2cd97" (OuterVolumeSpecName: "kube-api-access-2cd97") pod "3e97b08e-c54b-42c8-9e6f-893215182a8e" (UID: "3e97b08e-c54b-42c8-9e6f-893215182a8e"). InnerVolumeSpecName "kube-api-access-2cd97". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:30:04 crc kubenswrapper[4809]: I1206 06:30:04.572911 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e97b08e-c54b-42c8-9e6f-893215182a8e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3e97b08e-c54b-42c8-9e6f-893215182a8e" (UID: "3e97b08e-c54b-42c8-9e6f-893215182a8e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:30:04 crc kubenswrapper[4809]: I1206 06:30:04.671108 4809 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3e97b08e-c54b-42c8-9e6f-893215182a8e-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 06:30:04 crc kubenswrapper[4809]: I1206 06:30:04.671651 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2cd97\" (UniqueName: \"kubernetes.io/projected/3e97b08e-c54b-42c8-9e6f-893215182a8e-kube-api-access-2cd97\") on node \"crc\" DevicePath \"\"" Dec 06 06:30:05 crc kubenswrapper[4809]: I1206 06:30:05.092857 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416710-sjkdm" event={"ID":"3e97b08e-c54b-42c8-9e6f-893215182a8e","Type":"ContainerDied","Data":"cc75efd85d4bb7d13a738b4b14ed41a8f57a9112fc0dc3f9322fe164adefaa7c"} Dec 06 06:30:05 crc kubenswrapper[4809]: I1206 06:30:05.093173 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc75efd85d4bb7d13a738b4b14ed41a8f57a9112fc0dc3f9322fe164adefaa7c" Dec 06 06:30:05 crc kubenswrapper[4809]: I1206 06:30:05.092969 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416710-sjkdm" Dec 06 06:30:05 crc kubenswrapper[4809]: I1206 06:30:05.613545 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416665-lsfkp"] Dec 06 06:30:05 crc kubenswrapper[4809]: I1206 06:30:05.622895 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416665-lsfkp"] Dec 06 06:30:07 crc kubenswrapper[4809]: I1206 06:30:07.411230 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c892e996-27b7-4b8d-bd92-7eac2df7702b" path="/var/lib/kubelet/pods/c892e996-27b7-4b8d-bd92-7eac2df7702b/volumes" Dec 06 06:30:12 crc kubenswrapper[4809]: I1206 06:30:12.135089 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9fjmc"] Dec 06 06:30:12 crc kubenswrapper[4809]: E1206 06:30:12.137501 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e97b08e-c54b-42c8-9e6f-893215182a8e" containerName="collect-profiles" Dec 06 06:30:12 crc kubenswrapper[4809]: I1206 06:30:12.137525 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e97b08e-c54b-42c8-9e6f-893215182a8e" containerName="collect-profiles" Dec 06 06:30:12 crc kubenswrapper[4809]: I1206 06:30:12.138047 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e97b08e-c54b-42c8-9e6f-893215182a8e" containerName="collect-profiles" Dec 06 06:30:12 crc kubenswrapper[4809]: I1206 06:30:12.143861 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9fjmc" Dec 06 06:30:12 crc kubenswrapper[4809]: I1206 06:30:12.158328 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9fjmc"] Dec 06 06:30:12 crc kubenswrapper[4809]: I1206 06:30:12.288207 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e285912-94a1-441c-83b8-38e561cc3a0b-utilities\") pod \"redhat-marketplace-9fjmc\" (UID: \"2e285912-94a1-441c-83b8-38e561cc3a0b\") " pod="openshift-marketplace/redhat-marketplace-9fjmc" Dec 06 06:30:12 crc kubenswrapper[4809]: I1206 06:30:12.288469 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e285912-94a1-441c-83b8-38e561cc3a0b-catalog-content\") pod \"redhat-marketplace-9fjmc\" (UID: \"2e285912-94a1-441c-83b8-38e561cc3a0b\") " pod="openshift-marketplace/redhat-marketplace-9fjmc" Dec 06 06:30:12 crc kubenswrapper[4809]: I1206 06:30:12.288516 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5sj6s\" (UniqueName: \"kubernetes.io/projected/2e285912-94a1-441c-83b8-38e561cc3a0b-kube-api-access-5sj6s\") pod \"redhat-marketplace-9fjmc\" (UID: \"2e285912-94a1-441c-83b8-38e561cc3a0b\") " pod="openshift-marketplace/redhat-marketplace-9fjmc" Dec 06 06:30:12 crc kubenswrapper[4809]: I1206 06:30:12.290593 4809 generic.go:334] "Generic (PLEG): container finished" podID="cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad" containerID="5ab47ea2f039a313c7434082ac029fc16e227ff5d8d0f659dc4cc755ebfd3c27" exitCode=0 Dec 06 06:30:12 crc kubenswrapper[4809]: I1206 06:30:12.290690 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-xphmc" event={"ID":"cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad","Type":"ContainerDied","Data":"5ab47ea2f039a313c7434082ac029fc16e227ff5d8d0f659dc4cc755ebfd3c27"} Dec 06 06:30:12 crc kubenswrapper[4809]: I1206 06:30:12.392730 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e285912-94a1-441c-83b8-38e561cc3a0b-catalog-content\") pod \"redhat-marketplace-9fjmc\" (UID: \"2e285912-94a1-441c-83b8-38e561cc3a0b\") " pod="openshift-marketplace/redhat-marketplace-9fjmc" Dec 06 06:30:12 crc kubenswrapper[4809]: I1206 06:30:12.392794 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5sj6s\" (UniqueName: \"kubernetes.io/projected/2e285912-94a1-441c-83b8-38e561cc3a0b-kube-api-access-5sj6s\") pod \"redhat-marketplace-9fjmc\" (UID: \"2e285912-94a1-441c-83b8-38e561cc3a0b\") " pod="openshift-marketplace/redhat-marketplace-9fjmc" Dec 06 06:30:12 crc kubenswrapper[4809]: I1206 06:30:12.392910 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e285912-94a1-441c-83b8-38e561cc3a0b-utilities\") pod \"redhat-marketplace-9fjmc\" (UID: \"2e285912-94a1-441c-83b8-38e561cc3a0b\") " pod="openshift-marketplace/redhat-marketplace-9fjmc" Dec 06 06:30:12 crc kubenswrapper[4809]: I1206 06:30:12.393706 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e285912-94a1-441c-83b8-38e561cc3a0b-utilities\") pod \"redhat-marketplace-9fjmc\" (UID: \"2e285912-94a1-441c-83b8-38e561cc3a0b\") " pod="openshift-marketplace/redhat-marketplace-9fjmc" Dec 06 06:30:12 crc kubenswrapper[4809]: I1206 06:30:12.393763 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e285912-94a1-441c-83b8-38e561cc3a0b-catalog-content\") pod \"redhat-marketplace-9fjmc\" (UID: \"2e285912-94a1-441c-83b8-38e561cc3a0b\") " pod="openshift-marketplace/redhat-marketplace-9fjmc" Dec 06 06:30:12 crc kubenswrapper[4809]: I1206 06:30:12.437378 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5sj6s\" (UniqueName: \"kubernetes.io/projected/2e285912-94a1-441c-83b8-38e561cc3a0b-kube-api-access-5sj6s\") pod \"redhat-marketplace-9fjmc\" (UID: \"2e285912-94a1-441c-83b8-38e561cc3a0b\") " pod="openshift-marketplace/redhat-marketplace-9fjmc" Dec 06 06:30:12 crc kubenswrapper[4809]: I1206 06:30:12.482407 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9fjmc" Dec 06 06:30:13 crc kubenswrapper[4809]: I1206 06:30:13.004640 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9fjmc"] Dec 06 06:30:13 crc kubenswrapper[4809]: I1206 06:30:13.323122 4809 generic.go:334] "Generic (PLEG): container finished" podID="2e285912-94a1-441c-83b8-38e561cc3a0b" containerID="08dda443afeecc79da20d92093535d9a594dca97c320b2bc950371f59c2a808a" exitCode=0 Dec 06 06:30:13 crc kubenswrapper[4809]: I1206 06:30:13.323213 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9fjmc" event={"ID":"2e285912-94a1-441c-83b8-38e561cc3a0b","Type":"ContainerDied","Data":"08dda443afeecc79da20d92093535d9a594dca97c320b2bc950371f59c2a808a"} Dec 06 06:30:13 crc kubenswrapper[4809]: I1206 06:30:13.323535 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9fjmc" event={"ID":"2e285912-94a1-441c-83b8-38e561cc3a0b","Type":"ContainerStarted","Data":"bf37377d742fa5d7b680d257dea75ee608835af54eb0bdd86938258507faafb3"} Dec 06 06:30:13 crc kubenswrapper[4809]: I1206 06:30:13.327414 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 06:30:13 crc kubenswrapper[4809]: I1206 06:30:13.842169 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-xphmc" Dec 06 06:30:14 crc kubenswrapper[4809]: I1206 06:30:14.028856 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad-inventory\") pod \"cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad\" (UID: \"cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad\") " Dec 06 06:30:14 crc kubenswrapper[4809]: I1206 06:30:14.028901 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad-ssh-key\") pod \"cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad\" (UID: \"cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad\") " Dec 06 06:30:14 crc kubenswrapper[4809]: I1206 06:30:14.029092 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cqqhr\" (UniqueName: \"kubernetes.io/projected/cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad-kube-api-access-cqqhr\") pod \"cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad\" (UID: \"cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad\") " Dec 06 06:30:14 crc kubenswrapper[4809]: I1206 06:30:14.035028 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad-kube-api-access-cqqhr" (OuterVolumeSpecName: "kube-api-access-cqqhr") pod "cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad" (UID: "cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad"). InnerVolumeSpecName "kube-api-access-cqqhr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:30:14 crc kubenswrapper[4809]: I1206 06:30:14.066309 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad-inventory" (OuterVolumeSpecName: "inventory") pod "cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad" (UID: "cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:30:14 crc kubenswrapper[4809]: I1206 06:30:14.067474 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad" (UID: "cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:30:14 crc kubenswrapper[4809]: I1206 06:30:14.132406 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 06:30:14 crc kubenswrapper[4809]: I1206 06:30:14.132663 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 06:30:14 crc kubenswrapper[4809]: I1206 06:30:14.132749 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cqqhr\" (UniqueName: \"kubernetes.io/projected/cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad-kube-api-access-cqqhr\") on node \"crc\" DevicePath \"\"" Dec 06 06:30:14 crc kubenswrapper[4809]: I1206 06:30:14.347458 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-xphmc" event={"ID":"cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad","Type":"ContainerDied","Data":"44c912fe85b83153c4d53df5fbc41d9ef70e5eb4ae00cee10246ea4e6b4b121b"} Dec 06 06:30:14 crc kubenswrapper[4809]: I1206 06:30:14.347501 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="44c912fe85b83153c4d53df5fbc41d9ef70e5eb4ae00cee10246ea4e6b4b121b" Dec 06 06:30:14 crc kubenswrapper[4809]: I1206 06:30:14.347560 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-xphmc" Dec 06 06:30:14 crc kubenswrapper[4809]: I1206 06:30:14.392630 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-f4bfz"] Dec 06 06:30:14 crc kubenswrapper[4809]: E1206 06:30:14.393252 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 06 06:30:14 crc kubenswrapper[4809]: I1206 06:30:14.393277 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 06 06:30:14 crc kubenswrapper[4809]: I1206 06:30:14.393588 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 06 06:30:14 crc kubenswrapper[4809]: I1206 06:30:14.394624 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-f4bfz" Dec 06 06:30:14 crc kubenswrapper[4809]: I1206 06:30:14.399288 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 06:30:14 crc kubenswrapper[4809]: I1206 06:30:14.399511 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-bzlf5" Dec 06 06:30:14 crc kubenswrapper[4809]: I1206 06:30:14.399621 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 06:30:14 crc kubenswrapper[4809]: I1206 06:30:14.399658 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 06:30:14 crc kubenswrapper[4809]: I1206 06:30:14.404632 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-f4bfz"] Dec 06 06:30:14 crc kubenswrapper[4809]: I1206 06:30:14.545052 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e2ad775c-a1f4-4770-918b-4b3ab597634d-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-f4bfz\" (UID: \"e2ad775c-a1f4-4770-918b-4b3ab597634d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-f4bfz" Dec 06 06:30:14 crc kubenswrapper[4809]: I1206 06:30:14.545116 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e2ad775c-a1f4-4770-918b-4b3ab597634d-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-f4bfz\" (UID: \"e2ad775c-a1f4-4770-918b-4b3ab597634d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-f4bfz" Dec 06 06:30:14 crc kubenswrapper[4809]: I1206 06:30:14.545170 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwz95\" (UniqueName: \"kubernetes.io/projected/e2ad775c-a1f4-4770-918b-4b3ab597634d-kube-api-access-zwz95\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-f4bfz\" (UID: \"e2ad775c-a1f4-4770-918b-4b3ab597634d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-f4bfz" Dec 06 06:30:14 crc kubenswrapper[4809]: I1206 06:30:14.647743 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e2ad775c-a1f4-4770-918b-4b3ab597634d-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-f4bfz\" (UID: \"e2ad775c-a1f4-4770-918b-4b3ab597634d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-f4bfz" Dec 06 06:30:14 crc kubenswrapper[4809]: I1206 06:30:14.648046 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e2ad775c-a1f4-4770-918b-4b3ab597634d-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-f4bfz\" (UID: \"e2ad775c-a1f4-4770-918b-4b3ab597634d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-f4bfz" Dec 06 06:30:14 crc kubenswrapper[4809]: I1206 06:30:14.648090 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwz95\" (UniqueName: \"kubernetes.io/projected/e2ad775c-a1f4-4770-918b-4b3ab597634d-kube-api-access-zwz95\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-f4bfz\" (UID: \"e2ad775c-a1f4-4770-918b-4b3ab597634d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-f4bfz" Dec 06 06:30:14 crc kubenswrapper[4809]: I1206 06:30:14.653691 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e2ad775c-a1f4-4770-918b-4b3ab597634d-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-f4bfz\" (UID: \"e2ad775c-a1f4-4770-918b-4b3ab597634d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-f4bfz" Dec 06 06:30:14 crc kubenswrapper[4809]: I1206 06:30:14.654129 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e2ad775c-a1f4-4770-918b-4b3ab597634d-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-f4bfz\" (UID: \"e2ad775c-a1f4-4770-918b-4b3ab597634d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-f4bfz" Dec 06 06:30:14 crc kubenswrapper[4809]: I1206 06:30:14.666195 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwz95\" (UniqueName: \"kubernetes.io/projected/e2ad775c-a1f4-4770-918b-4b3ab597634d-kube-api-access-zwz95\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-f4bfz\" (UID: \"e2ad775c-a1f4-4770-918b-4b3ab597634d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-f4bfz" Dec 06 06:30:14 crc kubenswrapper[4809]: I1206 06:30:14.726678 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-f4bfz" Dec 06 06:30:15 crc kubenswrapper[4809]: I1206 06:30:15.314486 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-f4bfz"] Dec 06 06:30:15 crc kubenswrapper[4809]: I1206 06:30:15.359006 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-f4bfz" event={"ID":"e2ad775c-a1f4-4770-918b-4b3ab597634d","Type":"ContainerStarted","Data":"216a173e01de947b108bb86e0d0192bb821730242943735a9ebe60f667668fa1"} Dec 06 06:30:15 crc kubenswrapper[4809]: I1206 06:30:15.894727 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 06:30:16 crc kubenswrapper[4809]: I1206 06:30:16.372059 4809 generic.go:334] "Generic (PLEG): container finished" podID="2e285912-94a1-441c-83b8-38e561cc3a0b" containerID="c3687439e0df610ad45389e8f932372cc2401fdfdbf223c9e29ffc84291b1444" exitCode=0 Dec 06 06:30:16 crc kubenswrapper[4809]: I1206 06:30:16.372114 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9fjmc" event={"ID":"2e285912-94a1-441c-83b8-38e561cc3a0b","Type":"ContainerDied","Data":"c3687439e0df610ad45389e8f932372cc2401fdfdbf223c9e29ffc84291b1444"} Dec 06 06:30:17 crc kubenswrapper[4809]: I1206 06:30:17.383639 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-f4bfz" event={"ID":"e2ad775c-a1f4-4770-918b-4b3ab597634d","Type":"ContainerStarted","Data":"a88161d956b7d872e75e87f8e8d12cfc2e8457097c198d96cbaeee654e93a52b"} Dec 06 06:30:17 crc kubenswrapper[4809]: I1206 06:30:17.402227 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-f4bfz" podStartSLOduration=2.896380443 podStartE2EDuration="3.402209132s" podCreationTimestamp="2025-12-06 06:30:14 +0000 UTC" firstStartedPulling="2025-12-06 06:30:15.386031232 +0000 UTC m=+2340.275014174" lastFinishedPulling="2025-12-06 06:30:15.891859921 +0000 UTC m=+2340.780842863" observedRunningTime="2025-12-06 06:30:17.399361885 +0000 UTC m=+2342.288344837" watchObservedRunningTime="2025-12-06 06:30:17.402209132 +0000 UTC m=+2342.291192074" Dec 06 06:30:20 crc kubenswrapper[4809]: I1206 06:30:20.429123 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9fjmc" event={"ID":"2e285912-94a1-441c-83b8-38e561cc3a0b","Type":"ContainerStarted","Data":"fe6775ab1e3d8579813fd15b96ee303eec1212d8c66e64da6423882403b6d5e6"} Dec 06 06:30:20 crc kubenswrapper[4809]: I1206 06:30:20.455956 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9fjmc" podStartSLOduration=2.025357551 podStartE2EDuration="8.455914738s" podCreationTimestamp="2025-12-06 06:30:12 +0000 UTC" firstStartedPulling="2025-12-06 06:30:13.327109508 +0000 UTC m=+2338.216092450" lastFinishedPulling="2025-12-06 06:30:19.757666685 +0000 UTC m=+2344.646649637" observedRunningTime="2025-12-06 06:30:20.451288263 +0000 UTC m=+2345.340271205" watchObservedRunningTime="2025-12-06 06:30:20.455914738 +0000 UTC m=+2345.344897680" Dec 06 06:30:22 crc kubenswrapper[4809]: I1206 06:30:22.483327 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9fjmc" Dec 06 06:30:22 crc kubenswrapper[4809]: I1206 06:30:22.484824 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9fjmc" Dec 06 06:30:22 crc kubenswrapper[4809]: I1206 06:30:22.540179 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9fjmc" Dec 06 06:30:23 crc kubenswrapper[4809]: I1206 06:30:23.048641 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-tp9tt"] Dec 06 06:30:23 crc kubenswrapper[4809]: I1206 06:30:23.065771 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-tp9tt"] Dec 06 06:30:23 crc kubenswrapper[4809]: I1206 06:30:23.405411 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e106bab-59d8-4d2e-89f9-1f46e9b37f43" path="/var/lib/kubelet/pods/4e106bab-59d8-4d2e-89f9-1f46e9b37f43/volumes" Dec 06 06:30:30 crc kubenswrapper[4809]: I1206 06:30:30.989012 4809 scope.go:117] "RemoveContainer" containerID="ee227d98802b9bb43d7d6c3599870ec807f20cfa515b0f11e3a0a8d474c3cb99" Dec 06 06:30:31 crc kubenswrapper[4809]: I1206 06:30:31.043679 4809 scope.go:117] "RemoveContainer" containerID="0c398009cd796b9757f33073f5ae46f28d2c1cb725130ab094f4dd9d64918527" Dec 06 06:30:31 crc kubenswrapper[4809]: I1206 06:30:31.094970 4809 scope.go:117] "RemoveContainer" containerID="7ba2c355548a958343d1e97f6c244b4c1cb70b0083c1fb59afb79e37ce2ee82c" Dec 06 06:30:31 crc kubenswrapper[4809]: I1206 06:30:31.175314 4809 scope.go:117] "RemoveContainer" containerID="e31814af765d0bec9a5bad864050f9b02bb215972eb7f9ffc2d5fba01a2f13d8" Dec 06 06:30:32 crc kubenswrapper[4809]: I1206 06:30:32.550623 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9fjmc" Dec 06 06:30:32 crc kubenswrapper[4809]: I1206 06:30:32.611716 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9fjmc"] Dec 06 06:30:32 crc kubenswrapper[4809]: I1206 06:30:32.612530 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9fjmc" podUID="2e285912-94a1-441c-83b8-38e561cc3a0b" containerName="registry-server" containerID="cri-o://fe6775ab1e3d8579813fd15b96ee303eec1212d8c66e64da6423882403b6d5e6" gracePeriod=2 Dec 06 06:30:33 crc kubenswrapper[4809]: I1206 06:30:33.577083 4809 generic.go:334] "Generic (PLEG): container finished" podID="2e285912-94a1-441c-83b8-38e561cc3a0b" containerID="fe6775ab1e3d8579813fd15b96ee303eec1212d8c66e64da6423882403b6d5e6" exitCode=0 Dec 06 06:30:33 crc kubenswrapper[4809]: I1206 06:30:33.577153 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9fjmc" event={"ID":"2e285912-94a1-441c-83b8-38e561cc3a0b","Type":"ContainerDied","Data":"fe6775ab1e3d8579813fd15b96ee303eec1212d8c66e64da6423882403b6d5e6"} Dec 06 06:30:34 crc kubenswrapper[4809]: I1206 06:30:34.473995 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9fjmc" Dec 06 06:30:34 crc kubenswrapper[4809]: I1206 06:30:34.496074 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:30:34 crc kubenswrapper[4809]: I1206 06:30:34.496145 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:30:34 crc kubenswrapper[4809]: I1206 06:30:34.496206 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-npms2" Dec 06 06:30:34 crc kubenswrapper[4809]: I1206 06:30:34.497536 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3aa8c940468b111de41fd7720fd216eebfc0bb377ed933226628430658845c32"} pod="openshift-machine-config-operator/machine-config-daemon-npms2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 06:30:34 crc kubenswrapper[4809]: I1206 06:30:34.497634 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" containerID="cri-o://3aa8c940468b111de41fd7720fd216eebfc0bb377ed933226628430658845c32" gracePeriod=600 Dec 06 06:30:34 crc kubenswrapper[4809]: I1206 06:30:34.595770 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9fjmc" event={"ID":"2e285912-94a1-441c-83b8-38e561cc3a0b","Type":"ContainerDied","Data":"bf37377d742fa5d7b680d257dea75ee608835af54eb0bdd86938258507faafb3"} Dec 06 06:30:34 crc kubenswrapper[4809]: I1206 06:30:34.596181 4809 scope.go:117] "RemoveContainer" containerID="fe6775ab1e3d8579813fd15b96ee303eec1212d8c66e64da6423882403b6d5e6" Dec 06 06:30:34 crc kubenswrapper[4809]: I1206 06:30:34.595866 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9fjmc" Dec 06 06:30:34 crc kubenswrapper[4809]: I1206 06:30:34.619779 4809 scope.go:117] "RemoveContainer" containerID="c3687439e0df610ad45389e8f932372cc2401fdfdbf223c9e29ffc84291b1444" Dec 06 06:30:34 crc kubenswrapper[4809]: I1206 06:30:34.657707 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e285912-94a1-441c-83b8-38e561cc3a0b-utilities\") pod \"2e285912-94a1-441c-83b8-38e561cc3a0b\" (UID: \"2e285912-94a1-441c-83b8-38e561cc3a0b\") " Dec 06 06:30:34 crc kubenswrapper[4809]: I1206 06:30:34.657773 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5sj6s\" (UniqueName: \"kubernetes.io/projected/2e285912-94a1-441c-83b8-38e561cc3a0b-kube-api-access-5sj6s\") pod \"2e285912-94a1-441c-83b8-38e561cc3a0b\" (UID: \"2e285912-94a1-441c-83b8-38e561cc3a0b\") " Dec 06 06:30:34 crc kubenswrapper[4809]: I1206 06:30:34.657920 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e285912-94a1-441c-83b8-38e561cc3a0b-catalog-content\") pod \"2e285912-94a1-441c-83b8-38e561cc3a0b\" (UID: \"2e285912-94a1-441c-83b8-38e561cc3a0b\") " Dec 06 06:30:34 crc kubenswrapper[4809]: I1206 06:30:34.659124 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e285912-94a1-441c-83b8-38e561cc3a0b-utilities" (OuterVolumeSpecName: "utilities") pod "2e285912-94a1-441c-83b8-38e561cc3a0b" (UID: "2e285912-94a1-441c-83b8-38e561cc3a0b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:30:34 crc kubenswrapper[4809]: I1206 06:30:34.665124 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e285912-94a1-441c-83b8-38e561cc3a0b-kube-api-access-5sj6s" (OuterVolumeSpecName: "kube-api-access-5sj6s") pod "2e285912-94a1-441c-83b8-38e561cc3a0b" (UID: "2e285912-94a1-441c-83b8-38e561cc3a0b"). InnerVolumeSpecName "kube-api-access-5sj6s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:30:34 crc kubenswrapper[4809]: I1206 06:30:34.679439 4809 scope.go:117] "RemoveContainer" containerID="08dda443afeecc79da20d92093535d9a594dca97c320b2bc950371f59c2a808a" Dec 06 06:30:34 crc kubenswrapper[4809]: I1206 06:30:34.684054 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e285912-94a1-441c-83b8-38e561cc3a0b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2e285912-94a1-441c-83b8-38e561cc3a0b" (UID: "2e285912-94a1-441c-83b8-38e561cc3a0b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:30:34 crc kubenswrapper[4809]: I1206 06:30:34.761141 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e285912-94a1-441c-83b8-38e561cc3a0b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:30:34 crc kubenswrapper[4809]: I1206 06:30:34.761178 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e285912-94a1-441c-83b8-38e561cc3a0b-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:30:34 crc kubenswrapper[4809]: I1206 06:30:34.761197 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5sj6s\" (UniqueName: \"kubernetes.io/projected/2e285912-94a1-441c-83b8-38e561cc3a0b-kube-api-access-5sj6s\") on node \"crc\" DevicePath \"\"" Dec 06 06:30:34 crc kubenswrapper[4809]: I1206 06:30:34.939454 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9fjmc"] Dec 06 06:30:34 crc kubenswrapper[4809]: I1206 06:30:34.948269 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9fjmc"] Dec 06 06:30:35 crc kubenswrapper[4809]: I1206 06:30:35.409590 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e285912-94a1-441c-83b8-38e561cc3a0b" path="/var/lib/kubelet/pods/2e285912-94a1-441c-83b8-38e561cc3a0b/volumes" Dec 06 06:30:35 crc kubenswrapper[4809]: I1206 06:30:35.617128 4809 generic.go:334] "Generic (PLEG): container finished" podID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerID="3aa8c940468b111de41fd7720fd216eebfc0bb377ed933226628430658845c32" exitCode=0 Dec 06 06:30:35 crc kubenswrapper[4809]: I1206 06:30:35.617199 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" event={"ID":"cbf4cf62-024e-4703-a8b8-9aecda9cd26a","Type":"ContainerDied","Data":"3aa8c940468b111de41fd7720fd216eebfc0bb377ed933226628430658845c32"} Dec 06 06:30:35 crc kubenswrapper[4809]: I1206 06:30:35.617258 4809 scope.go:117] "RemoveContainer" containerID="54ba3d8af898b6b995ee9173b17fafab2449b8d6a7b61c005c29dabf843d35fe" Dec 06 06:30:36 crc kubenswrapper[4809]: E1206 06:30:36.568597 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:30:36 crc kubenswrapper[4809]: I1206 06:30:36.633091 4809 scope.go:117] "RemoveContainer" containerID="3aa8c940468b111de41fd7720fd216eebfc0bb377ed933226628430658845c32" Dec 06 06:30:36 crc kubenswrapper[4809]: E1206 06:30:36.633533 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:30:51 crc kubenswrapper[4809]: I1206 06:30:51.388597 4809 scope.go:117] "RemoveContainer" containerID="3aa8c940468b111de41fd7720fd216eebfc0bb377ed933226628430658845c32" Dec 06 06:30:51 crc kubenswrapper[4809]: E1206 06:30:51.389371 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:31:03 crc kubenswrapper[4809]: I1206 06:31:03.390247 4809 scope.go:117] "RemoveContainer" containerID="3aa8c940468b111de41fd7720fd216eebfc0bb377ed933226628430658845c32" Dec 06 06:31:03 crc kubenswrapper[4809]: E1206 06:31:03.391321 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:31:16 crc kubenswrapper[4809]: I1206 06:31:16.389218 4809 scope.go:117] "RemoveContainer" containerID="3aa8c940468b111de41fd7720fd216eebfc0bb377ed933226628430658845c32" Dec 06 06:31:16 crc kubenswrapper[4809]: E1206 06:31:16.390073 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:31:29 crc kubenswrapper[4809]: I1206 06:31:29.390213 4809 scope.go:117] "RemoveContainer" containerID="3aa8c940468b111de41fd7720fd216eebfc0bb377ed933226628430658845c32" Dec 06 06:31:29 crc kubenswrapper[4809]: E1206 06:31:29.391631 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:31:31 crc kubenswrapper[4809]: I1206 06:31:31.329686 4809 scope.go:117] "RemoveContainer" containerID="5a761d93ef4023d6a8f1aff0ac7aacb4f222ad1ae8d7014a514fb541b94eb75c" Dec 06 06:31:31 crc kubenswrapper[4809]: I1206 06:31:31.370155 4809 scope.go:117] "RemoveContainer" containerID="b4e56e00c58894cde7141cbdf58ce75924f9471af818d37e164cefd772ddb6e0" Dec 06 06:31:31 crc kubenswrapper[4809]: I1206 06:31:31.460359 4809 scope.go:117] "RemoveContainer" containerID="1c30484f54c1506391371b7860a5e6bd84e7ecb6d1a59a579217aed8e18fd7db" Dec 06 06:31:33 crc kubenswrapper[4809]: I1206 06:31:33.361242 4809 generic.go:334] "Generic (PLEG): container finished" podID="e2ad775c-a1f4-4770-918b-4b3ab597634d" containerID="a88161d956b7d872e75e87f8e8d12cfc2e8457097c198d96cbaeee654e93a52b" exitCode=0 Dec 06 06:31:33 crc kubenswrapper[4809]: I1206 06:31:33.361275 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-f4bfz" event={"ID":"e2ad775c-a1f4-4770-918b-4b3ab597634d","Type":"ContainerDied","Data":"a88161d956b7d872e75e87f8e8d12cfc2e8457097c198d96cbaeee654e93a52b"} Dec 06 06:31:34 crc kubenswrapper[4809]: I1206 06:31:34.844404 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-f4bfz" Dec 06 06:31:34 crc kubenswrapper[4809]: I1206 06:31:34.974161 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e2ad775c-a1f4-4770-918b-4b3ab597634d-ssh-key\") pod \"e2ad775c-a1f4-4770-918b-4b3ab597634d\" (UID: \"e2ad775c-a1f4-4770-918b-4b3ab597634d\") " Dec 06 06:31:34 crc kubenswrapper[4809]: I1206 06:31:34.974232 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwz95\" (UniqueName: \"kubernetes.io/projected/e2ad775c-a1f4-4770-918b-4b3ab597634d-kube-api-access-zwz95\") pod \"e2ad775c-a1f4-4770-918b-4b3ab597634d\" (UID: \"e2ad775c-a1f4-4770-918b-4b3ab597634d\") " Dec 06 06:31:34 crc kubenswrapper[4809]: I1206 06:31:34.974354 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e2ad775c-a1f4-4770-918b-4b3ab597634d-inventory\") pod \"e2ad775c-a1f4-4770-918b-4b3ab597634d\" (UID: \"e2ad775c-a1f4-4770-918b-4b3ab597634d\") " Dec 06 06:31:34 crc kubenswrapper[4809]: I1206 06:31:34.981677 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2ad775c-a1f4-4770-918b-4b3ab597634d-kube-api-access-zwz95" (OuterVolumeSpecName: "kube-api-access-zwz95") pod "e2ad775c-a1f4-4770-918b-4b3ab597634d" (UID: "e2ad775c-a1f4-4770-918b-4b3ab597634d"). InnerVolumeSpecName "kube-api-access-zwz95". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:31:35 crc kubenswrapper[4809]: I1206 06:31:35.022173 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2ad775c-a1f4-4770-918b-4b3ab597634d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e2ad775c-a1f4-4770-918b-4b3ab597634d" (UID: "e2ad775c-a1f4-4770-918b-4b3ab597634d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:31:35 crc kubenswrapper[4809]: I1206 06:31:35.028251 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2ad775c-a1f4-4770-918b-4b3ab597634d-inventory" (OuterVolumeSpecName: "inventory") pod "e2ad775c-a1f4-4770-918b-4b3ab597634d" (UID: "e2ad775c-a1f4-4770-918b-4b3ab597634d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:31:35 crc kubenswrapper[4809]: I1206 06:31:35.079852 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e2ad775c-a1f4-4770-918b-4b3ab597634d-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 06:31:35 crc kubenswrapper[4809]: I1206 06:31:35.079888 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwz95\" (UniqueName: \"kubernetes.io/projected/e2ad775c-a1f4-4770-918b-4b3ab597634d-kube-api-access-zwz95\") on node \"crc\" DevicePath \"\"" Dec 06 06:31:35 crc kubenswrapper[4809]: I1206 06:31:35.079902 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e2ad775c-a1f4-4770-918b-4b3ab597634d-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 06:31:35 crc kubenswrapper[4809]: I1206 06:31:35.451434 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-f4bfz" event={"ID":"e2ad775c-a1f4-4770-918b-4b3ab597634d","Type":"ContainerDied","Data":"216a173e01de947b108bb86e0d0192bb821730242943735a9ebe60f667668fa1"} Dec 06 06:31:35 crc kubenswrapper[4809]: I1206 06:31:35.451760 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-f4bfz" Dec 06 06:31:35 crc kubenswrapper[4809]: I1206 06:31:35.451921 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="216a173e01de947b108bb86e0d0192bb821730242943735a9ebe60f667668fa1" Dec 06 06:31:35 crc kubenswrapper[4809]: I1206 06:31:35.508607 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jzdth"] Dec 06 06:31:35 crc kubenswrapper[4809]: E1206 06:31:35.509350 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2ad775c-a1f4-4770-918b-4b3ab597634d" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 06 06:31:35 crc kubenswrapper[4809]: I1206 06:31:35.509378 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2ad775c-a1f4-4770-918b-4b3ab597634d" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 06 06:31:35 crc kubenswrapper[4809]: E1206 06:31:35.509392 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e285912-94a1-441c-83b8-38e561cc3a0b" containerName="registry-server" Dec 06 06:31:35 crc kubenswrapper[4809]: I1206 06:31:35.509401 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e285912-94a1-441c-83b8-38e561cc3a0b" containerName="registry-server" Dec 06 06:31:35 crc kubenswrapper[4809]: E1206 06:31:35.509429 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e285912-94a1-441c-83b8-38e561cc3a0b" containerName="extract-utilities" Dec 06 06:31:35 crc kubenswrapper[4809]: I1206 06:31:35.509440 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e285912-94a1-441c-83b8-38e561cc3a0b" containerName="extract-utilities" Dec 06 06:31:35 crc kubenswrapper[4809]: E1206 06:31:35.509482 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e285912-94a1-441c-83b8-38e561cc3a0b" containerName="extract-content" Dec 06 06:31:35 crc kubenswrapper[4809]: I1206 06:31:35.509490 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e285912-94a1-441c-83b8-38e561cc3a0b" containerName="extract-content" Dec 06 06:31:35 crc kubenswrapper[4809]: I1206 06:31:35.509848 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2ad775c-a1f4-4770-918b-4b3ab597634d" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 06 06:31:35 crc kubenswrapper[4809]: I1206 06:31:35.509905 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e285912-94a1-441c-83b8-38e561cc3a0b" containerName="registry-server" Dec 06 06:31:35 crc kubenswrapper[4809]: I1206 06:31:35.511176 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jzdth" Dec 06 06:31:35 crc kubenswrapper[4809]: I1206 06:31:35.514602 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-bzlf5" Dec 06 06:31:35 crc kubenswrapper[4809]: I1206 06:31:35.514640 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 06:31:35 crc kubenswrapper[4809]: I1206 06:31:35.514865 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 06:31:35 crc kubenswrapper[4809]: I1206 06:31:35.514970 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 06:31:35 crc kubenswrapper[4809]: I1206 06:31:35.540730 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jzdth"] Dec 06 06:31:35 crc kubenswrapper[4809]: I1206 06:31:35.601244 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d24e0d36-0864-4ecc-9e52-f4998a338776-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-jzdth\" (UID: \"d24e0d36-0864-4ecc-9e52-f4998a338776\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jzdth" Dec 06 06:31:35 crc kubenswrapper[4809]: I1206 06:31:35.601369 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jcd7l\" (UniqueName: \"kubernetes.io/projected/d24e0d36-0864-4ecc-9e52-f4998a338776-kube-api-access-jcd7l\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-jzdth\" (UID: \"d24e0d36-0864-4ecc-9e52-f4998a338776\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jzdth" Dec 06 06:31:35 crc kubenswrapper[4809]: I1206 06:31:35.601489 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d24e0d36-0864-4ecc-9e52-f4998a338776-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-jzdth\" (UID: \"d24e0d36-0864-4ecc-9e52-f4998a338776\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jzdth" Dec 06 06:31:35 crc kubenswrapper[4809]: I1206 06:31:35.704022 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jcd7l\" (UniqueName: \"kubernetes.io/projected/d24e0d36-0864-4ecc-9e52-f4998a338776-kube-api-access-jcd7l\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-jzdth\" (UID: \"d24e0d36-0864-4ecc-9e52-f4998a338776\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jzdth" Dec 06 06:31:35 crc kubenswrapper[4809]: I1206 06:31:35.704171 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d24e0d36-0864-4ecc-9e52-f4998a338776-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-jzdth\" (UID: \"d24e0d36-0864-4ecc-9e52-f4998a338776\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jzdth" Dec 06 06:31:35 crc kubenswrapper[4809]: I1206 06:31:35.704281 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d24e0d36-0864-4ecc-9e52-f4998a338776-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-jzdth\" (UID: \"d24e0d36-0864-4ecc-9e52-f4998a338776\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jzdth" Dec 06 06:31:35 crc kubenswrapper[4809]: I1206 06:31:35.709916 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d24e0d36-0864-4ecc-9e52-f4998a338776-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-jzdth\" (UID: \"d24e0d36-0864-4ecc-9e52-f4998a338776\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jzdth" Dec 06 06:31:35 crc kubenswrapper[4809]: I1206 06:31:35.714356 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d24e0d36-0864-4ecc-9e52-f4998a338776-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-jzdth\" (UID: \"d24e0d36-0864-4ecc-9e52-f4998a338776\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jzdth" Dec 06 06:31:36 crc kubenswrapper[4809]: I1206 06:31:36.564136 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jcd7l\" (UniqueName: \"kubernetes.io/projected/d24e0d36-0864-4ecc-9e52-f4998a338776-kube-api-access-jcd7l\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-jzdth\" (UID: \"d24e0d36-0864-4ecc-9e52-f4998a338776\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jzdth" Dec 06 06:31:36 crc kubenswrapper[4809]: I1206 06:31:36.728974 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jzdth" Dec 06 06:31:37 crc kubenswrapper[4809]: I1206 06:31:37.446538 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jzdth"] Dec 06 06:31:37 crc kubenswrapper[4809]: I1206 06:31:37.485122 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jzdth" event={"ID":"d24e0d36-0864-4ecc-9e52-f4998a338776","Type":"ContainerStarted","Data":"14fd67cbb6f685c52c3ed2c92374d3d76224f90610b37936d6a1118f2fe5b853"} Dec 06 06:31:43 crc kubenswrapper[4809]: I1206 06:31:43.388964 4809 scope.go:117] "RemoveContainer" containerID="3aa8c940468b111de41fd7720fd216eebfc0bb377ed933226628430658845c32" Dec 06 06:31:43 crc kubenswrapper[4809]: E1206 06:31:43.390325 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:31:55 crc kubenswrapper[4809]: I1206 06:31:55.713176 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jzdth" event={"ID":"d24e0d36-0864-4ecc-9e52-f4998a338776","Type":"ContainerStarted","Data":"d82b853bd09452c4c572b471c2c725db7478c193280e5402018c22978e849b8e"} Dec 06 06:31:55 crc kubenswrapper[4809]: I1206 06:31:55.734770 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jzdth" podStartSLOduration=3.031813098 podStartE2EDuration="20.734752897s" podCreationTimestamp="2025-12-06 06:31:35 +0000 UTC" firstStartedPulling="2025-12-06 06:31:37.451920942 +0000 UTC m=+2422.340903884" lastFinishedPulling="2025-12-06 06:31:55.154860741 +0000 UTC m=+2440.043843683" observedRunningTime="2025-12-06 06:31:55.730575464 +0000 UTC m=+2440.619558416" watchObservedRunningTime="2025-12-06 06:31:55.734752897 +0000 UTC m=+2440.623735839" Dec 06 06:31:56 crc kubenswrapper[4809]: I1206 06:31:56.388510 4809 scope.go:117] "RemoveContainer" containerID="3aa8c940468b111de41fd7720fd216eebfc0bb377ed933226628430658845c32" Dec 06 06:31:56 crc kubenswrapper[4809]: E1206 06:31:56.388832 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:31:59 crc kubenswrapper[4809]: E1206 06:31:59.158135 4809 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.771s" Dec 06 06:32:01 crc kubenswrapper[4809]: I1206 06:32:01.786202 4809 generic.go:334] "Generic (PLEG): container finished" podID="d24e0d36-0864-4ecc-9e52-f4998a338776" containerID="d82b853bd09452c4c572b471c2c725db7478c193280e5402018c22978e849b8e" exitCode=0 Dec 06 06:32:01 crc kubenswrapper[4809]: I1206 06:32:01.786312 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jzdth" event={"ID":"d24e0d36-0864-4ecc-9e52-f4998a338776","Type":"ContainerDied","Data":"d82b853bd09452c4c572b471c2c725db7478c193280e5402018c22978e849b8e"} Dec 06 06:32:03 crc kubenswrapper[4809]: I1206 06:32:03.366862 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jzdth" Dec 06 06:32:03 crc kubenswrapper[4809]: I1206 06:32:03.371822 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d24e0d36-0864-4ecc-9e52-f4998a338776-inventory\") pod \"d24e0d36-0864-4ecc-9e52-f4998a338776\" (UID: \"d24e0d36-0864-4ecc-9e52-f4998a338776\") " Dec 06 06:32:03 crc kubenswrapper[4809]: I1206 06:32:03.371876 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jcd7l\" (UniqueName: \"kubernetes.io/projected/d24e0d36-0864-4ecc-9e52-f4998a338776-kube-api-access-jcd7l\") pod \"d24e0d36-0864-4ecc-9e52-f4998a338776\" (UID: \"d24e0d36-0864-4ecc-9e52-f4998a338776\") " Dec 06 06:32:03 crc kubenswrapper[4809]: I1206 06:32:03.371976 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d24e0d36-0864-4ecc-9e52-f4998a338776-ssh-key\") pod \"d24e0d36-0864-4ecc-9e52-f4998a338776\" (UID: \"d24e0d36-0864-4ecc-9e52-f4998a338776\") " Dec 06 06:32:03 crc kubenswrapper[4809]: I1206 06:32:03.377685 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d24e0d36-0864-4ecc-9e52-f4998a338776-kube-api-access-jcd7l" (OuterVolumeSpecName: "kube-api-access-jcd7l") pod "d24e0d36-0864-4ecc-9e52-f4998a338776" (UID: "d24e0d36-0864-4ecc-9e52-f4998a338776"). InnerVolumeSpecName "kube-api-access-jcd7l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:32:03 crc kubenswrapper[4809]: I1206 06:32:03.449957 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d24e0d36-0864-4ecc-9e52-f4998a338776-inventory" (OuterVolumeSpecName: "inventory") pod "d24e0d36-0864-4ecc-9e52-f4998a338776" (UID: "d24e0d36-0864-4ecc-9e52-f4998a338776"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:32:03 crc kubenswrapper[4809]: I1206 06:32:03.452241 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d24e0d36-0864-4ecc-9e52-f4998a338776-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d24e0d36-0864-4ecc-9e52-f4998a338776" (UID: "d24e0d36-0864-4ecc-9e52-f4998a338776"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:32:03 crc kubenswrapper[4809]: I1206 06:32:03.474745 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d24e0d36-0864-4ecc-9e52-f4998a338776-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 06:32:03 crc kubenswrapper[4809]: I1206 06:32:03.474788 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jcd7l\" (UniqueName: \"kubernetes.io/projected/d24e0d36-0864-4ecc-9e52-f4998a338776-kube-api-access-jcd7l\") on node \"crc\" DevicePath \"\"" Dec 06 06:32:03 crc kubenswrapper[4809]: I1206 06:32:03.474803 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d24e0d36-0864-4ecc-9e52-f4998a338776-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 06:32:03 crc kubenswrapper[4809]: I1206 06:32:03.806918 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jzdth" event={"ID":"d24e0d36-0864-4ecc-9e52-f4998a338776","Type":"ContainerDied","Data":"14fd67cbb6f685c52c3ed2c92374d3d76224f90610b37936d6a1118f2fe5b853"} Dec 06 06:32:03 crc kubenswrapper[4809]: I1206 06:32:03.806970 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="14fd67cbb6f685c52c3ed2c92374d3d76224f90610b37936d6a1118f2fe5b853" Dec 06 06:32:03 crc kubenswrapper[4809]: I1206 06:32:03.806990 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jzdth" Dec 06 06:32:03 crc kubenswrapper[4809]: I1206 06:32:03.890127 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-p4jjg"] Dec 06 06:32:03 crc kubenswrapper[4809]: E1206 06:32:03.890725 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d24e0d36-0864-4ecc-9e52-f4998a338776" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 06 06:32:03 crc kubenswrapper[4809]: I1206 06:32:03.890746 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d24e0d36-0864-4ecc-9e52-f4998a338776" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 06 06:32:03 crc kubenswrapper[4809]: I1206 06:32:03.891046 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="d24e0d36-0864-4ecc-9e52-f4998a338776" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 06 06:32:03 crc kubenswrapper[4809]: I1206 06:32:03.891977 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-p4jjg" Dec 06 06:32:03 crc kubenswrapper[4809]: I1206 06:32:03.894306 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 06:32:03 crc kubenswrapper[4809]: I1206 06:32:03.894732 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 06:32:03 crc kubenswrapper[4809]: I1206 06:32:03.895110 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-bzlf5" Dec 06 06:32:03 crc kubenswrapper[4809]: I1206 06:32:03.896724 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 06:32:03 crc kubenswrapper[4809]: I1206 06:32:03.900603 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-p4jjg"] Dec 06 06:32:04 crc kubenswrapper[4809]: I1206 06:32:04.088794 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c492d1c2-3ee8-491a-a237-98691bfacec6-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-p4jjg\" (UID: \"c492d1c2-3ee8-491a-a237-98691bfacec6\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-p4jjg" Dec 06 06:32:04 crc kubenswrapper[4809]: I1206 06:32:04.088875 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c492d1c2-3ee8-491a-a237-98691bfacec6-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-p4jjg\" (UID: \"c492d1c2-3ee8-491a-a237-98691bfacec6\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-p4jjg" Dec 06 06:32:04 crc kubenswrapper[4809]: I1206 06:32:04.089202 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nntn8\" (UniqueName: \"kubernetes.io/projected/c492d1c2-3ee8-491a-a237-98691bfacec6-kube-api-access-nntn8\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-p4jjg\" (UID: \"c492d1c2-3ee8-491a-a237-98691bfacec6\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-p4jjg" Dec 06 06:32:04 crc kubenswrapper[4809]: I1206 06:32:04.192849 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nntn8\" (UniqueName: \"kubernetes.io/projected/c492d1c2-3ee8-491a-a237-98691bfacec6-kube-api-access-nntn8\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-p4jjg\" (UID: \"c492d1c2-3ee8-491a-a237-98691bfacec6\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-p4jjg" Dec 06 06:32:04 crc kubenswrapper[4809]: I1206 06:32:04.193028 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c492d1c2-3ee8-491a-a237-98691bfacec6-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-p4jjg\" (UID: \"c492d1c2-3ee8-491a-a237-98691bfacec6\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-p4jjg" Dec 06 06:32:04 crc kubenswrapper[4809]: I1206 06:32:04.193091 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c492d1c2-3ee8-491a-a237-98691bfacec6-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-p4jjg\" (UID: \"c492d1c2-3ee8-491a-a237-98691bfacec6\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-p4jjg" Dec 06 06:32:04 crc kubenswrapper[4809]: I1206 06:32:04.197922 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c492d1c2-3ee8-491a-a237-98691bfacec6-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-p4jjg\" (UID: \"c492d1c2-3ee8-491a-a237-98691bfacec6\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-p4jjg" Dec 06 06:32:04 crc kubenswrapper[4809]: I1206 06:32:04.197953 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c492d1c2-3ee8-491a-a237-98691bfacec6-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-p4jjg\" (UID: \"c492d1c2-3ee8-491a-a237-98691bfacec6\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-p4jjg" Dec 06 06:32:04 crc kubenswrapper[4809]: I1206 06:32:04.214523 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nntn8\" (UniqueName: \"kubernetes.io/projected/c492d1c2-3ee8-491a-a237-98691bfacec6-kube-api-access-nntn8\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-p4jjg\" (UID: \"c492d1c2-3ee8-491a-a237-98691bfacec6\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-p4jjg" Dec 06 06:32:04 crc kubenswrapper[4809]: I1206 06:32:04.218495 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-p4jjg" Dec 06 06:32:04 crc kubenswrapper[4809]: I1206 06:32:04.672455 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-p4jjg"] Dec 06 06:32:04 crc kubenswrapper[4809]: I1206 06:32:04.816654 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-p4jjg" event={"ID":"c492d1c2-3ee8-491a-a237-98691bfacec6","Type":"ContainerStarted","Data":"994b3068a66670886b3dbab8ffe784065afb7654f40e274c0e43264293a097ec"} Dec 06 06:32:06 crc kubenswrapper[4809]: I1206 06:32:06.844871 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-p4jjg" event={"ID":"c492d1c2-3ee8-491a-a237-98691bfacec6","Type":"ContainerStarted","Data":"ac011c30238ea1991971e3d0bb41f7bf0697c120fd2d109ca071130abb3e5414"} Dec 06 06:32:06 crc kubenswrapper[4809]: I1206 06:32:06.875653 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-p4jjg" podStartSLOduration=2.444561094 podStartE2EDuration="3.875627493s" podCreationTimestamp="2025-12-06 06:32:03 +0000 UTC" firstStartedPulling="2025-12-06 06:32:04.67565659 +0000 UTC m=+2449.564639532" lastFinishedPulling="2025-12-06 06:32:06.106722989 +0000 UTC m=+2450.995705931" observedRunningTime="2025-12-06 06:32:06.865013116 +0000 UTC m=+2451.753996068" watchObservedRunningTime="2025-12-06 06:32:06.875627493 +0000 UTC m=+2451.764610455" Dec 06 06:32:08 crc kubenswrapper[4809]: I1206 06:32:08.388825 4809 scope.go:117] "RemoveContainer" containerID="3aa8c940468b111de41fd7720fd216eebfc0bb377ed933226628430658845c32" Dec 06 06:32:08 crc kubenswrapper[4809]: E1206 06:32:08.389453 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:32:22 crc kubenswrapper[4809]: I1206 06:32:22.389239 4809 scope.go:117] "RemoveContainer" containerID="3aa8c940468b111de41fd7720fd216eebfc0bb377ed933226628430658845c32" Dec 06 06:32:22 crc kubenswrapper[4809]: E1206 06:32:22.390096 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:32:33 crc kubenswrapper[4809]: I1206 06:32:33.390822 4809 scope.go:117] "RemoveContainer" containerID="3aa8c940468b111de41fd7720fd216eebfc0bb377ed933226628430658845c32" Dec 06 06:32:33 crc kubenswrapper[4809]: E1206 06:32:33.391520 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:32:42 crc kubenswrapper[4809]: I1206 06:32:42.047337 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-ng4qq"] Dec 06 06:32:42 crc kubenswrapper[4809]: I1206 06:32:42.057992 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-ng4qq"] Dec 06 06:32:43 crc kubenswrapper[4809]: I1206 06:32:43.402419 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18fa982c-0572-4702-8c51-f0f7af7bf52f" path="/var/lib/kubelet/pods/18fa982c-0572-4702-8c51-f0f7af7bf52f/volumes" Dec 06 06:32:45 crc kubenswrapper[4809]: I1206 06:32:45.239186 4809 generic.go:334] "Generic (PLEG): container finished" podID="c492d1c2-3ee8-491a-a237-98691bfacec6" containerID="ac011c30238ea1991971e3d0bb41f7bf0697c120fd2d109ca071130abb3e5414" exitCode=0 Dec 06 06:32:45 crc kubenswrapper[4809]: I1206 06:32:45.239282 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-p4jjg" event={"ID":"c492d1c2-3ee8-491a-a237-98691bfacec6","Type":"ContainerDied","Data":"ac011c30238ea1991971e3d0bb41f7bf0697c120fd2d109ca071130abb3e5414"} Dec 06 06:32:46 crc kubenswrapper[4809]: I1206 06:32:46.389199 4809 scope.go:117] "RemoveContainer" containerID="3aa8c940468b111de41fd7720fd216eebfc0bb377ed933226628430658845c32" Dec 06 06:32:46 crc kubenswrapper[4809]: E1206 06:32:46.389753 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:32:46 crc kubenswrapper[4809]: I1206 06:32:46.789875 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-p4jjg" Dec 06 06:32:46 crc kubenswrapper[4809]: I1206 06:32:46.915171 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nntn8\" (UniqueName: \"kubernetes.io/projected/c492d1c2-3ee8-491a-a237-98691bfacec6-kube-api-access-nntn8\") pod \"c492d1c2-3ee8-491a-a237-98691bfacec6\" (UID: \"c492d1c2-3ee8-491a-a237-98691bfacec6\") " Dec 06 06:32:46 crc kubenswrapper[4809]: I1206 06:32:46.915522 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c492d1c2-3ee8-491a-a237-98691bfacec6-ssh-key\") pod \"c492d1c2-3ee8-491a-a237-98691bfacec6\" (UID: \"c492d1c2-3ee8-491a-a237-98691bfacec6\") " Dec 06 06:32:46 crc kubenswrapper[4809]: I1206 06:32:46.915835 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c492d1c2-3ee8-491a-a237-98691bfacec6-inventory\") pod \"c492d1c2-3ee8-491a-a237-98691bfacec6\" (UID: \"c492d1c2-3ee8-491a-a237-98691bfacec6\") " Dec 06 06:32:46 crc kubenswrapper[4809]: I1206 06:32:46.923400 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c492d1c2-3ee8-491a-a237-98691bfacec6-kube-api-access-nntn8" (OuterVolumeSpecName: "kube-api-access-nntn8") pod "c492d1c2-3ee8-491a-a237-98691bfacec6" (UID: "c492d1c2-3ee8-491a-a237-98691bfacec6"). InnerVolumeSpecName "kube-api-access-nntn8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:32:46 crc kubenswrapper[4809]: I1206 06:32:46.956148 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c492d1c2-3ee8-491a-a237-98691bfacec6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c492d1c2-3ee8-491a-a237-98691bfacec6" (UID: "c492d1c2-3ee8-491a-a237-98691bfacec6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:32:46 crc kubenswrapper[4809]: I1206 06:32:46.957314 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c492d1c2-3ee8-491a-a237-98691bfacec6-inventory" (OuterVolumeSpecName: "inventory") pod "c492d1c2-3ee8-491a-a237-98691bfacec6" (UID: "c492d1c2-3ee8-491a-a237-98691bfacec6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:32:47 crc kubenswrapper[4809]: I1206 06:32:47.018326 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nntn8\" (UniqueName: \"kubernetes.io/projected/c492d1c2-3ee8-491a-a237-98691bfacec6-kube-api-access-nntn8\") on node \"crc\" DevicePath \"\"" Dec 06 06:32:47 crc kubenswrapper[4809]: I1206 06:32:47.018372 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c492d1c2-3ee8-491a-a237-98691bfacec6-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 06:32:47 crc kubenswrapper[4809]: I1206 06:32:47.018381 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c492d1c2-3ee8-491a-a237-98691bfacec6-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 06:32:47 crc kubenswrapper[4809]: I1206 06:32:47.292344 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-p4jjg" event={"ID":"c492d1c2-3ee8-491a-a237-98691bfacec6","Type":"ContainerDied","Data":"994b3068a66670886b3dbab8ffe784065afb7654f40e274c0e43264293a097ec"} Dec 06 06:32:47 crc kubenswrapper[4809]: I1206 06:32:47.292391 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="994b3068a66670886b3dbab8ffe784065afb7654f40e274c0e43264293a097ec" Dec 06 06:32:47 crc kubenswrapper[4809]: I1206 06:32:47.292554 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-p4jjg" Dec 06 06:32:47 crc kubenswrapper[4809]: I1206 06:32:47.357873 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6zmrk"] Dec 06 06:32:47 crc kubenswrapper[4809]: E1206 06:32:47.358827 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c492d1c2-3ee8-491a-a237-98691bfacec6" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 06 06:32:47 crc kubenswrapper[4809]: I1206 06:32:47.358852 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c492d1c2-3ee8-491a-a237-98691bfacec6" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 06 06:32:47 crc kubenswrapper[4809]: I1206 06:32:47.359111 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="c492d1c2-3ee8-491a-a237-98691bfacec6" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 06 06:32:47 crc kubenswrapper[4809]: I1206 06:32:47.360361 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6zmrk" Dec 06 06:32:47 crc kubenswrapper[4809]: I1206 06:32:47.363421 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 06:32:47 crc kubenswrapper[4809]: I1206 06:32:47.363486 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 06:32:47 crc kubenswrapper[4809]: I1206 06:32:47.364083 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-bzlf5" Dec 06 06:32:47 crc kubenswrapper[4809]: I1206 06:32:47.364344 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 06:32:47 crc kubenswrapper[4809]: I1206 06:32:47.370203 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6zmrk"] Dec 06 06:32:47 crc kubenswrapper[4809]: I1206 06:32:47.430448 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plbtd\" (UniqueName: \"kubernetes.io/projected/c4012316-1bc7-4ca2-bdb8-d541646d6a9c-kube-api-access-plbtd\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6zmrk\" (UID: \"c4012316-1bc7-4ca2-bdb8-d541646d6a9c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6zmrk" Dec 06 06:32:47 crc kubenswrapper[4809]: I1206 06:32:47.430513 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c4012316-1bc7-4ca2-bdb8-d541646d6a9c-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6zmrk\" (UID: \"c4012316-1bc7-4ca2-bdb8-d541646d6a9c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6zmrk" Dec 06 06:32:47 crc kubenswrapper[4809]: I1206 06:32:47.431494 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c4012316-1bc7-4ca2-bdb8-d541646d6a9c-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6zmrk\" (UID: \"c4012316-1bc7-4ca2-bdb8-d541646d6a9c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6zmrk" Dec 06 06:32:47 crc kubenswrapper[4809]: I1206 06:32:47.533255 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plbtd\" (UniqueName: \"kubernetes.io/projected/c4012316-1bc7-4ca2-bdb8-d541646d6a9c-kube-api-access-plbtd\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6zmrk\" (UID: \"c4012316-1bc7-4ca2-bdb8-d541646d6a9c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6zmrk" Dec 06 06:32:47 crc kubenswrapper[4809]: I1206 06:32:47.533317 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c4012316-1bc7-4ca2-bdb8-d541646d6a9c-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6zmrk\" (UID: \"c4012316-1bc7-4ca2-bdb8-d541646d6a9c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6zmrk" Dec 06 06:32:47 crc kubenswrapper[4809]: I1206 06:32:47.533495 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c4012316-1bc7-4ca2-bdb8-d541646d6a9c-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6zmrk\" (UID: \"c4012316-1bc7-4ca2-bdb8-d541646d6a9c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6zmrk" Dec 06 06:32:47 crc kubenswrapper[4809]: I1206 06:32:47.538275 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c4012316-1bc7-4ca2-bdb8-d541646d6a9c-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6zmrk\" (UID: \"c4012316-1bc7-4ca2-bdb8-d541646d6a9c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6zmrk" Dec 06 06:32:47 crc kubenswrapper[4809]: I1206 06:32:47.539496 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c4012316-1bc7-4ca2-bdb8-d541646d6a9c-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6zmrk\" (UID: \"c4012316-1bc7-4ca2-bdb8-d541646d6a9c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6zmrk" Dec 06 06:32:47 crc kubenswrapper[4809]: I1206 06:32:47.549836 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plbtd\" (UniqueName: \"kubernetes.io/projected/c4012316-1bc7-4ca2-bdb8-d541646d6a9c-kube-api-access-plbtd\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6zmrk\" (UID: \"c4012316-1bc7-4ca2-bdb8-d541646d6a9c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6zmrk" Dec 06 06:32:47 crc kubenswrapper[4809]: I1206 06:32:47.698296 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6zmrk" Dec 06 06:32:48 crc kubenswrapper[4809]: I1206 06:32:48.228111 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6zmrk"] Dec 06 06:32:48 crc kubenswrapper[4809]: I1206 06:32:48.307014 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6zmrk" event={"ID":"c4012316-1bc7-4ca2-bdb8-d541646d6a9c","Type":"ContainerStarted","Data":"5ff149208d3dc3446a410f70f89a5c81e9025d74147894bb97fda8e4700b093a"} Dec 06 06:32:49 crc kubenswrapper[4809]: I1206 06:32:49.322048 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6zmrk" event={"ID":"c4012316-1bc7-4ca2-bdb8-d541646d6a9c","Type":"ContainerStarted","Data":"781b7f239765ac652afbbdf732d3b879ed6e7f05b096f8a481a6233a864adfc7"} Dec 06 06:32:49 crc kubenswrapper[4809]: I1206 06:32:49.367611 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6zmrk" podStartSLOduration=1.882765671 podStartE2EDuration="2.367589342s" podCreationTimestamp="2025-12-06 06:32:47 +0000 UTC" firstStartedPulling="2025-12-06 06:32:48.235054148 +0000 UTC m=+2493.124037090" lastFinishedPulling="2025-12-06 06:32:48.719877819 +0000 UTC m=+2493.608860761" observedRunningTime="2025-12-06 06:32:49.351738582 +0000 UTC m=+2494.240721544" watchObservedRunningTime="2025-12-06 06:32:49.367589342 +0000 UTC m=+2494.256572284" Dec 06 06:33:00 crc kubenswrapper[4809]: I1206 06:33:00.388150 4809 scope.go:117] "RemoveContainer" containerID="3aa8c940468b111de41fd7720fd216eebfc0bb377ed933226628430658845c32" Dec 06 06:33:00 crc kubenswrapper[4809]: E1206 06:33:00.388993 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:33:14 crc kubenswrapper[4809]: I1206 06:33:14.388693 4809 scope.go:117] "RemoveContainer" containerID="3aa8c940468b111de41fd7720fd216eebfc0bb377ed933226628430658845c32" Dec 06 06:33:14 crc kubenswrapper[4809]: E1206 06:33:14.389558 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:33:25 crc kubenswrapper[4809]: I1206 06:33:25.897195 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4jwhh"] Dec 06 06:33:25 crc kubenswrapper[4809]: I1206 06:33:25.901373 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4jwhh" Dec 06 06:33:25 crc kubenswrapper[4809]: I1206 06:33:25.926154 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4jwhh"] Dec 06 06:33:26 crc kubenswrapper[4809]: I1206 06:33:26.095122 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78f9b3d2-5ed4-4dfa-a2eb-577d694387b2-utilities\") pod \"redhat-operators-4jwhh\" (UID: \"78f9b3d2-5ed4-4dfa-a2eb-577d694387b2\") " pod="openshift-marketplace/redhat-operators-4jwhh" Dec 06 06:33:26 crc kubenswrapper[4809]: I1206 06:33:26.095964 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ms6s\" (UniqueName: \"kubernetes.io/projected/78f9b3d2-5ed4-4dfa-a2eb-577d694387b2-kube-api-access-7ms6s\") pod \"redhat-operators-4jwhh\" (UID: \"78f9b3d2-5ed4-4dfa-a2eb-577d694387b2\") " pod="openshift-marketplace/redhat-operators-4jwhh" Dec 06 06:33:26 crc kubenswrapper[4809]: I1206 06:33:26.096065 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78f9b3d2-5ed4-4dfa-a2eb-577d694387b2-catalog-content\") pod \"redhat-operators-4jwhh\" (UID: \"78f9b3d2-5ed4-4dfa-a2eb-577d694387b2\") " pod="openshift-marketplace/redhat-operators-4jwhh" Dec 06 06:33:26 crc kubenswrapper[4809]: I1206 06:33:26.198271 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78f9b3d2-5ed4-4dfa-a2eb-577d694387b2-utilities\") pod \"redhat-operators-4jwhh\" (UID: \"78f9b3d2-5ed4-4dfa-a2eb-577d694387b2\") " pod="openshift-marketplace/redhat-operators-4jwhh" Dec 06 06:33:26 crc kubenswrapper[4809]: I1206 06:33:26.198341 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ms6s\" (UniqueName: \"kubernetes.io/projected/78f9b3d2-5ed4-4dfa-a2eb-577d694387b2-kube-api-access-7ms6s\") pod \"redhat-operators-4jwhh\" (UID: \"78f9b3d2-5ed4-4dfa-a2eb-577d694387b2\") " pod="openshift-marketplace/redhat-operators-4jwhh" Dec 06 06:33:26 crc kubenswrapper[4809]: I1206 06:33:26.198374 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78f9b3d2-5ed4-4dfa-a2eb-577d694387b2-catalog-content\") pod \"redhat-operators-4jwhh\" (UID: \"78f9b3d2-5ed4-4dfa-a2eb-577d694387b2\") " pod="openshift-marketplace/redhat-operators-4jwhh" Dec 06 06:33:26 crc kubenswrapper[4809]: I1206 06:33:26.198768 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78f9b3d2-5ed4-4dfa-a2eb-577d694387b2-utilities\") pod \"redhat-operators-4jwhh\" (UID: \"78f9b3d2-5ed4-4dfa-a2eb-577d694387b2\") " pod="openshift-marketplace/redhat-operators-4jwhh" Dec 06 06:33:26 crc kubenswrapper[4809]: I1206 06:33:26.198819 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78f9b3d2-5ed4-4dfa-a2eb-577d694387b2-catalog-content\") pod \"redhat-operators-4jwhh\" (UID: \"78f9b3d2-5ed4-4dfa-a2eb-577d694387b2\") " pod="openshift-marketplace/redhat-operators-4jwhh" Dec 06 06:33:26 crc kubenswrapper[4809]: I1206 06:33:26.218317 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ms6s\" (UniqueName: \"kubernetes.io/projected/78f9b3d2-5ed4-4dfa-a2eb-577d694387b2-kube-api-access-7ms6s\") pod \"redhat-operators-4jwhh\" (UID: \"78f9b3d2-5ed4-4dfa-a2eb-577d694387b2\") " pod="openshift-marketplace/redhat-operators-4jwhh" Dec 06 06:33:26 crc kubenswrapper[4809]: I1206 06:33:26.225253 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4jwhh" Dec 06 06:33:26 crc kubenswrapper[4809]: I1206 06:33:26.748209 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4jwhh"] Dec 06 06:33:26 crc kubenswrapper[4809]: I1206 06:33:26.778384 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4jwhh" event={"ID":"78f9b3d2-5ed4-4dfa-a2eb-577d694387b2","Type":"ContainerStarted","Data":"2f9a417d27ef7ea9c6becdf9563d8b1697e962f4c21a36ccfd3891dd44bf8ee4"} Dec 06 06:33:28 crc kubenswrapper[4809]: I1206 06:33:28.803664 4809 generic.go:334] "Generic (PLEG): container finished" podID="78f9b3d2-5ed4-4dfa-a2eb-577d694387b2" containerID="69787a1985dcfb0030137c138151e298fbfdc05489ac7353c4687ab3cfa91a2f" exitCode=0 Dec 06 06:33:28 crc kubenswrapper[4809]: I1206 06:33:28.803860 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4jwhh" event={"ID":"78f9b3d2-5ed4-4dfa-a2eb-577d694387b2","Type":"ContainerDied","Data":"69787a1985dcfb0030137c138151e298fbfdc05489ac7353c4687ab3cfa91a2f"} Dec 06 06:33:29 crc kubenswrapper[4809]: I1206 06:33:29.388827 4809 scope.go:117] "RemoveContainer" containerID="3aa8c940468b111de41fd7720fd216eebfc0bb377ed933226628430658845c32" Dec 06 06:33:29 crc kubenswrapper[4809]: E1206 06:33:29.389392 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:33:30 crc kubenswrapper[4809]: I1206 06:33:30.827152 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4jwhh" event={"ID":"78f9b3d2-5ed4-4dfa-a2eb-577d694387b2","Type":"ContainerStarted","Data":"8c5f23df949dfd147c31ed3da5789f7c989161fffc6357ffe6a89762d95b8805"} Dec 06 06:33:31 crc kubenswrapper[4809]: I1206 06:33:31.568066 4809 scope.go:117] "RemoveContainer" containerID="d29e7e6ec1780e38eaf406f18e34fb3ad7982088153da6e08d1bbec4979270ac" Dec 06 06:33:43 crc kubenswrapper[4809]: I1206 06:33:43.389734 4809 scope.go:117] "RemoveContainer" containerID="3aa8c940468b111de41fd7720fd216eebfc0bb377ed933226628430658845c32" Dec 06 06:33:43 crc kubenswrapper[4809]: E1206 06:33:43.390736 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:33:43 crc kubenswrapper[4809]: I1206 06:33:43.986260 4809 generic.go:334] "Generic (PLEG): container finished" podID="78f9b3d2-5ed4-4dfa-a2eb-577d694387b2" containerID="8c5f23df949dfd147c31ed3da5789f7c989161fffc6357ffe6a89762d95b8805" exitCode=0 Dec 06 06:33:43 crc kubenswrapper[4809]: I1206 06:33:43.986354 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4jwhh" event={"ID":"78f9b3d2-5ed4-4dfa-a2eb-577d694387b2","Type":"ContainerDied","Data":"8c5f23df949dfd147c31ed3da5789f7c989161fffc6357ffe6a89762d95b8805"} Dec 06 06:33:43 crc kubenswrapper[4809]: I1206 06:33:43.989256 4809 generic.go:334] "Generic (PLEG): container finished" podID="c4012316-1bc7-4ca2-bdb8-d541646d6a9c" containerID="781b7f239765ac652afbbdf732d3b879ed6e7f05b096f8a481a6233a864adfc7" exitCode=0 Dec 06 06:33:43 crc kubenswrapper[4809]: I1206 06:33:43.989299 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6zmrk" event={"ID":"c4012316-1bc7-4ca2-bdb8-d541646d6a9c","Type":"ContainerDied","Data":"781b7f239765ac652afbbdf732d3b879ed6e7f05b096f8a481a6233a864adfc7"} Dec 06 06:33:45 crc kubenswrapper[4809]: I1206 06:33:45.568820 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6zmrk" Dec 06 06:33:45 crc kubenswrapper[4809]: I1206 06:33:45.650947 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-plbtd\" (UniqueName: \"kubernetes.io/projected/c4012316-1bc7-4ca2-bdb8-d541646d6a9c-kube-api-access-plbtd\") pod \"c4012316-1bc7-4ca2-bdb8-d541646d6a9c\" (UID: \"c4012316-1bc7-4ca2-bdb8-d541646d6a9c\") " Dec 06 06:33:45 crc kubenswrapper[4809]: I1206 06:33:45.651314 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c4012316-1bc7-4ca2-bdb8-d541646d6a9c-ssh-key\") pod \"c4012316-1bc7-4ca2-bdb8-d541646d6a9c\" (UID: \"c4012316-1bc7-4ca2-bdb8-d541646d6a9c\") " Dec 06 06:33:45 crc kubenswrapper[4809]: I1206 06:33:45.651479 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c4012316-1bc7-4ca2-bdb8-d541646d6a9c-inventory\") pod \"c4012316-1bc7-4ca2-bdb8-d541646d6a9c\" (UID: \"c4012316-1bc7-4ca2-bdb8-d541646d6a9c\") " Dec 06 06:33:45 crc kubenswrapper[4809]: I1206 06:33:45.660415 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4012316-1bc7-4ca2-bdb8-d541646d6a9c-kube-api-access-plbtd" (OuterVolumeSpecName: "kube-api-access-plbtd") pod "c4012316-1bc7-4ca2-bdb8-d541646d6a9c" (UID: "c4012316-1bc7-4ca2-bdb8-d541646d6a9c"). InnerVolumeSpecName "kube-api-access-plbtd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:33:45 crc kubenswrapper[4809]: I1206 06:33:45.691274 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4012316-1bc7-4ca2-bdb8-d541646d6a9c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c4012316-1bc7-4ca2-bdb8-d541646d6a9c" (UID: "c4012316-1bc7-4ca2-bdb8-d541646d6a9c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:33:45 crc kubenswrapper[4809]: I1206 06:33:45.713238 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4012316-1bc7-4ca2-bdb8-d541646d6a9c-inventory" (OuterVolumeSpecName: "inventory") pod "c4012316-1bc7-4ca2-bdb8-d541646d6a9c" (UID: "c4012316-1bc7-4ca2-bdb8-d541646d6a9c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:33:45 crc kubenswrapper[4809]: I1206 06:33:45.754338 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c4012316-1bc7-4ca2-bdb8-d541646d6a9c-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 06:33:45 crc kubenswrapper[4809]: I1206 06:33:45.754379 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c4012316-1bc7-4ca2-bdb8-d541646d6a9c-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 06:33:45 crc kubenswrapper[4809]: I1206 06:33:45.754393 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-plbtd\" (UniqueName: \"kubernetes.io/projected/c4012316-1bc7-4ca2-bdb8-d541646d6a9c-kube-api-access-plbtd\") on node \"crc\" DevicePath \"\"" Dec 06 06:33:46 crc kubenswrapper[4809]: I1206 06:33:46.014160 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4jwhh" event={"ID":"78f9b3d2-5ed4-4dfa-a2eb-577d694387b2","Type":"ContainerStarted","Data":"4317db95676b08496917b855ca265932f6ea982bcd35014ee23e80ff78ae2594"} Dec 06 06:33:46 crc kubenswrapper[4809]: I1206 06:33:46.015416 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6zmrk" event={"ID":"c4012316-1bc7-4ca2-bdb8-d541646d6a9c","Type":"ContainerDied","Data":"5ff149208d3dc3446a410f70f89a5c81e9025d74147894bb97fda8e4700b093a"} Dec 06 06:33:46 crc kubenswrapper[4809]: I1206 06:33:46.015464 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5ff149208d3dc3446a410f70f89a5c81e9025d74147894bb97fda8e4700b093a" Dec 06 06:33:46 crc kubenswrapper[4809]: I1206 06:33:46.015496 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6zmrk" Dec 06 06:33:46 crc kubenswrapper[4809]: I1206 06:33:46.045716 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4jwhh" podStartSLOduration=4.807768449 podStartE2EDuration="21.045693481s" podCreationTimestamp="2025-12-06 06:33:25 +0000 UTC" firstStartedPulling="2025-12-06 06:33:28.806018898 +0000 UTC m=+2533.695001850" lastFinishedPulling="2025-12-06 06:33:45.04394394 +0000 UTC m=+2549.932926882" observedRunningTime="2025-12-06 06:33:46.036190614 +0000 UTC m=+2550.925173566" watchObservedRunningTime="2025-12-06 06:33:46.045693481 +0000 UTC m=+2550.934676423" Dec 06 06:33:46 crc kubenswrapper[4809]: I1206 06:33:46.162027 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-4tq4x"] Dec 06 06:33:46 crc kubenswrapper[4809]: E1206 06:33:46.163116 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4012316-1bc7-4ca2-bdb8-d541646d6a9c" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 06 06:33:46 crc kubenswrapper[4809]: I1206 06:33:46.163253 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4012316-1bc7-4ca2-bdb8-d541646d6a9c" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 06 06:33:46 crc kubenswrapper[4809]: I1206 06:33:46.163667 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4012316-1bc7-4ca2-bdb8-d541646d6a9c" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 06 06:33:46 crc kubenswrapper[4809]: I1206 06:33:46.165158 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-4tq4x" Dec 06 06:33:46 crc kubenswrapper[4809]: I1206 06:33:46.168625 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 06:33:46 crc kubenswrapper[4809]: I1206 06:33:46.168786 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 06:33:46 crc kubenswrapper[4809]: I1206 06:33:46.168843 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 06:33:46 crc kubenswrapper[4809]: I1206 06:33:46.168892 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-bzlf5" Dec 06 06:33:46 crc kubenswrapper[4809]: I1206 06:33:46.179172 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-4tq4x"] Dec 06 06:33:46 crc kubenswrapper[4809]: I1206 06:33:46.226201 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4jwhh" Dec 06 06:33:46 crc kubenswrapper[4809]: I1206 06:33:46.226344 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4jwhh" Dec 06 06:33:46 crc kubenswrapper[4809]: I1206 06:33:46.267144 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rlhx\" (UniqueName: \"kubernetes.io/projected/5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280-kube-api-access-5rlhx\") pod \"ssh-known-hosts-edpm-deployment-4tq4x\" (UID: \"5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280\") " pod="openstack/ssh-known-hosts-edpm-deployment-4tq4x" Dec 06 06:33:46 crc kubenswrapper[4809]: I1206 06:33:46.267493 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-4tq4x\" (UID: \"5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280\") " pod="openstack/ssh-known-hosts-edpm-deployment-4tq4x" Dec 06 06:33:46 crc kubenswrapper[4809]: I1206 06:33:46.267777 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-4tq4x\" (UID: \"5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280\") " pod="openstack/ssh-known-hosts-edpm-deployment-4tq4x" Dec 06 06:33:46 crc kubenswrapper[4809]: I1206 06:33:46.371373 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-4tq4x\" (UID: \"5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280\") " pod="openstack/ssh-known-hosts-edpm-deployment-4tq4x" Dec 06 06:33:46 crc kubenswrapper[4809]: I1206 06:33:46.371632 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rlhx\" (UniqueName: \"kubernetes.io/projected/5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280-kube-api-access-5rlhx\") pod \"ssh-known-hosts-edpm-deployment-4tq4x\" (UID: \"5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280\") " pod="openstack/ssh-known-hosts-edpm-deployment-4tq4x" Dec 06 06:33:46 crc kubenswrapper[4809]: I1206 06:33:46.371716 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-4tq4x\" (UID: \"5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280\") " pod="openstack/ssh-known-hosts-edpm-deployment-4tq4x" Dec 06 06:33:46 crc kubenswrapper[4809]: I1206 06:33:46.383748 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-4tq4x\" (UID: \"5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280\") " pod="openstack/ssh-known-hosts-edpm-deployment-4tq4x" Dec 06 06:33:46 crc kubenswrapper[4809]: I1206 06:33:46.383773 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-4tq4x\" (UID: \"5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280\") " pod="openstack/ssh-known-hosts-edpm-deployment-4tq4x" Dec 06 06:33:46 crc kubenswrapper[4809]: I1206 06:33:46.391533 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rlhx\" (UniqueName: \"kubernetes.io/projected/5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280-kube-api-access-5rlhx\") pod \"ssh-known-hosts-edpm-deployment-4tq4x\" (UID: \"5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280\") " pod="openstack/ssh-known-hosts-edpm-deployment-4tq4x" Dec 06 06:33:46 crc kubenswrapper[4809]: I1206 06:33:46.490205 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-4tq4x" Dec 06 06:33:47 crc kubenswrapper[4809]: W1206 06:33:47.067665 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e4aa7e6_47ff_4f5f_abd9_4c4a193f2280.slice/crio-7567b3b5d9790b79ac625d507397d936132b9d99496a4cf9130e851f4b8bd0b0 WatchSource:0}: Error finding container 7567b3b5d9790b79ac625d507397d936132b9d99496a4cf9130e851f4b8bd0b0: Status 404 returned error can't find the container with id 7567b3b5d9790b79ac625d507397d936132b9d99496a4cf9130e851f4b8bd0b0 Dec 06 06:33:47 crc kubenswrapper[4809]: I1206 06:33:47.068128 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-4tq4x"] Dec 06 06:33:47 crc kubenswrapper[4809]: I1206 06:33:47.289853 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-4jwhh" podUID="78f9b3d2-5ed4-4dfa-a2eb-577d694387b2" containerName="registry-server" probeResult="failure" output=< Dec 06 06:33:47 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Dec 06 06:33:47 crc kubenswrapper[4809]: > Dec 06 06:33:48 crc kubenswrapper[4809]: I1206 06:33:48.037734 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-4tq4x" event={"ID":"5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280","Type":"ContainerStarted","Data":"7567b3b5d9790b79ac625d507397d936132b9d99496a4cf9130e851f4b8bd0b0"} Dec 06 06:33:51 crc kubenswrapper[4809]: I1206 06:33:51.059808 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-q7rkl"] Dec 06 06:33:51 crc kubenswrapper[4809]: I1206 06:33:51.072392 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-q7rkl"] Dec 06 06:33:51 crc kubenswrapper[4809]: I1206 06:33:51.072871 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-4tq4x" event={"ID":"5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280","Type":"ContainerStarted","Data":"325a5b3833785d0e0aa32073a36c9207c99c28fc1dd173e748189076beeae74c"} Dec 06 06:33:51 crc kubenswrapper[4809]: I1206 06:33:51.411510 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a75c2789-4b84-455c-9215-2c52308dbace" path="/var/lib/kubelet/pods/a75c2789-4b84-455c-9215-2c52308dbace/volumes" Dec 06 06:33:52 crc kubenswrapper[4809]: I1206 06:33:52.106233 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-4tq4x" podStartSLOduration=3.100125567 podStartE2EDuration="6.106207002s" podCreationTimestamp="2025-12-06 06:33:46 +0000 UTC" firstStartedPulling="2025-12-06 06:33:47.069624423 +0000 UTC m=+2551.958607375" lastFinishedPulling="2025-12-06 06:33:50.075705868 +0000 UTC m=+2554.964688810" observedRunningTime="2025-12-06 06:33:52.100349723 +0000 UTC m=+2556.989332665" watchObservedRunningTime="2025-12-06 06:33:52.106207002 +0000 UTC m=+2556.995189954" Dec 06 06:33:55 crc kubenswrapper[4809]: I1206 06:33:55.395408 4809 scope.go:117] "RemoveContainer" containerID="3aa8c940468b111de41fd7720fd216eebfc0bb377ed933226628430658845c32" Dec 06 06:33:55 crc kubenswrapper[4809]: E1206 06:33:55.397724 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:33:56 crc kubenswrapper[4809]: I1206 06:33:56.298053 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4jwhh" Dec 06 06:33:56 crc kubenswrapper[4809]: I1206 06:33:56.388522 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4jwhh" Dec 06 06:33:57 crc kubenswrapper[4809]: I1206 06:33:57.102472 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4jwhh"] Dec 06 06:33:58 crc kubenswrapper[4809]: I1206 06:33:58.177306 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4jwhh" podUID="78f9b3d2-5ed4-4dfa-a2eb-577d694387b2" containerName="registry-server" containerID="cri-o://4317db95676b08496917b855ca265932f6ea982bcd35014ee23e80ff78ae2594" gracePeriod=2 Dec 06 06:33:59 crc kubenswrapper[4809]: I1206 06:33:59.195643 4809 generic.go:334] "Generic (PLEG): container finished" podID="78f9b3d2-5ed4-4dfa-a2eb-577d694387b2" containerID="4317db95676b08496917b855ca265932f6ea982bcd35014ee23e80ff78ae2594" exitCode=0 Dec 06 06:33:59 crc kubenswrapper[4809]: I1206 06:33:59.195919 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4jwhh" event={"ID":"78f9b3d2-5ed4-4dfa-a2eb-577d694387b2","Type":"ContainerDied","Data":"4317db95676b08496917b855ca265932f6ea982bcd35014ee23e80ff78ae2594"} Dec 06 06:33:59 crc kubenswrapper[4809]: I1206 06:33:59.195964 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4jwhh" event={"ID":"78f9b3d2-5ed4-4dfa-a2eb-577d694387b2","Type":"ContainerDied","Data":"2f9a417d27ef7ea9c6becdf9563d8b1697e962f4c21a36ccfd3891dd44bf8ee4"} Dec 06 06:33:59 crc kubenswrapper[4809]: I1206 06:33:59.195978 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f9a417d27ef7ea9c6becdf9563d8b1697e962f4c21a36ccfd3891dd44bf8ee4" Dec 06 06:33:59 crc kubenswrapper[4809]: I1206 06:33:59.239226 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4jwhh" Dec 06 06:33:59 crc kubenswrapper[4809]: I1206 06:33:59.330262 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78f9b3d2-5ed4-4dfa-a2eb-577d694387b2-catalog-content\") pod \"78f9b3d2-5ed4-4dfa-a2eb-577d694387b2\" (UID: \"78f9b3d2-5ed4-4dfa-a2eb-577d694387b2\") " Dec 06 06:33:59 crc kubenswrapper[4809]: I1206 06:33:59.331025 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ms6s\" (UniqueName: \"kubernetes.io/projected/78f9b3d2-5ed4-4dfa-a2eb-577d694387b2-kube-api-access-7ms6s\") pod \"78f9b3d2-5ed4-4dfa-a2eb-577d694387b2\" (UID: \"78f9b3d2-5ed4-4dfa-a2eb-577d694387b2\") " Dec 06 06:33:59 crc kubenswrapper[4809]: I1206 06:33:59.331095 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78f9b3d2-5ed4-4dfa-a2eb-577d694387b2-utilities\") pod \"78f9b3d2-5ed4-4dfa-a2eb-577d694387b2\" (UID: \"78f9b3d2-5ed4-4dfa-a2eb-577d694387b2\") " Dec 06 06:33:59 crc kubenswrapper[4809]: I1206 06:33:59.332216 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78f9b3d2-5ed4-4dfa-a2eb-577d694387b2-utilities" (OuterVolumeSpecName: "utilities") pod "78f9b3d2-5ed4-4dfa-a2eb-577d694387b2" (UID: "78f9b3d2-5ed4-4dfa-a2eb-577d694387b2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:33:59 crc kubenswrapper[4809]: I1206 06:33:59.338161 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78f9b3d2-5ed4-4dfa-a2eb-577d694387b2-kube-api-access-7ms6s" (OuterVolumeSpecName: "kube-api-access-7ms6s") pod "78f9b3d2-5ed4-4dfa-a2eb-577d694387b2" (UID: "78f9b3d2-5ed4-4dfa-a2eb-577d694387b2"). InnerVolumeSpecName "kube-api-access-7ms6s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:33:59 crc kubenswrapper[4809]: I1206 06:33:59.434351 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ms6s\" (UniqueName: \"kubernetes.io/projected/78f9b3d2-5ed4-4dfa-a2eb-577d694387b2-kube-api-access-7ms6s\") on node \"crc\" DevicePath \"\"" Dec 06 06:33:59 crc kubenswrapper[4809]: I1206 06:33:59.434383 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78f9b3d2-5ed4-4dfa-a2eb-577d694387b2-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:33:59 crc kubenswrapper[4809]: I1206 06:33:59.471971 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78f9b3d2-5ed4-4dfa-a2eb-577d694387b2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "78f9b3d2-5ed4-4dfa-a2eb-577d694387b2" (UID: "78f9b3d2-5ed4-4dfa-a2eb-577d694387b2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:33:59 crc kubenswrapper[4809]: I1206 06:33:59.538746 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78f9b3d2-5ed4-4dfa-a2eb-577d694387b2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:34:00 crc kubenswrapper[4809]: I1206 06:34:00.210495 4809 generic.go:334] "Generic (PLEG): container finished" podID="5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280" containerID="325a5b3833785d0e0aa32073a36c9207c99c28fc1dd173e748189076beeae74c" exitCode=0 Dec 06 06:34:00 crc kubenswrapper[4809]: I1206 06:34:00.211571 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-4tq4x" event={"ID":"5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280","Type":"ContainerDied","Data":"325a5b3833785d0e0aa32073a36c9207c99c28fc1dd173e748189076beeae74c"} Dec 06 06:34:00 crc kubenswrapper[4809]: I1206 06:34:00.212145 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4jwhh" Dec 06 06:34:00 crc kubenswrapper[4809]: I1206 06:34:00.267835 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4jwhh"] Dec 06 06:34:00 crc kubenswrapper[4809]: I1206 06:34:00.278768 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4jwhh"] Dec 06 06:34:01 crc kubenswrapper[4809]: I1206 06:34:01.402416 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78f9b3d2-5ed4-4dfa-a2eb-577d694387b2" path="/var/lib/kubelet/pods/78f9b3d2-5ed4-4dfa-a2eb-577d694387b2/volumes" Dec 06 06:34:01 crc kubenswrapper[4809]: I1206 06:34:01.725164 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-4tq4x" Dec 06 06:34:01 crc kubenswrapper[4809]: I1206 06:34:01.794689 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5rlhx\" (UniqueName: \"kubernetes.io/projected/5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280-kube-api-access-5rlhx\") pod \"5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280\" (UID: \"5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280\") " Dec 06 06:34:01 crc kubenswrapper[4809]: I1206 06:34:01.794807 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280-ssh-key-openstack-edpm-ipam\") pod \"5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280\" (UID: \"5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280\") " Dec 06 06:34:01 crc kubenswrapper[4809]: I1206 06:34:01.794978 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280-inventory-0\") pod \"5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280\" (UID: \"5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280\") " Dec 06 06:34:01 crc kubenswrapper[4809]: I1206 06:34:01.802850 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280-kube-api-access-5rlhx" (OuterVolumeSpecName: "kube-api-access-5rlhx") pod "5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280" (UID: "5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280"). InnerVolumeSpecName "kube-api-access-5rlhx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:34:01 crc kubenswrapper[4809]: I1206 06:34:01.836978 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280" (UID: "5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:34:01 crc kubenswrapper[4809]: I1206 06:34:01.850673 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280" (UID: "5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:34:01 crc kubenswrapper[4809]: I1206 06:34:01.898763 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5rlhx\" (UniqueName: \"kubernetes.io/projected/5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280-kube-api-access-5rlhx\") on node \"crc\" DevicePath \"\"" Dec 06 06:34:01 crc kubenswrapper[4809]: I1206 06:34:01.898808 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 06 06:34:01 crc kubenswrapper[4809]: I1206 06:34:01.898821 4809 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:34:02 crc kubenswrapper[4809]: I1206 06:34:02.232833 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-4tq4x" event={"ID":"5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280","Type":"ContainerDied","Data":"7567b3b5d9790b79ac625d507397d936132b9d99496a4cf9130e851f4b8bd0b0"} Dec 06 06:34:02 crc kubenswrapper[4809]: I1206 06:34:02.233264 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7567b3b5d9790b79ac625d507397d936132b9d99496a4cf9130e851f4b8bd0b0" Dec 06 06:34:02 crc kubenswrapper[4809]: I1206 06:34:02.232870 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-4tq4x" Dec 06 06:34:02 crc kubenswrapper[4809]: I1206 06:34:02.355248 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-mcdd9"] Dec 06 06:34:02 crc kubenswrapper[4809]: E1206 06:34:02.356464 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280" containerName="ssh-known-hosts-edpm-deployment" Dec 06 06:34:02 crc kubenswrapper[4809]: I1206 06:34:02.356492 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280" containerName="ssh-known-hosts-edpm-deployment" Dec 06 06:34:02 crc kubenswrapper[4809]: E1206 06:34:02.356522 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78f9b3d2-5ed4-4dfa-a2eb-577d694387b2" containerName="extract-content" Dec 06 06:34:02 crc kubenswrapper[4809]: I1206 06:34:02.356534 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="78f9b3d2-5ed4-4dfa-a2eb-577d694387b2" containerName="extract-content" Dec 06 06:34:02 crc kubenswrapper[4809]: E1206 06:34:02.356603 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78f9b3d2-5ed4-4dfa-a2eb-577d694387b2" containerName="extract-utilities" Dec 06 06:34:02 crc kubenswrapper[4809]: I1206 06:34:02.356612 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="78f9b3d2-5ed4-4dfa-a2eb-577d694387b2" containerName="extract-utilities" Dec 06 06:34:02 crc kubenswrapper[4809]: E1206 06:34:02.356648 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78f9b3d2-5ed4-4dfa-a2eb-577d694387b2" containerName="registry-server" Dec 06 06:34:02 crc kubenswrapper[4809]: I1206 06:34:02.356657 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="78f9b3d2-5ed4-4dfa-a2eb-577d694387b2" containerName="registry-server" Dec 06 06:34:02 crc kubenswrapper[4809]: I1206 06:34:02.357358 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="78f9b3d2-5ed4-4dfa-a2eb-577d694387b2" containerName="registry-server" Dec 06 06:34:02 crc kubenswrapper[4809]: I1206 06:34:02.357445 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280" containerName="ssh-known-hosts-edpm-deployment" Dec 06 06:34:02 crc kubenswrapper[4809]: I1206 06:34:02.359114 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mcdd9" Dec 06 06:34:02 crc kubenswrapper[4809]: I1206 06:34:02.363902 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 06:34:02 crc kubenswrapper[4809]: I1206 06:34:02.364265 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 06:34:02 crc kubenswrapper[4809]: I1206 06:34:02.364740 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 06:34:02 crc kubenswrapper[4809]: I1206 06:34:02.373309 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-bzlf5" Dec 06 06:34:02 crc kubenswrapper[4809]: I1206 06:34:02.389975 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-mcdd9"] Dec 06 06:34:02 crc kubenswrapper[4809]: I1206 06:34:02.409951 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcx7r\" (UniqueName: \"kubernetes.io/projected/896d8df1-4885-4ffa-a368-87b680cd909c-kube-api-access-qcx7r\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-mcdd9\" (UID: \"896d8df1-4885-4ffa-a368-87b680cd909c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mcdd9" Dec 06 06:34:02 crc kubenswrapper[4809]: I1206 06:34:02.409998 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/896d8df1-4885-4ffa-a368-87b680cd909c-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-mcdd9\" (UID: \"896d8df1-4885-4ffa-a368-87b680cd909c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mcdd9" Dec 06 06:34:02 crc kubenswrapper[4809]: I1206 06:34:02.410153 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/896d8df1-4885-4ffa-a368-87b680cd909c-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-mcdd9\" (UID: \"896d8df1-4885-4ffa-a368-87b680cd909c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mcdd9" Dec 06 06:34:02 crc kubenswrapper[4809]: I1206 06:34:02.512320 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/896d8df1-4885-4ffa-a368-87b680cd909c-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-mcdd9\" (UID: \"896d8df1-4885-4ffa-a368-87b680cd909c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mcdd9" Dec 06 06:34:02 crc kubenswrapper[4809]: I1206 06:34:02.512723 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcx7r\" (UniqueName: \"kubernetes.io/projected/896d8df1-4885-4ffa-a368-87b680cd909c-kube-api-access-qcx7r\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-mcdd9\" (UID: \"896d8df1-4885-4ffa-a368-87b680cd909c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mcdd9" Dec 06 06:34:02 crc kubenswrapper[4809]: I1206 06:34:02.513082 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/896d8df1-4885-4ffa-a368-87b680cd909c-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-mcdd9\" (UID: \"896d8df1-4885-4ffa-a368-87b680cd909c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mcdd9" Dec 06 06:34:02 crc kubenswrapper[4809]: I1206 06:34:02.516593 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/896d8df1-4885-4ffa-a368-87b680cd909c-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-mcdd9\" (UID: \"896d8df1-4885-4ffa-a368-87b680cd909c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mcdd9" Dec 06 06:34:02 crc kubenswrapper[4809]: I1206 06:34:02.518127 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/896d8df1-4885-4ffa-a368-87b680cd909c-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-mcdd9\" (UID: \"896d8df1-4885-4ffa-a368-87b680cd909c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mcdd9" Dec 06 06:34:02 crc kubenswrapper[4809]: I1206 06:34:02.534635 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcx7r\" (UniqueName: \"kubernetes.io/projected/896d8df1-4885-4ffa-a368-87b680cd909c-kube-api-access-qcx7r\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-mcdd9\" (UID: \"896d8df1-4885-4ffa-a368-87b680cd909c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mcdd9" Dec 06 06:34:02 crc kubenswrapper[4809]: I1206 06:34:02.690561 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mcdd9" Dec 06 06:34:03 crc kubenswrapper[4809]: I1206 06:34:03.261243 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-mcdd9"] Dec 06 06:34:04 crc kubenswrapper[4809]: I1206 06:34:04.256197 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mcdd9" event={"ID":"896d8df1-4885-4ffa-a368-87b680cd909c","Type":"ContainerStarted","Data":"35417b3bd48b8e3dcdba77aa06dfcde9d1056ee3ea9faec2842e7788a8fc96fd"} Dec 06 06:34:05 crc kubenswrapper[4809]: I1206 06:34:05.428231 4809 patch_prober.go:28] interesting pod/metrics-server-556bbc5579-8wblw container/metrics-server namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.75:10250/livez\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 06:34:05 crc kubenswrapper[4809]: I1206 06:34:05.428583 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/metrics-server-556bbc5579-8wblw" podUID="b4f96141-d70e-4715-9e30-516c9e783cc9" containerName="metrics-server" probeResult="failure" output="Get \"https://10.217.0.75:10250/livez\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 06:34:05 crc kubenswrapper[4809]: I1206 06:34:05.441115 4809 patch_prober.go:28] interesting pod/console-operator-58897d9998-hlvjb container/console-operator namespace/openshift-console-operator: Liveness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 06:34:05 crc kubenswrapper[4809]: I1206 06:34:05.441175 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console-operator/console-operator-58897d9998-hlvjb" podUID="f3a30985-f268-44de-93b9-bce0f9aebc6d" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 06:34:06 crc kubenswrapper[4809]: I1206 06:34:06.388846 4809 scope.go:117] "RemoveContainer" containerID="3aa8c940468b111de41fd7720fd216eebfc0bb377ed933226628430658845c32" Dec 06 06:34:06 crc kubenswrapper[4809]: E1206 06:34:06.389143 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:34:08 crc kubenswrapper[4809]: I1206 06:34:08.783485 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Dec 06 06:34:09 crc kubenswrapper[4809]: I1206 06:34:09.683159 4809 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-qqv6x container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.30:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 06:34:09 crc kubenswrapper[4809]: I1206 06:34:09.683564 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qqv6x" podUID="8a225b5c-4ad5-4f08-a0cb-ae38a8fa83d4" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.30:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 06:34:11 crc kubenswrapper[4809]: I1206 06:34:11.954099 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/speaker-5t9z6" podUID="21411d89-4a48-4556-8ad2-193f1c8be0db" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 06:34:13 crc kubenswrapper[4809]: I1206 06:34:13.809738 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Dec 06 06:34:13 crc kubenswrapper[4809]: I1206 06:34:13.811380 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerName="ceilometer-notification-agent" probeResult="failure" output="command timed out" Dec 06 06:34:15 crc kubenswrapper[4809]: I1206 06:34:15.301194 4809 patch_prober.go:28] interesting pod/authentication-operator-69f744f599-7qpjs container/authentication-operator namespace/openshift-authentication-operator: Liveness probe status=failure output="Get \"https://10.217.0.18:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 06:34:15 crc kubenswrapper[4809]: I1206 06:34:15.301643 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication-operator/authentication-operator-69f744f599-7qpjs" podUID="2a1ed741-489c-455f-a344-404bbab66c22" containerName="authentication-operator" probeResult="failure" output="Get \"https://10.217.0.18:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 06:34:15 crc kubenswrapper[4809]: I1206 06:34:15.442157 4809 patch_prober.go:28] interesting pod/console-operator-58897d9998-hlvjb container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 06:34:15 crc kubenswrapper[4809]: I1206 06:34:15.442195 4809 patch_prober.go:28] interesting pod/console-operator-58897d9998-hlvjb container/console-operator namespace/openshift-console-operator: Liveness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 06:34:15 crc kubenswrapper[4809]: I1206 06:34:15.442209 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-hlvjb" podUID="f3a30985-f268-44de-93b9-bce0f9aebc6d" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 06:34:15 crc kubenswrapper[4809]: I1206 06:34:15.442253 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console-operator/console-operator-58897d9998-hlvjb" podUID="f3a30985-f268-44de-93b9-bce0f9aebc6d" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 06:34:16 crc kubenswrapper[4809]: I1206 06:34:16.576009 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerName="ceilometer-central-agent" probeResult="failure" output=< Dec 06 06:34:16 crc kubenswrapper[4809]: Unkown error: Expecting value: line 1 column 1 (char 0) Dec 06 06:34:16 crc kubenswrapper[4809]: > Dec 06 06:34:16 crc kubenswrapper[4809]: I1206 06:34:16.576281 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/ceilometer-0" Dec 06 06:34:16 crc kubenswrapper[4809]: I1206 06:34:16.577157 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="ceilometer-central-agent" containerStatusID={"Type":"cri-o","ID":"644e6da2ea964da6dc3103aea27cc874c7220fb372f243ee56009f3395b2d887"} pod="openstack/ceilometer-0" containerMessage="Container ceilometer-central-agent failed liveness probe, will be restarted" Dec 06 06:34:16 crc kubenswrapper[4809]: I1206 06:34:16.577228 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerName="ceilometer-central-agent" containerID="cri-o://644e6da2ea964da6dc3103aea27cc874c7220fb372f243ee56009f3395b2d887" gracePeriod=30 Dec 06 06:34:19 crc kubenswrapper[4809]: I1206 06:34:19.388610 4809 scope.go:117] "RemoveContainer" containerID="3aa8c940468b111de41fd7720fd216eebfc0bb377ed933226628430658845c32" Dec 06 06:34:19 crc kubenswrapper[4809]: E1206 06:34:19.389691 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:34:19 crc kubenswrapper[4809]: I1206 06:34:19.453031 4809 generic.go:334] "Generic (PLEG): container finished" podID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerID="644e6da2ea964da6dc3103aea27cc874c7220fb372f243ee56009f3395b2d887" exitCode=0 Dec 06 06:34:19 crc kubenswrapper[4809]: I1206 06:34:19.453127 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3d4b42d3-04a4-40e3-803b-11c77024f612","Type":"ContainerDied","Data":"644e6da2ea964da6dc3103aea27cc874c7220fb372f243ee56009f3395b2d887"} Dec 06 06:34:19 crc kubenswrapper[4809]: I1206 06:34:19.455360 4809 generic.go:334] "Generic (PLEG): container finished" podID="b376f569-0116-41f4-8a57-b5e5769e6a10" containerID="5c7db629ed3edf74098384a0a0132f41fe046c98a11cfe620d66c8f600323dc5" exitCode=1 Dec 06 06:34:19 crc kubenswrapper[4809]: I1206 06:34:19.455387 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-lqkvl" event={"ID":"b376f569-0116-41f4-8a57-b5e5769e6a10","Type":"ContainerDied","Data":"5c7db629ed3edf74098384a0a0132f41fe046c98a11cfe620d66c8f600323dc5"} Dec 06 06:34:19 crc kubenswrapper[4809]: I1206 06:34:19.456204 4809 scope.go:117] "RemoveContainer" containerID="5c7db629ed3edf74098384a0a0132f41fe046c98a11cfe620d66c8f600323dc5" Dec 06 06:34:20 crc kubenswrapper[4809]: I1206 06:34:20.565231 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 06:34:21 crc kubenswrapper[4809]: I1206 06:34:21.480198 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-lqkvl" event={"ID":"b376f569-0116-41f4-8a57-b5e5769e6a10","Type":"ContainerStarted","Data":"b317e5312c43701b451e86d74c12b7f999f7e606ea9ed3c6f3a5c991a820a635"} Dec 06 06:34:21 crc kubenswrapper[4809]: I1206 06:34:21.480738 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-lqkvl" Dec 06 06:34:21 crc kubenswrapper[4809]: I1206 06:34:21.482258 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mcdd9" event={"ID":"896d8df1-4885-4ffa-a368-87b680cd909c","Type":"ContainerStarted","Data":"c999cab42d9e6a4080b8645507c998b0f885bef1dda027927ff97b251a1555b2"} Dec 06 06:34:21 crc kubenswrapper[4809]: I1206 06:34:21.520372 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mcdd9" podStartSLOduration=2.222736962 podStartE2EDuration="19.520353294s" podCreationTimestamp="2025-12-06 06:34:02 +0000 UTC" firstStartedPulling="2025-12-06 06:34:03.262973559 +0000 UTC m=+2568.151956511" lastFinishedPulling="2025-12-06 06:34:20.560589901 +0000 UTC m=+2585.449572843" observedRunningTime="2025-12-06 06:34:21.514409053 +0000 UTC m=+2586.403391995" watchObservedRunningTime="2025-12-06 06:34:21.520353294 +0000 UTC m=+2586.409336236" Dec 06 06:34:24 crc kubenswrapper[4809]: I1206 06:34:24.532610 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3d4b42d3-04a4-40e3-803b-11c77024f612","Type":"ContainerStarted","Data":"92f800a6ae2ed5e64c1afe14ad830a380ec334e44b69d5e189323d25b3c22531"} Dec 06 06:34:30 crc kubenswrapper[4809]: I1206 06:34:30.610743 4809 generic.go:334] "Generic (PLEG): container finished" podID="896d8df1-4885-4ffa-a368-87b680cd909c" containerID="c999cab42d9e6a4080b8645507c998b0f885bef1dda027927ff97b251a1555b2" exitCode=0 Dec 06 06:34:30 crc kubenswrapper[4809]: I1206 06:34:30.610828 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mcdd9" event={"ID":"896d8df1-4885-4ffa-a368-87b680cd909c","Type":"ContainerDied","Data":"c999cab42d9e6a4080b8645507c998b0f885bef1dda027927ff97b251a1555b2"} Dec 06 06:34:31 crc kubenswrapper[4809]: I1206 06:34:31.866263 4809 scope.go:117] "RemoveContainer" containerID="afcce747e881895f738981fccf8e65088a80ac900f8d433121f09c6ca21f80da" Dec 06 06:34:32 crc kubenswrapper[4809]: I1206 06:34:32.196818 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mcdd9" Dec 06 06:34:32 crc kubenswrapper[4809]: I1206 06:34:32.327440 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/896d8df1-4885-4ffa-a368-87b680cd909c-inventory\") pod \"896d8df1-4885-4ffa-a368-87b680cd909c\" (UID: \"896d8df1-4885-4ffa-a368-87b680cd909c\") " Dec 06 06:34:32 crc kubenswrapper[4809]: I1206 06:34:32.327633 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/896d8df1-4885-4ffa-a368-87b680cd909c-ssh-key\") pod \"896d8df1-4885-4ffa-a368-87b680cd909c\" (UID: \"896d8df1-4885-4ffa-a368-87b680cd909c\") " Dec 06 06:34:32 crc kubenswrapper[4809]: I1206 06:34:32.327879 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qcx7r\" (UniqueName: \"kubernetes.io/projected/896d8df1-4885-4ffa-a368-87b680cd909c-kube-api-access-qcx7r\") pod \"896d8df1-4885-4ffa-a368-87b680cd909c\" (UID: \"896d8df1-4885-4ffa-a368-87b680cd909c\") " Dec 06 06:34:32 crc kubenswrapper[4809]: I1206 06:34:32.340871 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/896d8df1-4885-4ffa-a368-87b680cd909c-kube-api-access-qcx7r" (OuterVolumeSpecName: "kube-api-access-qcx7r") pod "896d8df1-4885-4ffa-a368-87b680cd909c" (UID: "896d8df1-4885-4ffa-a368-87b680cd909c"). InnerVolumeSpecName "kube-api-access-qcx7r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:34:32 crc kubenswrapper[4809]: I1206 06:34:32.365738 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/896d8df1-4885-4ffa-a368-87b680cd909c-inventory" (OuterVolumeSpecName: "inventory") pod "896d8df1-4885-4ffa-a368-87b680cd909c" (UID: "896d8df1-4885-4ffa-a368-87b680cd909c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:34:32 crc kubenswrapper[4809]: I1206 06:34:32.366731 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/896d8df1-4885-4ffa-a368-87b680cd909c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "896d8df1-4885-4ffa-a368-87b680cd909c" (UID: "896d8df1-4885-4ffa-a368-87b680cd909c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:34:32 crc kubenswrapper[4809]: I1206 06:34:32.431676 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/896d8df1-4885-4ffa-a368-87b680cd909c-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 06:34:32 crc kubenswrapper[4809]: I1206 06:34:32.431726 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/896d8df1-4885-4ffa-a368-87b680cd909c-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 06:34:32 crc kubenswrapper[4809]: I1206 06:34:32.431739 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qcx7r\" (UniqueName: \"kubernetes.io/projected/896d8df1-4885-4ffa-a368-87b680cd909c-kube-api-access-qcx7r\") on node \"crc\" DevicePath \"\"" Dec 06 06:34:32 crc kubenswrapper[4809]: I1206 06:34:32.632350 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mcdd9" event={"ID":"896d8df1-4885-4ffa-a368-87b680cd909c","Type":"ContainerDied","Data":"35417b3bd48b8e3dcdba77aa06dfcde9d1056ee3ea9faec2842e7788a8fc96fd"} Dec 06 06:34:32 crc kubenswrapper[4809]: I1206 06:34:32.632674 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="35417b3bd48b8e3dcdba77aa06dfcde9d1056ee3ea9faec2842e7788a8fc96fd" Dec 06 06:34:32 crc kubenswrapper[4809]: I1206 06:34:32.632480 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mcdd9" Dec 06 06:34:32 crc kubenswrapper[4809]: I1206 06:34:32.740216 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7tb8w"] Dec 06 06:34:32 crc kubenswrapper[4809]: E1206 06:34:32.740816 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="896d8df1-4885-4ffa-a368-87b680cd909c" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 06 06:34:32 crc kubenswrapper[4809]: I1206 06:34:32.740845 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="896d8df1-4885-4ffa-a368-87b680cd909c" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 06 06:34:32 crc kubenswrapper[4809]: I1206 06:34:32.741138 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="896d8df1-4885-4ffa-a368-87b680cd909c" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 06 06:34:32 crc kubenswrapper[4809]: I1206 06:34:32.742694 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7tb8w" Dec 06 06:34:32 crc kubenswrapper[4809]: I1206 06:34:32.746136 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 06:34:32 crc kubenswrapper[4809]: I1206 06:34:32.746595 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 06:34:32 crc kubenswrapper[4809]: I1206 06:34:32.746774 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-bzlf5" Dec 06 06:34:32 crc kubenswrapper[4809]: I1206 06:34:32.747616 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 06:34:32 crc kubenswrapper[4809]: I1206 06:34:32.764463 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7tb8w"] Dec 06 06:34:32 crc kubenswrapper[4809]: I1206 06:34:32.842789 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e254b2e6-7bfa-4244-b1ce-d336ae05fa90-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7tb8w\" (UID: \"e254b2e6-7bfa-4244-b1ce-d336ae05fa90\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7tb8w" Dec 06 06:34:32 crc kubenswrapper[4809]: I1206 06:34:32.842891 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e254b2e6-7bfa-4244-b1ce-d336ae05fa90-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7tb8w\" (UID: \"e254b2e6-7bfa-4244-b1ce-d336ae05fa90\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7tb8w" Dec 06 06:34:32 crc kubenswrapper[4809]: I1206 06:34:32.843427 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbwvp\" (UniqueName: \"kubernetes.io/projected/e254b2e6-7bfa-4244-b1ce-d336ae05fa90-kube-api-access-rbwvp\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7tb8w\" (UID: \"e254b2e6-7bfa-4244-b1ce-d336ae05fa90\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7tb8w" Dec 06 06:34:32 crc kubenswrapper[4809]: I1206 06:34:32.945438 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbwvp\" (UniqueName: \"kubernetes.io/projected/e254b2e6-7bfa-4244-b1ce-d336ae05fa90-kube-api-access-rbwvp\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7tb8w\" (UID: \"e254b2e6-7bfa-4244-b1ce-d336ae05fa90\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7tb8w" Dec 06 06:34:32 crc kubenswrapper[4809]: I1206 06:34:32.945634 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e254b2e6-7bfa-4244-b1ce-d336ae05fa90-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7tb8w\" (UID: \"e254b2e6-7bfa-4244-b1ce-d336ae05fa90\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7tb8w" Dec 06 06:34:32 crc kubenswrapper[4809]: I1206 06:34:32.945682 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e254b2e6-7bfa-4244-b1ce-d336ae05fa90-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7tb8w\" (UID: \"e254b2e6-7bfa-4244-b1ce-d336ae05fa90\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7tb8w" Dec 06 06:34:32 crc kubenswrapper[4809]: I1206 06:34:32.950207 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e254b2e6-7bfa-4244-b1ce-d336ae05fa90-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7tb8w\" (UID: \"e254b2e6-7bfa-4244-b1ce-d336ae05fa90\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7tb8w" Dec 06 06:34:32 crc kubenswrapper[4809]: I1206 06:34:32.952281 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e254b2e6-7bfa-4244-b1ce-d336ae05fa90-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7tb8w\" (UID: \"e254b2e6-7bfa-4244-b1ce-d336ae05fa90\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7tb8w" Dec 06 06:34:32 crc kubenswrapper[4809]: I1206 06:34:32.980384 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbwvp\" (UniqueName: \"kubernetes.io/projected/e254b2e6-7bfa-4244-b1ce-d336ae05fa90-kube-api-access-rbwvp\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7tb8w\" (UID: \"e254b2e6-7bfa-4244-b1ce-d336ae05fa90\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7tb8w" Dec 06 06:34:33 crc kubenswrapper[4809]: I1206 06:34:33.081760 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7tb8w" Dec 06 06:34:33 crc kubenswrapper[4809]: I1206 06:34:33.388455 4809 scope.go:117] "RemoveContainer" containerID="3aa8c940468b111de41fd7720fd216eebfc0bb377ed933226628430658845c32" Dec 06 06:34:33 crc kubenswrapper[4809]: E1206 06:34:33.389150 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:34:33 crc kubenswrapper[4809]: W1206 06:34:33.606209 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode254b2e6_7bfa_4244_b1ce_d336ae05fa90.slice/crio-c8c30a60cf2152642b6f5bee993e63907b62da49b707d23b11b8598ec92dfaa8 WatchSource:0}: Error finding container c8c30a60cf2152642b6f5bee993e63907b62da49b707d23b11b8598ec92dfaa8: Status 404 returned error can't find the container with id c8c30a60cf2152642b6f5bee993e63907b62da49b707d23b11b8598ec92dfaa8 Dec 06 06:34:33 crc kubenswrapper[4809]: I1206 06:34:33.611897 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7tb8w"] Dec 06 06:34:33 crc kubenswrapper[4809]: I1206 06:34:33.648252 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7tb8w" event={"ID":"e254b2e6-7bfa-4244-b1ce-d336ae05fa90","Type":"ContainerStarted","Data":"c8c30a60cf2152642b6f5bee993e63907b62da49b707d23b11b8598ec92dfaa8"} Dec 06 06:34:34 crc kubenswrapper[4809]: I1206 06:34:34.659918 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7tb8w" event={"ID":"e254b2e6-7bfa-4244-b1ce-d336ae05fa90","Type":"ContainerStarted","Data":"94180316e668480c9795647156b9934043b880dc12043a91b491c06841febdbf"} Dec 06 06:34:34 crc kubenswrapper[4809]: I1206 06:34:34.692021 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7tb8w" podStartSLOduration=2.148390698 podStartE2EDuration="2.69199996s" podCreationTimestamp="2025-12-06 06:34:32 +0000 UTC" firstStartedPulling="2025-12-06 06:34:33.608493755 +0000 UTC m=+2598.497476697" lastFinishedPulling="2025-12-06 06:34:34.152102987 +0000 UTC m=+2599.041085959" observedRunningTime="2025-12-06 06:34:34.683700355 +0000 UTC m=+2599.572683297" watchObservedRunningTime="2025-12-06 06:34:34.69199996 +0000 UTC m=+2599.580982902" Dec 06 06:34:35 crc kubenswrapper[4809]: I1206 06:34:35.177333 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-lqkvl" Dec 06 06:34:44 crc kubenswrapper[4809]: I1206 06:34:44.389348 4809 scope.go:117] "RemoveContainer" containerID="3aa8c940468b111de41fd7720fd216eebfc0bb377ed933226628430658845c32" Dec 06 06:34:44 crc kubenswrapper[4809]: E1206 06:34:44.390114 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:34:44 crc kubenswrapper[4809]: I1206 06:34:44.780398 4809 generic.go:334] "Generic (PLEG): container finished" podID="e254b2e6-7bfa-4244-b1ce-d336ae05fa90" containerID="94180316e668480c9795647156b9934043b880dc12043a91b491c06841febdbf" exitCode=0 Dec 06 06:34:44 crc kubenswrapper[4809]: I1206 06:34:44.780451 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7tb8w" event={"ID":"e254b2e6-7bfa-4244-b1ce-d336ae05fa90","Type":"ContainerDied","Data":"94180316e668480c9795647156b9934043b880dc12043a91b491c06841febdbf"} Dec 06 06:34:46 crc kubenswrapper[4809]: I1206 06:34:46.286349 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7tb8w" Dec 06 06:34:46 crc kubenswrapper[4809]: I1206 06:34:46.377876 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e254b2e6-7bfa-4244-b1ce-d336ae05fa90-inventory\") pod \"e254b2e6-7bfa-4244-b1ce-d336ae05fa90\" (UID: \"e254b2e6-7bfa-4244-b1ce-d336ae05fa90\") " Dec 06 06:34:46 crc kubenswrapper[4809]: I1206 06:34:46.378047 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e254b2e6-7bfa-4244-b1ce-d336ae05fa90-ssh-key\") pod \"e254b2e6-7bfa-4244-b1ce-d336ae05fa90\" (UID: \"e254b2e6-7bfa-4244-b1ce-d336ae05fa90\") " Dec 06 06:34:46 crc kubenswrapper[4809]: I1206 06:34:46.378165 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbwvp\" (UniqueName: \"kubernetes.io/projected/e254b2e6-7bfa-4244-b1ce-d336ae05fa90-kube-api-access-rbwvp\") pod \"e254b2e6-7bfa-4244-b1ce-d336ae05fa90\" (UID: \"e254b2e6-7bfa-4244-b1ce-d336ae05fa90\") " Dec 06 06:34:46 crc kubenswrapper[4809]: I1206 06:34:46.386472 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e254b2e6-7bfa-4244-b1ce-d336ae05fa90-kube-api-access-rbwvp" (OuterVolumeSpecName: "kube-api-access-rbwvp") pod "e254b2e6-7bfa-4244-b1ce-d336ae05fa90" (UID: "e254b2e6-7bfa-4244-b1ce-d336ae05fa90"). InnerVolumeSpecName "kube-api-access-rbwvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:34:46 crc kubenswrapper[4809]: I1206 06:34:46.411486 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e254b2e6-7bfa-4244-b1ce-d336ae05fa90-inventory" (OuterVolumeSpecName: "inventory") pod "e254b2e6-7bfa-4244-b1ce-d336ae05fa90" (UID: "e254b2e6-7bfa-4244-b1ce-d336ae05fa90"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:34:46 crc kubenswrapper[4809]: I1206 06:34:46.413208 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e254b2e6-7bfa-4244-b1ce-d336ae05fa90-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e254b2e6-7bfa-4244-b1ce-d336ae05fa90" (UID: "e254b2e6-7bfa-4244-b1ce-d336ae05fa90"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:34:46 crc kubenswrapper[4809]: I1206 06:34:46.484284 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e254b2e6-7bfa-4244-b1ce-d336ae05fa90-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 06:34:46 crc kubenswrapper[4809]: I1206 06:34:46.485054 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e254b2e6-7bfa-4244-b1ce-d336ae05fa90-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 06:34:46 crc kubenswrapper[4809]: I1206 06:34:46.485159 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbwvp\" (UniqueName: \"kubernetes.io/projected/e254b2e6-7bfa-4244-b1ce-d336ae05fa90-kube-api-access-rbwvp\") on node \"crc\" DevicePath \"\"" Dec 06 06:34:46 crc kubenswrapper[4809]: I1206 06:34:46.805045 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7tb8w" Dec 06 06:34:46 crc kubenswrapper[4809]: I1206 06:34:46.805088 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7tb8w" event={"ID":"e254b2e6-7bfa-4244-b1ce-d336ae05fa90","Type":"ContainerDied","Data":"c8c30a60cf2152642b6f5bee993e63907b62da49b707d23b11b8598ec92dfaa8"} Dec 06 06:34:46 crc kubenswrapper[4809]: I1206 06:34:46.805614 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c8c30a60cf2152642b6f5bee993e63907b62da49b707d23b11b8598ec92dfaa8" Dec 06 06:34:46 crc kubenswrapper[4809]: I1206 06:34:46.924979 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb"] Dec 06 06:34:46 crc kubenswrapper[4809]: E1206 06:34:46.925612 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e254b2e6-7bfa-4244-b1ce-d336ae05fa90" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 06 06:34:46 crc kubenswrapper[4809]: I1206 06:34:46.925638 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e254b2e6-7bfa-4244-b1ce-d336ae05fa90" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 06 06:34:46 crc kubenswrapper[4809]: I1206 06:34:46.959093 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="e254b2e6-7bfa-4244-b1ce-d336ae05fa90" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 06 06:34:46 crc kubenswrapper[4809]: I1206 06:34:46.960417 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb"] Dec 06 06:34:46 crc kubenswrapper[4809]: I1206 06:34:46.960533 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:46 crc kubenswrapper[4809]: I1206 06:34:46.964909 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 06:34:46 crc kubenswrapper[4809]: I1206 06:34:46.965186 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Dec 06 06:34:46 crc kubenswrapper[4809]: I1206 06:34:46.965791 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0" Dec 06 06:34:46 crc kubenswrapper[4809]: I1206 06:34:46.966005 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 06:34:46 crc kubenswrapper[4809]: I1206 06:34:46.966157 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 06:34:46 crc kubenswrapper[4809]: I1206 06:34:46.966280 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 06 06:34:46 crc kubenswrapper[4809]: I1206 06:34:46.968527 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 06 06:34:46 crc kubenswrapper[4809]: I1206 06:34:46.968977 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 06 06:34:46 crc kubenswrapper[4809]: I1206 06:34:46.969245 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-bzlf5" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.000073 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.000147 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.000182 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c2a5ffc7-6399-4cca-b2b1-d73542617259-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.000211 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.000242 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.000290 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrblk\" (UniqueName: \"kubernetes.io/projected/c2a5ffc7-6399-4cca-b2b1-d73542617259-kube-api-access-vrblk\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.000309 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c2a5ffc7-6399-4cca-b2b1-d73542617259-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.000338 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c2a5ffc7-6399-4cca-b2b1-d73542617259-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.000451 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.000563 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c2a5ffc7-6399-4cca-b2b1-d73542617259-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.000613 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c2a5ffc7-6399-4cca-b2b1-d73542617259-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.000665 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.000694 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.000747 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.000839 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.001068 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.102815 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c2a5ffc7-6399-4cca-b2b1-d73542617259-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.102902 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c2a5ffc7-6399-4cca-b2b1-d73542617259-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.102964 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.102996 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.103047 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.103144 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.103182 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.103246 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.103288 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.103326 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c2a5ffc7-6399-4cca-b2b1-d73542617259-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.103353 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.103394 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.103428 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrblk\" (UniqueName: \"kubernetes.io/projected/c2a5ffc7-6399-4cca-b2b1-d73542617259-kube-api-access-vrblk\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.103456 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c2a5ffc7-6399-4cca-b2b1-d73542617259-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.103496 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c2a5ffc7-6399-4cca-b2b1-d73542617259-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.103530 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.108071 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.108164 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.108734 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.109031 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.109461 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c2a5ffc7-6399-4cca-b2b1-d73542617259-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.110273 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.110548 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.111840 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c2a5ffc7-6399-4cca-b2b1-d73542617259-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.112260 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c2a5ffc7-6399-4cca-b2b1-d73542617259-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.113413 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c2a5ffc7-6399-4cca-b2b1-d73542617259-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.114671 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c2a5ffc7-6399-4cca-b2b1-d73542617259-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.116380 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.117567 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.121600 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.122006 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.130362 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrblk\" (UniqueName: \"kubernetes.io/projected/c2a5ffc7-6399-4cca-b2b1-d73542617259-kube-api-access-vrblk\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.291719 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:34:47 crc kubenswrapper[4809]: I1206 06:34:47.859994 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb"] Dec 06 06:34:48 crc kubenswrapper[4809]: I1206 06:34:48.828997 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" event={"ID":"c2a5ffc7-6399-4cca-b2b1-d73542617259","Type":"ContainerStarted","Data":"9880cb816fe5693bc33af85f316eaafe72112b207e89981bcfc8529b78b2c1b9"} Dec 06 06:34:48 crc kubenswrapper[4809]: I1206 06:34:48.829588 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" event={"ID":"c2a5ffc7-6399-4cca-b2b1-d73542617259","Type":"ContainerStarted","Data":"e83b4a50b28fdcd00807a3e2168f5b4c590f79125cdab6929d5872ae8353f62c"} Dec 06 06:34:48 crc kubenswrapper[4809]: I1206 06:34:48.863619 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" podStartSLOduration=2.231067038 podStartE2EDuration="2.863564497s" podCreationTimestamp="2025-12-06 06:34:46 +0000 UTC" firstStartedPulling="2025-12-06 06:34:47.871666473 +0000 UTC m=+2612.760649415" lastFinishedPulling="2025-12-06 06:34:48.504163932 +0000 UTC m=+2613.393146874" observedRunningTime="2025-12-06 06:34:48.846915396 +0000 UTC m=+2613.735898358" watchObservedRunningTime="2025-12-06 06:34:48.863564497 +0000 UTC m=+2613.752547449" Dec 06 06:34:59 crc kubenswrapper[4809]: I1206 06:34:59.390571 4809 scope.go:117] "RemoveContainer" containerID="3aa8c940468b111de41fd7720fd216eebfc0bb377ed933226628430658845c32" Dec 06 06:34:59 crc kubenswrapper[4809]: E1206 06:34:59.391407 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:35:04 crc kubenswrapper[4809]: I1206 06:35:04.088463 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mlcnp"] Dec 06 06:35:04 crc kubenswrapper[4809]: I1206 06:35:04.093830 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mlcnp" Dec 06 06:35:04 crc kubenswrapper[4809]: I1206 06:35:04.102907 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mlcnp"] Dec 06 06:35:04 crc kubenswrapper[4809]: I1206 06:35:04.144716 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32c90188-abb7-40d1-99bf-0befcec30ac2-catalog-content\") pod \"certified-operators-mlcnp\" (UID: \"32c90188-abb7-40d1-99bf-0befcec30ac2\") " pod="openshift-marketplace/certified-operators-mlcnp" Dec 06 06:35:04 crc kubenswrapper[4809]: I1206 06:35:04.144802 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qltnq\" (UniqueName: \"kubernetes.io/projected/32c90188-abb7-40d1-99bf-0befcec30ac2-kube-api-access-qltnq\") pod \"certified-operators-mlcnp\" (UID: \"32c90188-abb7-40d1-99bf-0befcec30ac2\") " pod="openshift-marketplace/certified-operators-mlcnp" Dec 06 06:35:04 crc kubenswrapper[4809]: I1206 06:35:04.145325 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32c90188-abb7-40d1-99bf-0befcec30ac2-utilities\") pod \"certified-operators-mlcnp\" (UID: \"32c90188-abb7-40d1-99bf-0befcec30ac2\") " pod="openshift-marketplace/certified-operators-mlcnp" Dec 06 06:35:04 crc kubenswrapper[4809]: I1206 06:35:04.247204 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32c90188-abb7-40d1-99bf-0befcec30ac2-catalog-content\") pod \"certified-operators-mlcnp\" (UID: \"32c90188-abb7-40d1-99bf-0befcec30ac2\") " pod="openshift-marketplace/certified-operators-mlcnp" Dec 06 06:35:04 crc kubenswrapper[4809]: I1206 06:35:04.247257 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qltnq\" (UniqueName: \"kubernetes.io/projected/32c90188-abb7-40d1-99bf-0befcec30ac2-kube-api-access-qltnq\") pod \"certified-operators-mlcnp\" (UID: \"32c90188-abb7-40d1-99bf-0befcec30ac2\") " pod="openshift-marketplace/certified-operators-mlcnp" Dec 06 06:35:04 crc kubenswrapper[4809]: I1206 06:35:04.247452 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32c90188-abb7-40d1-99bf-0befcec30ac2-utilities\") pod \"certified-operators-mlcnp\" (UID: \"32c90188-abb7-40d1-99bf-0befcec30ac2\") " pod="openshift-marketplace/certified-operators-mlcnp" Dec 06 06:35:04 crc kubenswrapper[4809]: I1206 06:35:04.249176 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32c90188-abb7-40d1-99bf-0befcec30ac2-utilities\") pod \"certified-operators-mlcnp\" (UID: \"32c90188-abb7-40d1-99bf-0befcec30ac2\") " pod="openshift-marketplace/certified-operators-mlcnp" Dec 06 06:35:04 crc kubenswrapper[4809]: I1206 06:35:04.249489 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32c90188-abb7-40d1-99bf-0befcec30ac2-catalog-content\") pod \"certified-operators-mlcnp\" (UID: \"32c90188-abb7-40d1-99bf-0befcec30ac2\") " pod="openshift-marketplace/certified-operators-mlcnp" Dec 06 06:35:04 crc kubenswrapper[4809]: I1206 06:35:04.278735 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qltnq\" (UniqueName: \"kubernetes.io/projected/32c90188-abb7-40d1-99bf-0befcec30ac2-kube-api-access-qltnq\") pod \"certified-operators-mlcnp\" (UID: \"32c90188-abb7-40d1-99bf-0befcec30ac2\") " pod="openshift-marketplace/certified-operators-mlcnp" Dec 06 06:35:04 crc kubenswrapper[4809]: I1206 06:35:04.424525 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mlcnp" Dec 06 06:35:04 crc kubenswrapper[4809]: I1206 06:35:04.978565 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mlcnp"] Dec 06 06:35:05 crc kubenswrapper[4809]: I1206 06:35:05.018207 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mlcnp" event={"ID":"32c90188-abb7-40d1-99bf-0befcec30ac2","Type":"ContainerStarted","Data":"c2b4202d7ab9779cd96697d2708b07af2bdb73142ff0d14b12d3911f330deadb"} Dec 06 06:35:06 crc kubenswrapper[4809]: I1206 06:35:06.030682 4809 generic.go:334] "Generic (PLEG): container finished" podID="32c90188-abb7-40d1-99bf-0befcec30ac2" containerID="11440c5db1afe46530a3136f3b18ddafe09977fe85313a4c21ebed9c02fd2ddb" exitCode=0 Dec 06 06:35:06 crc kubenswrapper[4809]: I1206 06:35:06.030793 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mlcnp" event={"ID":"32c90188-abb7-40d1-99bf-0befcec30ac2","Type":"ContainerDied","Data":"11440c5db1afe46530a3136f3b18ddafe09977fe85313a4c21ebed9c02fd2ddb"} Dec 06 06:35:08 crc kubenswrapper[4809]: I1206 06:35:08.053888 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mlcnp" event={"ID":"32c90188-abb7-40d1-99bf-0befcec30ac2","Type":"ContainerStarted","Data":"22bf2a5b8d97c6fdcc0485687ea2cbca5ae9fe781492260cff5a65bf916c0eda"} Dec 06 06:35:10 crc kubenswrapper[4809]: I1206 06:35:10.079095 4809 generic.go:334] "Generic (PLEG): container finished" podID="32c90188-abb7-40d1-99bf-0befcec30ac2" containerID="22bf2a5b8d97c6fdcc0485687ea2cbca5ae9fe781492260cff5a65bf916c0eda" exitCode=0 Dec 06 06:35:10 crc kubenswrapper[4809]: I1206 06:35:10.079163 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mlcnp" event={"ID":"32c90188-abb7-40d1-99bf-0befcec30ac2","Type":"ContainerDied","Data":"22bf2a5b8d97c6fdcc0485687ea2cbca5ae9fe781492260cff5a65bf916c0eda"} Dec 06 06:35:10 crc kubenswrapper[4809]: I1206 06:35:10.389053 4809 scope.go:117] "RemoveContainer" containerID="3aa8c940468b111de41fd7720fd216eebfc0bb377ed933226628430658845c32" Dec 06 06:35:10 crc kubenswrapper[4809]: E1206 06:35:10.389612 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:35:11 crc kubenswrapper[4809]: I1206 06:35:11.094288 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mlcnp" event={"ID":"32c90188-abb7-40d1-99bf-0befcec30ac2","Type":"ContainerStarted","Data":"992aa4b82f49d69fd9a7bead90e91b063960ca1e8d6c9c5c7eeec644b85848c3"} Dec 06 06:35:11 crc kubenswrapper[4809]: I1206 06:35:11.147871 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mlcnp" podStartSLOduration=2.698167863 podStartE2EDuration="7.147843448s" podCreationTimestamp="2025-12-06 06:35:04 +0000 UTC" firstStartedPulling="2025-12-06 06:35:06.0327044 +0000 UTC m=+2630.921687342" lastFinishedPulling="2025-12-06 06:35:10.482379985 +0000 UTC m=+2635.371362927" observedRunningTime="2025-12-06 06:35:11.118208755 +0000 UTC m=+2636.007191697" watchObservedRunningTime="2025-12-06 06:35:11.147843448 +0000 UTC m=+2636.036826390" Dec 06 06:35:14 crc kubenswrapper[4809]: I1206 06:35:14.425549 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mlcnp" Dec 06 06:35:14 crc kubenswrapper[4809]: I1206 06:35:14.426140 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mlcnp" Dec 06 06:35:14 crc kubenswrapper[4809]: I1206 06:35:14.490615 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mlcnp" Dec 06 06:35:15 crc kubenswrapper[4809]: I1206 06:35:15.220719 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mlcnp" Dec 06 06:35:15 crc kubenswrapper[4809]: I1206 06:35:15.298158 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mlcnp"] Dec 06 06:35:17 crc kubenswrapper[4809]: I1206 06:35:17.161107 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mlcnp" podUID="32c90188-abb7-40d1-99bf-0befcec30ac2" containerName="registry-server" containerID="cri-o://992aa4b82f49d69fd9a7bead90e91b063960ca1e8d6c9c5c7eeec644b85848c3" gracePeriod=2 Dec 06 06:35:17 crc kubenswrapper[4809]: I1206 06:35:17.728959 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mlcnp" Dec 06 06:35:17 crc kubenswrapper[4809]: I1206 06:35:17.921256 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qltnq\" (UniqueName: \"kubernetes.io/projected/32c90188-abb7-40d1-99bf-0befcec30ac2-kube-api-access-qltnq\") pod \"32c90188-abb7-40d1-99bf-0befcec30ac2\" (UID: \"32c90188-abb7-40d1-99bf-0befcec30ac2\") " Dec 06 06:35:17 crc kubenswrapper[4809]: I1206 06:35:17.921406 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32c90188-abb7-40d1-99bf-0befcec30ac2-catalog-content\") pod \"32c90188-abb7-40d1-99bf-0befcec30ac2\" (UID: \"32c90188-abb7-40d1-99bf-0befcec30ac2\") " Dec 06 06:35:17 crc kubenswrapper[4809]: I1206 06:35:17.921437 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32c90188-abb7-40d1-99bf-0befcec30ac2-utilities\") pod \"32c90188-abb7-40d1-99bf-0befcec30ac2\" (UID: \"32c90188-abb7-40d1-99bf-0befcec30ac2\") " Dec 06 06:35:17 crc kubenswrapper[4809]: I1206 06:35:17.922371 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32c90188-abb7-40d1-99bf-0befcec30ac2-utilities" (OuterVolumeSpecName: "utilities") pod "32c90188-abb7-40d1-99bf-0befcec30ac2" (UID: "32c90188-abb7-40d1-99bf-0befcec30ac2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:35:17 crc kubenswrapper[4809]: I1206 06:35:17.923089 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32c90188-abb7-40d1-99bf-0befcec30ac2-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:35:17 crc kubenswrapper[4809]: I1206 06:35:17.930270 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32c90188-abb7-40d1-99bf-0befcec30ac2-kube-api-access-qltnq" (OuterVolumeSpecName: "kube-api-access-qltnq") pod "32c90188-abb7-40d1-99bf-0befcec30ac2" (UID: "32c90188-abb7-40d1-99bf-0befcec30ac2"). InnerVolumeSpecName "kube-api-access-qltnq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:35:17 crc kubenswrapper[4809]: I1206 06:35:17.972800 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32c90188-abb7-40d1-99bf-0befcec30ac2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "32c90188-abb7-40d1-99bf-0befcec30ac2" (UID: "32c90188-abb7-40d1-99bf-0befcec30ac2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:35:18 crc kubenswrapper[4809]: I1206 06:35:18.026601 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32c90188-abb7-40d1-99bf-0befcec30ac2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:35:18 crc kubenswrapper[4809]: I1206 06:35:18.026631 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qltnq\" (UniqueName: \"kubernetes.io/projected/32c90188-abb7-40d1-99bf-0befcec30ac2-kube-api-access-qltnq\") on node \"crc\" DevicePath \"\"" Dec 06 06:35:18 crc kubenswrapper[4809]: I1206 06:35:18.173443 4809 generic.go:334] "Generic (PLEG): container finished" podID="32c90188-abb7-40d1-99bf-0befcec30ac2" containerID="992aa4b82f49d69fd9a7bead90e91b063960ca1e8d6c9c5c7eeec644b85848c3" exitCode=0 Dec 06 06:35:18 crc kubenswrapper[4809]: I1206 06:35:18.173492 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mlcnp" event={"ID":"32c90188-abb7-40d1-99bf-0befcec30ac2","Type":"ContainerDied","Data":"992aa4b82f49d69fd9a7bead90e91b063960ca1e8d6c9c5c7eeec644b85848c3"} Dec 06 06:35:18 crc kubenswrapper[4809]: I1206 06:35:18.173530 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mlcnp" Dec 06 06:35:18 crc kubenswrapper[4809]: I1206 06:35:18.173562 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mlcnp" event={"ID":"32c90188-abb7-40d1-99bf-0befcec30ac2","Type":"ContainerDied","Data":"c2b4202d7ab9779cd96697d2708b07af2bdb73142ff0d14b12d3911f330deadb"} Dec 06 06:35:18 crc kubenswrapper[4809]: I1206 06:35:18.173587 4809 scope.go:117] "RemoveContainer" containerID="992aa4b82f49d69fd9a7bead90e91b063960ca1e8d6c9c5c7eeec644b85848c3" Dec 06 06:35:18 crc kubenswrapper[4809]: I1206 06:35:18.199407 4809 scope.go:117] "RemoveContainer" containerID="22bf2a5b8d97c6fdcc0485687ea2cbca5ae9fe781492260cff5a65bf916c0eda" Dec 06 06:35:18 crc kubenswrapper[4809]: I1206 06:35:18.226069 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mlcnp"] Dec 06 06:35:18 crc kubenswrapper[4809]: I1206 06:35:18.243827 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mlcnp"] Dec 06 06:35:18 crc kubenswrapper[4809]: I1206 06:35:18.259362 4809 scope.go:117] "RemoveContainer" containerID="11440c5db1afe46530a3136f3b18ddafe09977fe85313a4c21ebed9c02fd2ddb" Dec 06 06:35:18 crc kubenswrapper[4809]: I1206 06:35:18.279432 4809 scope.go:117] "RemoveContainer" containerID="992aa4b82f49d69fd9a7bead90e91b063960ca1e8d6c9c5c7eeec644b85848c3" Dec 06 06:35:18 crc kubenswrapper[4809]: E1206 06:35:18.279899 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"992aa4b82f49d69fd9a7bead90e91b063960ca1e8d6c9c5c7eeec644b85848c3\": container with ID starting with 992aa4b82f49d69fd9a7bead90e91b063960ca1e8d6c9c5c7eeec644b85848c3 not found: ID does not exist" containerID="992aa4b82f49d69fd9a7bead90e91b063960ca1e8d6c9c5c7eeec644b85848c3" Dec 06 06:35:18 crc kubenswrapper[4809]: I1206 06:35:18.279945 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"992aa4b82f49d69fd9a7bead90e91b063960ca1e8d6c9c5c7eeec644b85848c3"} err="failed to get container status \"992aa4b82f49d69fd9a7bead90e91b063960ca1e8d6c9c5c7eeec644b85848c3\": rpc error: code = NotFound desc = could not find container \"992aa4b82f49d69fd9a7bead90e91b063960ca1e8d6c9c5c7eeec644b85848c3\": container with ID starting with 992aa4b82f49d69fd9a7bead90e91b063960ca1e8d6c9c5c7eeec644b85848c3 not found: ID does not exist" Dec 06 06:35:18 crc kubenswrapper[4809]: I1206 06:35:18.279966 4809 scope.go:117] "RemoveContainer" containerID="22bf2a5b8d97c6fdcc0485687ea2cbca5ae9fe781492260cff5a65bf916c0eda" Dec 06 06:35:18 crc kubenswrapper[4809]: E1206 06:35:18.280334 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22bf2a5b8d97c6fdcc0485687ea2cbca5ae9fe781492260cff5a65bf916c0eda\": container with ID starting with 22bf2a5b8d97c6fdcc0485687ea2cbca5ae9fe781492260cff5a65bf916c0eda not found: ID does not exist" containerID="22bf2a5b8d97c6fdcc0485687ea2cbca5ae9fe781492260cff5a65bf916c0eda" Dec 06 06:35:18 crc kubenswrapper[4809]: I1206 06:35:18.280505 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22bf2a5b8d97c6fdcc0485687ea2cbca5ae9fe781492260cff5a65bf916c0eda"} err="failed to get container status \"22bf2a5b8d97c6fdcc0485687ea2cbca5ae9fe781492260cff5a65bf916c0eda\": rpc error: code = NotFound desc = could not find container \"22bf2a5b8d97c6fdcc0485687ea2cbca5ae9fe781492260cff5a65bf916c0eda\": container with ID starting with 22bf2a5b8d97c6fdcc0485687ea2cbca5ae9fe781492260cff5a65bf916c0eda not found: ID does not exist" Dec 06 06:35:18 crc kubenswrapper[4809]: I1206 06:35:18.280658 4809 scope.go:117] "RemoveContainer" containerID="11440c5db1afe46530a3136f3b18ddafe09977fe85313a4c21ebed9c02fd2ddb" Dec 06 06:35:18 crc kubenswrapper[4809]: E1206 06:35:18.281119 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11440c5db1afe46530a3136f3b18ddafe09977fe85313a4c21ebed9c02fd2ddb\": container with ID starting with 11440c5db1afe46530a3136f3b18ddafe09977fe85313a4c21ebed9c02fd2ddb not found: ID does not exist" containerID="11440c5db1afe46530a3136f3b18ddafe09977fe85313a4c21ebed9c02fd2ddb" Dec 06 06:35:18 crc kubenswrapper[4809]: I1206 06:35:18.281145 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11440c5db1afe46530a3136f3b18ddafe09977fe85313a4c21ebed9c02fd2ddb"} err="failed to get container status \"11440c5db1afe46530a3136f3b18ddafe09977fe85313a4c21ebed9c02fd2ddb\": rpc error: code = NotFound desc = could not find container \"11440c5db1afe46530a3136f3b18ddafe09977fe85313a4c21ebed9c02fd2ddb\": container with ID starting with 11440c5db1afe46530a3136f3b18ddafe09977fe85313a4c21ebed9c02fd2ddb not found: ID does not exist" Dec 06 06:35:19 crc kubenswrapper[4809]: I1206 06:35:19.403339 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32c90188-abb7-40d1-99bf-0befcec30ac2" path="/var/lib/kubelet/pods/32c90188-abb7-40d1-99bf-0befcec30ac2/volumes" Dec 06 06:35:21 crc kubenswrapper[4809]: I1206 06:35:21.388984 4809 scope.go:117] "RemoveContainer" containerID="3aa8c940468b111de41fd7720fd216eebfc0bb377ed933226628430658845c32" Dec 06 06:35:21 crc kubenswrapper[4809]: E1206 06:35:21.389776 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:35:36 crc kubenswrapper[4809]: I1206 06:35:36.366786 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" event={"ID":"c2a5ffc7-6399-4cca-b2b1-d73542617259","Type":"ContainerDied","Data":"9880cb816fe5693bc33af85f316eaafe72112b207e89981bcfc8529b78b2c1b9"} Dec 06 06:35:36 crc kubenswrapper[4809]: I1206 06:35:36.366847 4809 generic.go:334] "Generic (PLEG): container finished" podID="c2a5ffc7-6399-4cca-b2b1-d73542617259" containerID="9880cb816fe5693bc33af85f316eaafe72112b207e89981bcfc8529b78b2c1b9" exitCode=0 Dec 06 06:35:36 crc kubenswrapper[4809]: I1206 06:35:36.392292 4809 scope.go:117] "RemoveContainer" containerID="3aa8c940468b111de41fd7720fd216eebfc0bb377ed933226628430658845c32" Dec 06 06:35:37 crc kubenswrapper[4809]: I1206 06:35:37.384629 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" event={"ID":"cbf4cf62-024e-4703-a8b8-9aecda9cd26a","Type":"ContainerStarted","Data":"96a97222b57fb6887f7526caa62d3152cf10b334fdc68d7a809ceb204d462e92"} Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.065297 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.121811 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-ovn-combined-ca-bundle\") pod \"c2a5ffc7-6399-4cca-b2b1-d73542617259\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.122139 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c2a5ffc7-6399-4cca-b2b1-d73542617259-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"c2a5ffc7-6399-4cca-b2b1-d73542617259\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.122199 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c2a5ffc7-6399-4cca-b2b1-d73542617259-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"c2a5ffc7-6399-4cca-b2b1-d73542617259\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.122274 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-libvirt-combined-ca-bundle\") pod \"c2a5ffc7-6399-4cca-b2b1-d73542617259\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.122312 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c2a5ffc7-6399-4cca-b2b1-d73542617259-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"c2a5ffc7-6399-4cca-b2b1-d73542617259\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.122351 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-ssh-key\") pod \"c2a5ffc7-6399-4cca-b2b1-d73542617259\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.122399 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-bootstrap-combined-ca-bundle\") pod \"c2a5ffc7-6399-4cca-b2b1-d73542617259\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.122428 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-inventory\") pod \"c2a5ffc7-6399-4cca-b2b1-d73542617259\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.122462 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c2a5ffc7-6399-4cca-b2b1-d73542617259-openstack-edpm-ipam-ovn-default-certs-0\") pod \"c2a5ffc7-6399-4cca-b2b1-d73542617259\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.122506 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-telemetry-combined-ca-bundle\") pod \"c2a5ffc7-6399-4cca-b2b1-d73542617259\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.122530 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-neutron-metadata-combined-ca-bundle\") pod \"c2a5ffc7-6399-4cca-b2b1-d73542617259\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.122560 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrblk\" (UniqueName: \"kubernetes.io/projected/c2a5ffc7-6399-4cca-b2b1-d73542617259-kube-api-access-vrblk\") pod \"c2a5ffc7-6399-4cca-b2b1-d73542617259\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.122585 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-repo-setup-combined-ca-bundle\") pod \"c2a5ffc7-6399-4cca-b2b1-d73542617259\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.122666 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-telemetry-power-monitoring-combined-ca-bundle\") pod \"c2a5ffc7-6399-4cca-b2b1-d73542617259\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.122698 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-nova-combined-ca-bundle\") pod \"c2a5ffc7-6399-4cca-b2b1-d73542617259\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.122746 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c2a5ffc7-6399-4cca-b2b1-d73542617259-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"c2a5ffc7-6399-4cca-b2b1-d73542617259\" (UID: \"c2a5ffc7-6399-4cca-b2b1-d73542617259\") " Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.129493 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "c2a5ffc7-6399-4cca-b2b1-d73542617259" (UID: "c2a5ffc7-6399-4cca-b2b1-d73542617259"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.130229 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "c2a5ffc7-6399-4cca-b2b1-d73542617259" (UID: "c2a5ffc7-6399-4cca-b2b1-d73542617259"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.130314 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2a5ffc7-6399-4cca-b2b1-d73542617259-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0") pod "c2a5ffc7-6399-4cca-b2b1-d73542617259" (UID: "c2a5ffc7-6399-4cca-b2b1-d73542617259"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.130343 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2a5ffc7-6399-4cca-b2b1-d73542617259-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "c2a5ffc7-6399-4cca-b2b1-d73542617259" (UID: "c2a5ffc7-6399-4cca-b2b1-d73542617259"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.131989 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2a5ffc7-6399-4cca-b2b1-d73542617259-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "c2a5ffc7-6399-4cca-b2b1-d73542617259" (UID: "c2a5ffc7-6399-4cca-b2b1-d73542617259"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.132817 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2a5ffc7-6399-4cca-b2b1-d73542617259-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "c2a5ffc7-6399-4cca-b2b1-d73542617259" (UID: "c2a5ffc7-6399-4cca-b2b1-d73542617259"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.132929 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2a5ffc7-6399-4cca-b2b1-d73542617259-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "c2a5ffc7-6399-4cca-b2b1-d73542617259" (UID: "c2a5ffc7-6399-4cca-b2b1-d73542617259"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.142052 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "c2a5ffc7-6399-4cca-b2b1-d73542617259" (UID: "c2a5ffc7-6399-4cca-b2b1-d73542617259"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.146379 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "c2a5ffc7-6399-4cca-b2b1-d73542617259" (UID: "c2a5ffc7-6399-4cca-b2b1-d73542617259"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.147114 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "c2a5ffc7-6399-4cca-b2b1-d73542617259" (UID: "c2a5ffc7-6399-4cca-b2b1-d73542617259"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.147654 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "c2a5ffc7-6399-4cca-b2b1-d73542617259" (UID: "c2a5ffc7-6399-4cca-b2b1-d73542617259"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.147752 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2a5ffc7-6399-4cca-b2b1-d73542617259-kube-api-access-vrblk" (OuterVolumeSpecName: "kube-api-access-vrblk") pod "c2a5ffc7-6399-4cca-b2b1-d73542617259" (UID: "c2a5ffc7-6399-4cca-b2b1-d73542617259"). InnerVolumeSpecName "kube-api-access-vrblk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.148437 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "c2a5ffc7-6399-4cca-b2b1-d73542617259" (UID: "c2a5ffc7-6399-4cca-b2b1-d73542617259"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.165188 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-telemetry-power-monitoring-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-power-monitoring-combined-ca-bundle") pod "c2a5ffc7-6399-4cca-b2b1-d73542617259" (UID: "c2a5ffc7-6399-4cca-b2b1-d73542617259"). InnerVolumeSpecName "telemetry-power-monitoring-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.177361 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c2a5ffc7-6399-4cca-b2b1-d73542617259" (UID: "c2a5ffc7-6399-4cca-b2b1-d73542617259"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.181142 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-inventory" (OuterVolumeSpecName: "inventory") pod "c2a5ffc7-6399-4cca-b2b1-d73542617259" (UID: "c2a5ffc7-6399-4cca-b2b1-d73542617259"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.227254 4809 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.227287 4809 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c2a5ffc7-6399-4cca-b2b1-d73542617259-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.227298 4809 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c2a5ffc7-6399-4cca-b2b1-d73542617259-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.227309 4809 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.227319 4809 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c2a5ffc7-6399-4cca-b2b1-d73542617259-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.227333 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.227344 4809 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.227355 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.227367 4809 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c2a5ffc7-6399-4cca-b2b1-d73542617259-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.227384 4809 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.227431 4809 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.227447 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrblk\" (UniqueName: \"kubernetes.io/projected/c2a5ffc7-6399-4cca-b2b1-d73542617259-kube-api-access-vrblk\") on node \"crc\" DevicePath \"\"" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.227462 4809 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.227471 4809 reconciler_common.go:293] "Volume detached for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-telemetry-power-monitoring-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.227503 4809 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a5ffc7-6399-4cca-b2b1-d73542617259-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.227514 4809 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c2a5ffc7-6399-4cca-b2b1-d73542617259-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.399034 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" event={"ID":"c2a5ffc7-6399-4cca-b2b1-d73542617259","Type":"ContainerDied","Data":"e83b4a50b28fdcd00807a3e2168f5b4c590f79125cdab6929d5872ae8353f62c"} Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.399526 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e83b4a50b28fdcd00807a3e2168f5b4c590f79125cdab6929d5872ae8353f62c" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.399114 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.499976 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-xm4wm"] Dec 06 06:35:38 crc kubenswrapper[4809]: E1206 06:35:38.500742 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32c90188-abb7-40d1-99bf-0befcec30ac2" containerName="registry-server" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.501000 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="32c90188-abb7-40d1-99bf-0befcec30ac2" containerName="registry-server" Dec 06 06:35:38 crc kubenswrapper[4809]: E1206 06:35:38.501123 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32c90188-abb7-40d1-99bf-0befcec30ac2" containerName="extract-utilities" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.501181 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="32c90188-abb7-40d1-99bf-0befcec30ac2" containerName="extract-utilities" Dec 06 06:35:38 crc kubenswrapper[4809]: E1206 06:35:38.501253 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2a5ffc7-6399-4cca-b2b1-d73542617259" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.501324 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2a5ffc7-6399-4cca-b2b1-d73542617259" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 06 06:35:38 crc kubenswrapper[4809]: E1206 06:35:38.501405 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32c90188-abb7-40d1-99bf-0befcec30ac2" containerName="extract-content" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.501487 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="32c90188-abb7-40d1-99bf-0befcec30ac2" containerName="extract-content" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.501858 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2a5ffc7-6399-4cca-b2b1-d73542617259" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.502014 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="32c90188-abb7-40d1-99bf-0befcec30ac2" containerName="registry-server" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.503365 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xm4wm" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.508495 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.508728 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.508847 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.508974 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.509085 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-bzlf5" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.514641 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-xm4wm"] Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.534689 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/1fd335d6-64e4-4bb6-8617-4a08cfc65951-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xm4wm\" (UID: \"1fd335d6-64e4-4bb6-8617-4a08cfc65951\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xm4wm" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.534789 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1fd335d6-64e4-4bb6-8617-4a08cfc65951-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xm4wm\" (UID: \"1fd335d6-64e4-4bb6-8617-4a08cfc65951\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xm4wm" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.534969 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1fd335d6-64e4-4bb6-8617-4a08cfc65951-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xm4wm\" (UID: \"1fd335d6-64e4-4bb6-8617-4a08cfc65951\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xm4wm" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.535000 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tcttq\" (UniqueName: \"kubernetes.io/projected/1fd335d6-64e4-4bb6-8617-4a08cfc65951-kube-api-access-tcttq\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xm4wm\" (UID: \"1fd335d6-64e4-4bb6-8617-4a08cfc65951\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xm4wm" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.535037 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fd335d6-64e4-4bb6-8617-4a08cfc65951-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xm4wm\" (UID: \"1fd335d6-64e4-4bb6-8617-4a08cfc65951\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xm4wm" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.637125 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1fd335d6-64e4-4bb6-8617-4a08cfc65951-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xm4wm\" (UID: \"1fd335d6-64e4-4bb6-8617-4a08cfc65951\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xm4wm" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.637168 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tcttq\" (UniqueName: \"kubernetes.io/projected/1fd335d6-64e4-4bb6-8617-4a08cfc65951-kube-api-access-tcttq\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xm4wm\" (UID: \"1fd335d6-64e4-4bb6-8617-4a08cfc65951\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xm4wm" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.637198 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fd335d6-64e4-4bb6-8617-4a08cfc65951-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xm4wm\" (UID: \"1fd335d6-64e4-4bb6-8617-4a08cfc65951\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xm4wm" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.637270 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/1fd335d6-64e4-4bb6-8617-4a08cfc65951-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xm4wm\" (UID: \"1fd335d6-64e4-4bb6-8617-4a08cfc65951\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xm4wm" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.637314 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1fd335d6-64e4-4bb6-8617-4a08cfc65951-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xm4wm\" (UID: \"1fd335d6-64e4-4bb6-8617-4a08cfc65951\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xm4wm" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.638504 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/1fd335d6-64e4-4bb6-8617-4a08cfc65951-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xm4wm\" (UID: \"1fd335d6-64e4-4bb6-8617-4a08cfc65951\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xm4wm" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.641573 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1fd335d6-64e4-4bb6-8617-4a08cfc65951-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xm4wm\" (UID: \"1fd335d6-64e4-4bb6-8617-4a08cfc65951\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xm4wm" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.642385 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1fd335d6-64e4-4bb6-8617-4a08cfc65951-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xm4wm\" (UID: \"1fd335d6-64e4-4bb6-8617-4a08cfc65951\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xm4wm" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.643024 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fd335d6-64e4-4bb6-8617-4a08cfc65951-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xm4wm\" (UID: \"1fd335d6-64e4-4bb6-8617-4a08cfc65951\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xm4wm" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.660312 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tcttq\" (UniqueName: \"kubernetes.io/projected/1fd335d6-64e4-4bb6-8617-4a08cfc65951-kube-api-access-tcttq\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xm4wm\" (UID: \"1fd335d6-64e4-4bb6-8617-4a08cfc65951\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xm4wm" Dec 06 06:35:38 crc kubenswrapper[4809]: I1206 06:35:38.832964 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xm4wm" Dec 06 06:35:39 crc kubenswrapper[4809]: I1206 06:35:39.433811 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-xm4wm"] Dec 06 06:35:39 crc kubenswrapper[4809]: W1206 06:35:39.434677 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1fd335d6_64e4_4bb6_8617_4a08cfc65951.slice/crio-fcfbbc0e4bf3104da689a834ce91c503584ff0a2ad21b91fba96c6e484e66c59 WatchSource:0}: Error finding container fcfbbc0e4bf3104da689a834ce91c503584ff0a2ad21b91fba96c6e484e66c59: Status 404 returned error can't find the container with id fcfbbc0e4bf3104da689a834ce91c503584ff0a2ad21b91fba96c6e484e66c59 Dec 06 06:35:39 crc kubenswrapper[4809]: I1206 06:35:39.438366 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 06:35:40 crc kubenswrapper[4809]: I1206 06:35:40.433238 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xm4wm" event={"ID":"1fd335d6-64e4-4bb6-8617-4a08cfc65951","Type":"ContainerStarted","Data":"43022267bf6d2eca5618c89856bde192f3420533aaf1532739b27e8e1d7ffb66"} Dec 06 06:35:40 crc kubenswrapper[4809]: I1206 06:35:40.433876 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xm4wm" event={"ID":"1fd335d6-64e4-4bb6-8617-4a08cfc65951","Type":"ContainerStarted","Data":"fcfbbc0e4bf3104da689a834ce91c503584ff0a2ad21b91fba96c6e484e66c59"} Dec 06 06:35:40 crc kubenswrapper[4809]: I1206 06:35:40.456258 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xm4wm" podStartSLOduration=2.030099024 podStartE2EDuration="2.456237208s" podCreationTimestamp="2025-12-06 06:35:38 +0000 UTC" firstStartedPulling="2025-12-06 06:35:39.438163909 +0000 UTC m=+2664.327146851" lastFinishedPulling="2025-12-06 06:35:39.864302093 +0000 UTC m=+2664.753285035" observedRunningTime="2025-12-06 06:35:40.446916327 +0000 UTC m=+2665.335899269" watchObservedRunningTime="2025-12-06 06:35:40.456237208 +0000 UTC m=+2665.345220150" Dec 06 06:36:46 crc kubenswrapper[4809]: I1206 06:36:46.230611 4809 generic.go:334] "Generic (PLEG): container finished" podID="1fd335d6-64e4-4bb6-8617-4a08cfc65951" containerID="43022267bf6d2eca5618c89856bde192f3420533aaf1532739b27e8e1d7ffb66" exitCode=0 Dec 06 06:36:46 crc kubenswrapper[4809]: I1206 06:36:46.230676 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xm4wm" event={"ID":"1fd335d6-64e4-4bb6-8617-4a08cfc65951","Type":"ContainerDied","Data":"43022267bf6d2eca5618c89856bde192f3420533aaf1532739b27e8e1d7ffb66"} Dec 06 06:36:47 crc kubenswrapper[4809]: I1206 06:36:47.732372 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xm4wm" Dec 06 06:36:47 crc kubenswrapper[4809]: I1206 06:36:47.909101 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1fd335d6-64e4-4bb6-8617-4a08cfc65951-ssh-key\") pod \"1fd335d6-64e4-4bb6-8617-4a08cfc65951\" (UID: \"1fd335d6-64e4-4bb6-8617-4a08cfc65951\") " Dec 06 06:36:47 crc kubenswrapper[4809]: I1206 06:36:47.909279 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/1fd335d6-64e4-4bb6-8617-4a08cfc65951-ovncontroller-config-0\") pod \"1fd335d6-64e4-4bb6-8617-4a08cfc65951\" (UID: \"1fd335d6-64e4-4bb6-8617-4a08cfc65951\") " Dec 06 06:36:47 crc kubenswrapper[4809]: I1206 06:36:47.909404 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1fd335d6-64e4-4bb6-8617-4a08cfc65951-inventory\") pod \"1fd335d6-64e4-4bb6-8617-4a08cfc65951\" (UID: \"1fd335d6-64e4-4bb6-8617-4a08cfc65951\") " Dec 06 06:36:47 crc kubenswrapper[4809]: I1206 06:36:47.909514 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fd335d6-64e4-4bb6-8617-4a08cfc65951-ovn-combined-ca-bundle\") pod \"1fd335d6-64e4-4bb6-8617-4a08cfc65951\" (UID: \"1fd335d6-64e4-4bb6-8617-4a08cfc65951\") " Dec 06 06:36:47 crc kubenswrapper[4809]: I1206 06:36:47.909606 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tcttq\" (UniqueName: \"kubernetes.io/projected/1fd335d6-64e4-4bb6-8617-4a08cfc65951-kube-api-access-tcttq\") pod \"1fd335d6-64e4-4bb6-8617-4a08cfc65951\" (UID: \"1fd335d6-64e4-4bb6-8617-4a08cfc65951\") " Dec 06 06:36:47 crc kubenswrapper[4809]: I1206 06:36:47.915072 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fd335d6-64e4-4bb6-8617-4a08cfc65951-kube-api-access-tcttq" (OuterVolumeSpecName: "kube-api-access-tcttq") pod "1fd335d6-64e4-4bb6-8617-4a08cfc65951" (UID: "1fd335d6-64e4-4bb6-8617-4a08cfc65951"). InnerVolumeSpecName "kube-api-access-tcttq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:36:47 crc kubenswrapper[4809]: I1206 06:36:47.916563 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tcttq\" (UniqueName: \"kubernetes.io/projected/1fd335d6-64e4-4bb6-8617-4a08cfc65951-kube-api-access-tcttq\") on node \"crc\" DevicePath \"\"" Dec 06 06:36:47 crc kubenswrapper[4809]: I1206 06:36:47.920454 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fd335d6-64e4-4bb6-8617-4a08cfc65951-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "1fd335d6-64e4-4bb6-8617-4a08cfc65951" (UID: "1fd335d6-64e4-4bb6-8617-4a08cfc65951"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:36:47 crc kubenswrapper[4809]: I1206 06:36:47.943847 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1fd335d6-64e4-4bb6-8617-4a08cfc65951-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "1fd335d6-64e4-4bb6-8617-4a08cfc65951" (UID: "1fd335d6-64e4-4bb6-8617-4a08cfc65951"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:36:47 crc kubenswrapper[4809]: I1206 06:36:47.948310 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fd335d6-64e4-4bb6-8617-4a08cfc65951-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1fd335d6-64e4-4bb6-8617-4a08cfc65951" (UID: "1fd335d6-64e4-4bb6-8617-4a08cfc65951"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:36:47 crc kubenswrapper[4809]: I1206 06:36:47.958035 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fd335d6-64e4-4bb6-8617-4a08cfc65951-inventory" (OuterVolumeSpecName: "inventory") pod "1fd335d6-64e4-4bb6-8617-4a08cfc65951" (UID: "1fd335d6-64e4-4bb6-8617-4a08cfc65951"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:36:48 crc kubenswrapper[4809]: I1206 06:36:48.019543 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1fd335d6-64e4-4bb6-8617-4a08cfc65951-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 06:36:48 crc kubenswrapper[4809]: I1206 06:36:48.019592 4809 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/1fd335d6-64e4-4bb6-8617-4a08cfc65951-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:36:48 crc kubenswrapper[4809]: I1206 06:36:48.019610 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1fd335d6-64e4-4bb6-8617-4a08cfc65951-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 06:36:48 crc kubenswrapper[4809]: I1206 06:36:48.019624 4809 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fd335d6-64e4-4bb6-8617-4a08cfc65951-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:36:48 crc kubenswrapper[4809]: I1206 06:36:48.254045 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xm4wm" event={"ID":"1fd335d6-64e4-4bb6-8617-4a08cfc65951","Type":"ContainerDied","Data":"fcfbbc0e4bf3104da689a834ce91c503584ff0a2ad21b91fba96c6e484e66c59"} Dec 06 06:36:48 crc kubenswrapper[4809]: I1206 06:36:48.254359 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fcfbbc0e4bf3104da689a834ce91c503584ff0a2ad21b91fba96c6e484e66c59" Dec 06 06:36:48 crc kubenswrapper[4809]: I1206 06:36:48.254279 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xm4wm" Dec 06 06:36:48 crc kubenswrapper[4809]: I1206 06:36:48.364993 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5"] Dec 06 06:36:48 crc kubenswrapper[4809]: E1206 06:36:48.365744 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fd335d6-64e4-4bb6-8617-4a08cfc65951" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 06 06:36:48 crc kubenswrapper[4809]: I1206 06:36:48.365762 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fd335d6-64e4-4bb6-8617-4a08cfc65951" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 06 06:36:48 crc kubenswrapper[4809]: I1206 06:36:48.366099 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="1fd335d6-64e4-4bb6-8617-4a08cfc65951" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 06 06:36:48 crc kubenswrapper[4809]: I1206 06:36:48.367125 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5" Dec 06 06:36:48 crc kubenswrapper[4809]: I1206 06:36:48.378258 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 06 06:36:48 crc kubenswrapper[4809]: I1206 06:36:48.380519 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5"] Dec 06 06:36:48 crc kubenswrapper[4809]: I1206 06:36:48.385488 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 06:36:48 crc kubenswrapper[4809]: I1206 06:36:48.385577 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 06 06:36:48 crc kubenswrapper[4809]: I1206 06:36:48.385596 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 06:36:48 crc kubenswrapper[4809]: I1206 06:36:48.385957 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-bzlf5" Dec 06 06:36:48 crc kubenswrapper[4809]: I1206 06:36:48.386128 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 06:36:48 crc kubenswrapper[4809]: I1206 06:36:48.527804 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb2b0f4f-7e1b-4e7d-a376-50a53f24879e-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5\" (UID: \"eb2b0f4f-7e1b-4e7d-a376-50a53f24879e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5" Dec 06 06:36:48 crc kubenswrapper[4809]: I1206 06:36:48.528020 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/eb2b0f4f-7e1b-4e7d-a376-50a53f24879e-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5\" (UID: \"eb2b0f4f-7e1b-4e7d-a376-50a53f24879e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5" Dec 06 06:36:48 crc kubenswrapper[4809]: I1206 06:36:48.528042 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxvgs\" (UniqueName: \"kubernetes.io/projected/eb2b0f4f-7e1b-4e7d-a376-50a53f24879e-kube-api-access-zxvgs\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5\" (UID: \"eb2b0f4f-7e1b-4e7d-a376-50a53f24879e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5" Dec 06 06:36:48 crc kubenswrapper[4809]: I1206 06:36:48.528063 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eb2b0f4f-7e1b-4e7d-a376-50a53f24879e-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5\" (UID: \"eb2b0f4f-7e1b-4e7d-a376-50a53f24879e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5" Dec 06 06:36:48 crc kubenswrapper[4809]: I1206 06:36:48.528110 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eb2b0f4f-7e1b-4e7d-a376-50a53f24879e-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5\" (UID: \"eb2b0f4f-7e1b-4e7d-a376-50a53f24879e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5" Dec 06 06:36:48 crc kubenswrapper[4809]: I1206 06:36:48.528158 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/eb2b0f4f-7e1b-4e7d-a376-50a53f24879e-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5\" (UID: \"eb2b0f4f-7e1b-4e7d-a376-50a53f24879e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5" Dec 06 06:36:48 crc kubenswrapper[4809]: I1206 06:36:48.630475 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/eb2b0f4f-7e1b-4e7d-a376-50a53f24879e-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5\" (UID: \"eb2b0f4f-7e1b-4e7d-a376-50a53f24879e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5" Dec 06 06:36:48 crc kubenswrapper[4809]: I1206 06:36:48.630844 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxvgs\" (UniqueName: \"kubernetes.io/projected/eb2b0f4f-7e1b-4e7d-a376-50a53f24879e-kube-api-access-zxvgs\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5\" (UID: \"eb2b0f4f-7e1b-4e7d-a376-50a53f24879e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5" Dec 06 06:36:48 crc kubenswrapper[4809]: I1206 06:36:48.630884 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eb2b0f4f-7e1b-4e7d-a376-50a53f24879e-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5\" (UID: \"eb2b0f4f-7e1b-4e7d-a376-50a53f24879e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5" Dec 06 06:36:48 crc kubenswrapper[4809]: I1206 06:36:48.630968 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eb2b0f4f-7e1b-4e7d-a376-50a53f24879e-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5\" (UID: \"eb2b0f4f-7e1b-4e7d-a376-50a53f24879e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5" Dec 06 06:36:48 crc kubenswrapper[4809]: I1206 06:36:48.631039 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/eb2b0f4f-7e1b-4e7d-a376-50a53f24879e-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5\" (UID: \"eb2b0f4f-7e1b-4e7d-a376-50a53f24879e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5" Dec 06 06:36:48 crc kubenswrapper[4809]: I1206 06:36:48.631124 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb2b0f4f-7e1b-4e7d-a376-50a53f24879e-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5\" (UID: \"eb2b0f4f-7e1b-4e7d-a376-50a53f24879e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5" Dec 06 06:36:48 crc kubenswrapper[4809]: I1206 06:36:48.635838 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/eb2b0f4f-7e1b-4e7d-a376-50a53f24879e-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5\" (UID: \"eb2b0f4f-7e1b-4e7d-a376-50a53f24879e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5" Dec 06 06:36:48 crc kubenswrapper[4809]: I1206 06:36:48.636209 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/eb2b0f4f-7e1b-4e7d-a376-50a53f24879e-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5\" (UID: \"eb2b0f4f-7e1b-4e7d-a376-50a53f24879e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5" Dec 06 06:36:48 crc kubenswrapper[4809]: I1206 06:36:48.636359 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eb2b0f4f-7e1b-4e7d-a376-50a53f24879e-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5\" (UID: \"eb2b0f4f-7e1b-4e7d-a376-50a53f24879e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5" Dec 06 06:36:48 crc kubenswrapper[4809]: I1206 06:36:48.636744 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eb2b0f4f-7e1b-4e7d-a376-50a53f24879e-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5\" (UID: \"eb2b0f4f-7e1b-4e7d-a376-50a53f24879e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5" Dec 06 06:36:48 crc kubenswrapper[4809]: I1206 06:36:48.637239 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb2b0f4f-7e1b-4e7d-a376-50a53f24879e-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5\" (UID: \"eb2b0f4f-7e1b-4e7d-a376-50a53f24879e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5" Dec 06 06:36:48 crc kubenswrapper[4809]: I1206 06:36:48.652372 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxvgs\" (UniqueName: \"kubernetes.io/projected/eb2b0f4f-7e1b-4e7d-a376-50a53f24879e-kube-api-access-zxvgs\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5\" (UID: \"eb2b0f4f-7e1b-4e7d-a376-50a53f24879e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5" Dec 06 06:36:48 crc kubenswrapper[4809]: I1206 06:36:48.711718 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5" Dec 06 06:36:51 crc kubenswrapper[4809]: I1206 06:36:51.175742 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5"] Dec 06 06:36:51 crc kubenswrapper[4809]: I1206 06:36:51.299751 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5" event={"ID":"eb2b0f4f-7e1b-4e7d-a376-50a53f24879e","Type":"ContainerStarted","Data":"085183c7db72ab93798788f808ce55bae78c4a318035879cb410298b22f8241d"} Dec 06 06:36:53 crc kubenswrapper[4809]: I1206 06:36:53.327588 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5" event={"ID":"eb2b0f4f-7e1b-4e7d-a376-50a53f24879e","Type":"ContainerStarted","Data":"940d1f7442d404d7354b27af91d2210bdddd501e86d930145e3a64c441dc1df9"} Dec 06 06:36:53 crc kubenswrapper[4809]: I1206 06:36:53.364519 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5" podStartSLOduration=4.086054106 podStartE2EDuration="5.364458709s" podCreationTimestamp="2025-12-06 06:36:48 +0000 UTC" firstStartedPulling="2025-12-06 06:36:51.185710175 +0000 UTC m=+2736.074693117" lastFinishedPulling="2025-12-06 06:36:52.464114778 +0000 UTC m=+2737.353097720" observedRunningTime="2025-12-06 06:36:53.349503974 +0000 UTC m=+2738.238486926" watchObservedRunningTime="2025-12-06 06:36:53.364458709 +0000 UTC m=+2738.253441651" Dec 06 06:37:14 crc kubenswrapper[4809]: I1206 06:37:14.599277 4809 generic.go:334] "Generic (PLEG): container finished" podID="19fc8500-fddb-4650-839a-82166e98a567" containerID="fb4c1d3b58b0b0436f44913e8e182a2234dd3fde8bf8f57bff9bc4c4237b6157" exitCode=1 Dec 06 06:37:14 crc kubenswrapper[4809]: I1206 06:37:14.600120 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-697657f76c-5stmg" event={"ID":"19fc8500-fddb-4650-839a-82166e98a567","Type":"ContainerDied","Data":"fb4c1d3b58b0b0436f44913e8e182a2234dd3fde8bf8f57bff9bc4c4237b6157"} Dec 06 06:37:14 crc kubenswrapper[4809]: I1206 06:37:14.602905 4809 scope.go:117] "RemoveContainer" containerID="fb4c1d3b58b0b0436f44913e8e182a2234dd3fde8bf8f57bff9bc4c4237b6157" Dec 06 06:37:15 crc kubenswrapper[4809]: I1206 06:37:15.038881 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators-redhat/loki-operator-controller-manager-697657f76c-5stmg" Dec 06 06:37:15 crc kubenswrapper[4809]: I1206 06:37:15.615045 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-697657f76c-5stmg" event={"ID":"19fc8500-fddb-4650-839a-82166e98a567","Type":"ContainerStarted","Data":"9d5fccea66b2135c94b5e915974f3adc7a7689dbd251769b353e65801f02f8b2"} Dec 06 06:37:15 crc kubenswrapper[4809]: I1206 06:37:15.615421 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators-redhat/loki-operator-controller-manager-697657f76c-5stmg" Dec 06 06:37:25 crc kubenswrapper[4809]: I1206 06:37:25.043792 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators-redhat/loki-operator-controller-manager-697657f76c-5stmg" Dec 06 06:37:30 crc kubenswrapper[4809]: I1206 06:37:30.685131 4809 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-qqv6x container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.30:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 06:37:30 crc kubenswrapper[4809]: I1206 06:37:30.685652 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qqv6x" podUID="8a225b5c-4ad5-4f08-a0cb-ae38a8fa83d4" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.30:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 06:37:35 crc kubenswrapper[4809]: I1206 06:37:35.646688 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7trbt" podUID="45f87079-e852-4b47-a240-6f2bcf247509" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.105:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 06:37:55 crc kubenswrapper[4809]: I1206 06:37:55.035950 4809 generic.go:334] "Generic (PLEG): container finished" podID="eb2b0f4f-7e1b-4e7d-a376-50a53f24879e" containerID="940d1f7442d404d7354b27af91d2210bdddd501e86d930145e3a64c441dc1df9" exitCode=0 Dec 06 06:37:55 crc kubenswrapper[4809]: I1206 06:37:55.035971 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5" event={"ID":"eb2b0f4f-7e1b-4e7d-a376-50a53f24879e","Type":"ContainerDied","Data":"940d1f7442d404d7354b27af91d2210bdddd501e86d930145e3a64c441dc1df9"} Dec 06 06:37:56 crc kubenswrapper[4809]: I1206 06:37:56.603243 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5" Dec 06 06:37:56 crc kubenswrapper[4809]: I1206 06:37:56.690079 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/eb2b0f4f-7e1b-4e7d-a376-50a53f24879e-nova-metadata-neutron-config-0\") pod \"eb2b0f4f-7e1b-4e7d-a376-50a53f24879e\" (UID: \"eb2b0f4f-7e1b-4e7d-a376-50a53f24879e\") " Dec 06 06:37:56 crc kubenswrapper[4809]: I1206 06:37:56.690117 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eb2b0f4f-7e1b-4e7d-a376-50a53f24879e-ssh-key\") pod \"eb2b0f4f-7e1b-4e7d-a376-50a53f24879e\" (UID: \"eb2b0f4f-7e1b-4e7d-a376-50a53f24879e\") " Dec 06 06:37:56 crc kubenswrapper[4809]: I1206 06:37:56.690258 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb2b0f4f-7e1b-4e7d-a376-50a53f24879e-neutron-metadata-combined-ca-bundle\") pod \"eb2b0f4f-7e1b-4e7d-a376-50a53f24879e\" (UID: \"eb2b0f4f-7e1b-4e7d-a376-50a53f24879e\") " Dec 06 06:37:56 crc kubenswrapper[4809]: I1206 06:37:56.690313 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/eb2b0f4f-7e1b-4e7d-a376-50a53f24879e-neutron-ovn-metadata-agent-neutron-config-0\") pod \"eb2b0f4f-7e1b-4e7d-a376-50a53f24879e\" (UID: \"eb2b0f4f-7e1b-4e7d-a376-50a53f24879e\") " Dec 06 06:37:56 crc kubenswrapper[4809]: I1206 06:37:56.690362 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxvgs\" (UniqueName: \"kubernetes.io/projected/eb2b0f4f-7e1b-4e7d-a376-50a53f24879e-kube-api-access-zxvgs\") pod \"eb2b0f4f-7e1b-4e7d-a376-50a53f24879e\" (UID: \"eb2b0f4f-7e1b-4e7d-a376-50a53f24879e\") " Dec 06 06:37:56 crc kubenswrapper[4809]: I1206 06:37:56.690464 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eb2b0f4f-7e1b-4e7d-a376-50a53f24879e-inventory\") pod \"eb2b0f4f-7e1b-4e7d-a376-50a53f24879e\" (UID: \"eb2b0f4f-7e1b-4e7d-a376-50a53f24879e\") " Dec 06 06:37:56 crc kubenswrapper[4809]: I1206 06:37:56.696436 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb2b0f4f-7e1b-4e7d-a376-50a53f24879e-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "eb2b0f4f-7e1b-4e7d-a376-50a53f24879e" (UID: "eb2b0f4f-7e1b-4e7d-a376-50a53f24879e"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:37:56 crc kubenswrapper[4809]: I1206 06:37:56.705384 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb2b0f4f-7e1b-4e7d-a376-50a53f24879e-kube-api-access-zxvgs" (OuterVolumeSpecName: "kube-api-access-zxvgs") pod "eb2b0f4f-7e1b-4e7d-a376-50a53f24879e" (UID: "eb2b0f4f-7e1b-4e7d-a376-50a53f24879e"). InnerVolumeSpecName "kube-api-access-zxvgs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:37:56 crc kubenswrapper[4809]: I1206 06:37:56.725207 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb2b0f4f-7e1b-4e7d-a376-50a53f24879e-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "eb2b0f4f-7e1b-4e7d-a376-50a53f24879e" (UID: "eb2b0f4f-7e1b-4e7d-a376-50a53f24879e"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:37:56 crc kubenswrapper[4809]: I1206 06:37:56.731733 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb2b0f4f-7e1b-4e7d-a376-50a53f24879e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "eb2b0f4f-7e1b-4e7d-a376-50a53f24879e" (UID: "eb2b0f4f-7e1b-4e7d-a376-50a53f24879e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:37:56 crc kubenswrapper[4809]: I1206 06:37:56.733269 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb2b0f4f-7e1b-4e7d-a376-50a53f24879e-inventory" (OuterVolumeSpecName: "inventory") pod "eb2b0f4f-7e1b-4e7d-a376-50a53f24879e" (UID: "eb2b0f4f-7e1b-4e7d-a376-50a53f24879e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:37:56 crc kubenswrapper[4809]: I1206 06:37:56.737214 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb2b0f4f-7e1b-4e7d-a376-50a53f24879e-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "eb2b0f4f-7e1b-4e7d-a376-50a53f24879e" (UID: "eb2b0f4f-7e1b-4e7d-a376-50a53f24879e"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:37:56 crc kubenswrapper[4809]: I1206 06:37:56.793224 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxvgs\" (UniqueName: \"kubernetes.io/projected/eb2b0f4f-7e1b-4e7d-a376-50a53f24879e-kube-api-access-zxvgs\") on node \"crc\" DevicePath \"\"" Dec 06 06:37:56 crc kubenswrapper[4809]: I1206 06:37:56.793263 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eb2b0f4f-7e1b-4e7d-a376-50a53f24879e-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 06:37:56 crc kubenswrapper[4809]: I1206 06:37:56.793275 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eb2b0f4f-7e1b-4e7d-a376-50a53f24879e-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 06:37:56 crc kubenswrapper[4809]: I1206 06:37:56.793289 4809 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/eb2b0f4f-7e1b-4e7d-a376-50a53f24879e-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:37:56 crc kubenswrapper[4809]: I1206 06:37:56.793304 4809 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb2b0f4f-7e1b-4e7d-a376-50a53f24879e-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:37:56 crc kubenswrapper[4809]: I1206 06:37:56.793320 4809 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/eb2b0f4f-7e1b-4e7d-a376-50a53f24879e-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:37:57 crc kubenswrapper[4809]: I1206 06:37:57.059126 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5" event={"ID":"eb2b0f4f-7e1b-4e7d-a376-50a53f24879e","Type":"ContainerDied","Data":"085183c7db72ab93798788f808ce55bae78c4a318035879cb410298b22f8241d"} Dec 06 06:37:57 crc kubenswrapper[4809]: I1206 06:37:57.059462 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="085183c7db72ab93798788f808ce55bae78c4a318035879cb410298b22f8241d" Dec 06 06:37:57 crc kubenswrapper[4809]: I1206 06:37:57.059200 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5" Dec 06 06:37:57 crc kubenswrapper[4809]: I1206 06:37:57.173016 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmgn9"] Dec 06 06:37:57 crc kubenswrapper[4809]: E1206 06:37:57.173546 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb2b0f4f-7e1b-4e7d-a376-50a53f24879e" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 06 06:37:57 crc kubenswrapper[4809]: I1206 06:37:57.173572 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb2b0f4f-7e1b-4e7d-a376-50a53f24879e" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 06 06:37:57 crc kubenswrapper[4809]: I1206 06:37:57.173864 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb2b0f4f-7e1b-4e7d-a376-50a53f24879e" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 06 06:37:57 crc kubenswrapper[4809]: I1206 06:37:57.176086 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmgn9" Dec 06 06:37:57 crc kubenswrapper[4809]: I1206 06:37:57.178743 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 06 06:37:57 crc kubenswrapper[4809]: I1206 06:37:57.178817 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 06:37:57 crc kubenswrapper[4809]: I1206 06:37:57.180822 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-bzlf5" Dec 06 06:37:57 crc kubenswrapper[4809]: I1206 06:37:57.180915 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 06:37:57 crc kubenswrapper[4809]: I1206 06:37:57.181063 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 06:37:57 crc kubenswrapper[4809]: I1206 06:37:57.215815 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmgn9"] Dec 06 06:37:57 crc kubenswrapper[4809]: I1206 06:37:57.318227 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/c4b57a32-8e04-4fbe-864d-387273af423f-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lmgn9\" (UID: \"c4b57a32-8e04-4fbe-864d-387273af423f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmgn9" Dec 06 06:37:57 crc kubenswrapper[4809]: I1206 06:37:57.318331 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xwnx\" (UniqueName: \"kubernetes.io/projected/c4b57a32-8e04-4fbe-864d-387273af423f-kube-api-access-6xwnx\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lmgn9\" (UID: \"c4b57a32-8e04-4fbe-864d-387273af423f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmgn9" Dec 06 06:37:57 crc kubenswrapper[4809]: I1206 06:37:57.318376 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c4b57a32-8e04-4fbe-864d-387273af423f-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lmgn9\" (UID: \"c4b57a32-8e04-4fbe-864d-387273af423f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmgn9" Dec 06 06:37:57 crc kubenswrapper[4809]: I1206 06:37:57.318410 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4b57a32-8e04-4fbe-864d-387273af423f-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lmgn9\" (UID: \"c4b57a32-8e04-4fbe-864d-387273af423f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmgn9" Dec 06 06:37:57 crc kubenswrapper[4809]: I1206 06:37:57.318736 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c4b57a32-8e04-4fbe-864d-387273af423f-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lmgn9\" (UID: \"c4b57a32-8e04-4fbe-864d-387273af423f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmgn9" Dec 06 06:37:57 crc kubenswrapper[4809]: I1206 06:37:57.423580 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xwnx\" (UniqueName: \"kubernetes.io/projected/c4b57a32-8e04-4fbe-864d-387273af423f-kube-api-access-6xwnx\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lmgn9\" (UID: \"c4b57a32-8e04-4fbe-864d-387273af423f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmgn9" Dec 06 06:37:57 crc kubenswrapper[4809]: I1206 06:37:57.423689 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c4b57a32-8e04-4fbe-864d-387273af423f-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lmgn9\" (UID: \"c4b57a32-8e04-4fbe-864d-387273af423f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmgn9" Dec 06 06:37:57 crc kubenswrapper[4809]: I1206 06:37:57.423741 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4b57a32-8e04-4fbe-864d-387273af423f-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lmgn9\" (UID: \"c4b57a32-8e04-4fbe-864d-387273af423f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmgn9" Dec 06 06:37:57 crc kubenswrapper[4809]: I1206 06:37:57.424001 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c4b57a32-8e04-4fbe-864d-387273af423f-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lmgn9\" (UID: \"c4b57a32-8e04-4fbe-864d-387273af423f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmgn9" Dec 06 06:37:57 crc kubenswrapper[4809]: I1206 06:37:57.424405 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/c4b57a32-8e04-4fbe-864d-387273af423f-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lmgn9\" (UID: \"c4b57a32-8e04-4fbe-864d-387273af423f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmgn9" Dec 06 06:37:57 crc kubenswrapper[4809]: I1206 06:37:57.428427 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/c4b57a32-8e04-4fbe-864d-387273af423f-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lmgn9\" (UID: \"c4b57a32-8e04-4fbe-864d-387273af423f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmgn9" Dec 06 06:37:57 crc kubenswrapper[4809]: I1206 06:37:57.428467 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4b57a32-8e04-4fbe-864d-387273af423f-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lmgn9\" (UID: \"c4b57a32-8e04-4fbe-864d-387273af423f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmgn9" Dec 06 06:37:57 crc kubenswrapper[4809]: I1206 06:37:57.430101 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c4b57a32-8e04-4fbe-864d-387273af423f-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lmgn9\" (UID: \"c4b57a32-8e04-4fbe-864d-387273af423f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmgn9" Dec 06 06:37:57 crc kubenswrapper[4809]: I1206 06:37:57.433050 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c4b57a32-8e04-4fbe-864d-387273af423f-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lmgn9\" (UID: \"c4b57a32-8e04-4fbe-864d-387273af423f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmgn9" Dec 06 06:37:57 crc kubenswrapper[4809]: I1206 06:37:57.443846 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xwnx\" (UniqueName: \"kubernetes.io/projected/c4b57a32-8e04-4fbe-864d-387273af423f-kube-api-access-6xwnx\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lmgn9\" (UID: \"c4b57a32-8e04-4fbe-864d-387273af423f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmgn9" Dec 06 06:37:57 crc kubenswrapper[4809]: I1206 06:37:57.504002 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmgn9" Dec 06 06:37:58 crc kubenswrapper[4809]: I1206 06:37:58.317870 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmgn9"] Dec 06 06:37:59 crc kubenswrapper[4809]: I1206 06:37:59.082207 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmgn9" event={"ID":"c4b57a32-8e04-4fbe-864d-387273af423f","Type":"ContainerStarted","Data":"367cb841eb6e329a056a41254f0f7126c1d029d32bf41987fe37ce6ce1a83f3e"} Dec 06 06:38:00 crc kubenswrapper[4809]: I1206 06:38:00.093493 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmgn9" event={"ID":"c4b57a32-8e04-4fbe-864d-387273af423f","Type":"ContainerStarted","Data":"3f1ee8401752600c08a780800491544e46437e10a61712ed115a912edb66200a"} Dec 06 06:38:00 crc kubenswrapper[4809]: I1206 06:38:00.111532 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmgn9" podStartSLOduration=2.6713213639999998 podStartE2EDuration="3.111510589s" podCreationTimestamp="2025-12-06 06:37:57 +0000 UTC" firstStartedPulling="2025-12-06 06:37:58.355116017 +0000 UTC m=+2803.244098959" lastFinishedPulling="2025-12-06 06:37:58.795305242 +0000 UTC m=+2803.684288184" observedRunningTime="2025-12-06 06:38:00.109289818 +0000 UTC m=+2804.998272760" watchObservedRunningTime="2025-12-06 06:38:00.111510589 +0000 UTC m=+2805.000493531" Dec 06 06:38:04 crc kubenswrapper[4809]: I1206 06:38:04.497207 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:38:04 crc kubenswrapper[4809]: I1206 06:38:04.497868 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:38:32 crc kubenswrapper[4809]: I1206 06:38:32.783798 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vz8vw"] Dec 06 06:38:32 crc kubenswrapper[4809]: I1206 06:38:32.787744 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vz8vw" Dec 06 06:38:32 crc kubenswrapper[4809]: I1206 06:38:32.797468 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vz8vw"] Dec 06 06:38:32 crc kubenswrapper[4809]: I1206 06:38:32.966575 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53ff183f-11d1-4232-ad78-5bdd33a2c180-catalog-content\") pod \"community-operators-vz8vw\" (UID: \"53ff183f-11d1-4232-ad78-5bdd33a2c180\") " pod="openshift-marketplace/community-operators-vz8vw" Dec 06 06:38:32 crc kubenswrapper[4809]: I1206 06:38:32.966630 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2p92v\" (UniqueName: \"kubernetes.io/projected/53ff183f-11d1-4232-ad78-5bdd33a2c180-kube-api-access-2p92v\") pod \"community-operators-vz8vw\" (UID: \"53ff183f-11d1-4232-ad78-5bdd33a2c180\") " pod="openshift-marketplace/community-operators-vz8vw" Dec 06 06:38:32 crc kubenswrapper[4809]: I1206 06:38:32.966673 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53ff183f-11d1-4232-ad78-5bdd33a2c180-utilities\") pod \"community-operators-vz8vw\" (UID: \"53ff183f-11d1-4232-ad78-5bdd33a2c180\") " pod="openshift-marketplace/community-operators-vz8vw" Dec 06 06:38:33 crc kubenswrapper[4809]: I1206 06:38:33.069089 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53ff183f-11d1-4232-ad78-5bdd33a2c180-utilities\") pod \"community-operators-vz8vw\" (UID: \"53ff183f-11d1-4232-ad78-5bdd33a2c180\") " pod="openshift-marketplace/community-operators-vz8vw" Dec 06 06:38:33 crc kubenswrapper[4809]: I1206 06:38:33.069582 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53ff183f-11d1-4232-ad78-5bdd33a2c180-catalog-content\") pod \"community-operators-vz8vw\" (UID: \"53ff183f-11d1-4232-ad78-5bdd33a2c180\") " pod="openshift-marketplace/community-operators-vz8vw" Dec 06 06:38:33 crc kubenswrapper[4809]: I1206 06:38:33.069670 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2p92v\" (UniqueName: \"kubernetes.io/projected/53ff183f-11d1-4232-ad78-5bdd33a2c180-kube-api-access-2p92v\") pod \"community-operators-vz8vw\" (UID: \"53ff183f-11d1-4232-ad78-5bdd33a2c180\") " pod="openshift-marketplace/community-operators-vz8vw" Dec 06 06:38:33 crc kubenswrapper[4809]: I1206 06:38:33.070046 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53ff183f-11d1-4232-ad78-5bdd33a2c180-catalog-content\") pod \"community-operators-vz8vw\" (UID: \"53ff183f-11d1-4232-ad78-5bdd33a2c180\") " pod="openshift-marketplace/community-operators-vz8vw" Dec 06 06:38:33 crc kubenswrapper[4809]: I1206 06:38:33.070132 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53ff183f-11d1-4232-ad78-5bdd33a2c180-utilities\") pod \"community-operators-vz8vw\" (UID: \"53ff183f-11d1-4232-ad78-5bdd33a2c180\") " pod="openshift-marketplace/community-operators-vz8vw" Dec 06 06:38:33 crc kubenswrapper[4809]: I1206 06:38:33.090063 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2p92v\" (UniqueName: \"kubernetes.io/projected/53ff183f-11d1-4232-ad78-5bdd33a2c180-kube-api-access-2p92v\") pod \"community-operators-vz8vw\" (UID: \"53ff183f-11d1-4232-ad78-5bdd33a2c180\") " pod="openshift-marketplace/community-operators-vz8vw" Dec 06 06:38:33 crc kubenswrapper[4809]: I1206 06:38:33.111224 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vz8vw" Dec 06 06:38:33 crc kubenswrapper[4809]: I1206 06:38:33.777179 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vz8vw"] Dec 06 06:38:34 crc kubenswrapper[4809]: I1206 06:38:34.496150 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:38:34 crc kubenswrapper[4809]: I1206 06:38:34.496488 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:38:34 crc kubenswrapper[4809]: I1206 06:38:34.502128 4809 generic.go:334] "Generic (PLEG): container finished" podID="53ff183f-11d1-4232-ad78-5bdd33a2c180" containerID="8e1df3efd957fd03f1bbe3c93e84dbd268f74c04b8f566e341b049858c1518f3" exitCode=0 Dec 06 06:38:34 crc kubenswrapper[4809]: I1206 06:38:34.502203 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vz8vw" event={"ID":"53ff183f-11d1-4232-ad78-5bdd33a2c180","Type":"ContainerDied","Data":"8e1df3efd957fd03f1bbe3c93e84dbd268f74c04b8f566e341b049858c1518f3"} Dec 06 06:38:34 crc kubenswrapper[4809]: I1206 06:38:34.502386 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vz8vw" event={"ID":"53ff183f-11d1-4232-ad78-5bdd33a2c180","Type":"ContainerStarted","Data":"066de22473e7d338bcfefb8c26deab2167388aef1cd659c5da0f2982c5050316"} Dec 06 06:38:35 crc kubenswrapper[4809]: I1206 06:38:35.516746 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vz8vw" event={"ID":"53ff183f-11d1-4232-ad78-5bdd33a2c180","Type":"ContainerStarted","Data":"965fa12055c28f45592b6559e68f27128ddf5c297e726a755231961189ec1bac"} Dec 06 06:38:37 crc kubenswrapper[4809]: I1206 06:38:37.541242 4809 generic.go:334] "Generic (PLEG): container finished" podID="53ff183f-11d1-4232-ad78-5bdd33a2c180" containerID="965fa12055c28f45592b6559e68f27128ddf5c297e726a755231961189ec1bac" exitCode=0 Dec 06 06:38:37 crc kubenswrapper[4809]: I1206 06:38:37.541325 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vz8vw" event={"ID":"53ff183f-11d1-4232-ad78-5bdd33a2c180","Type":"ContainerDied","Data":"965fa12055c28f45592b6559e68f27128ddf5c297e726a755231961189ec1bac"} Dec 06 06:38:38 crc kubenswrapper[4809]: I1206 06:38:38.554562 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vz8vw" event={"ID":"53ff183f-11d1-4232-ad78-5bdd33a2c180","Type":"ContainerStarted","Data":"aa97e8ec541b70cdebbe530fdff77e80cf8bc276132aae479b066d6e58fbc1e7"} Dec 06 06:38:38 crc kubenswrapper[4809]: I1206 06:38:38.578106 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vz8vw" podStartSLOduration=2.904529346 podStartE2EDuration="6.578087185s" podCreationTimestamp="2025-12-06 06:38:32 +0000 UTC" firstStartedPulling="2025-12-06 06:38:34.504170551 +0000 UTC m=+2839.393153493" lastFinishedPulling="2025-12-06 06:38:38.17772839 +0000 UTC m=+2843.066711332" observedRunningTime="2025-12-06 06:38:38.569744379 +0000 UTC m=+2843.458727341" watchObservedRunningTime="2025-12-06 06:38:38.578087185 +0000 UTC m=+2843.467070127" Dec 06 06:38:43 crc kubenswrapper[4809]: I1206 06:38:43.112248 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vz8vw" Dec 06 06:38:43 crc kubenswrapper[4809]: I1206 06:38:43.112720 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vz8vw" Dec 06 06:38:43 crc kubenswrapper[4809]: I1206 06:38:43.178533 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vz8vw" Dec 06 06:38:43 crc kubenswrapper[4809]: I1206 06:38:43.668584 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vz8vw" Dec 06 06:38:48 crc kubenswrapper[4809]: I1206 06:38:48.897721 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vz8vw"] Dec 06 06:38:48 crc kubenswrapper[4809]: I1206 06:38:48.898590 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vz8vw" podUID="53ff183f-11d1-4232-ad78-5bdd33a2c180" containerName="registry-server" containerID="cri-o://aa97e8ec541b70cdebbe530fdff77e80cf8bc276132aae479b066d6e58fbc1e7" gracePeriod=2 Dec 06 06:38:49 crc kubenswrapper[4809]: I1206 06:38:49.521863 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vz8vw" Dec 06 06:38:49 crc kubenswrapper[4809]: I1206 06:38:49.605402 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53ff183f-11d1-4232-ad78-5bdd33a2c180-catalog-content\") pod \"53ff183f-11d1-4232-ad78-5bdd33a2c180\" (UID: \"53ff183f-11d1-4232-ad78-5bdd33a2c180\") " Dec 06 06:38:49 crc kubenswrapper[4809]: I1206 06:38:49.605477 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53ff183f-11d1-4232-ad78-5bdd33a2c180-utilities\") pod \"53ff183f-11d1-4232-ad78-5bdd33a2c180\" (UID: \"53ff183f-11d1-4232-ad78-5bdd33a2c180\") " Dec 06 06:38:49 crc kubenswrapper[4809]: I1206 06:38:49.605641 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2p92v\" (UniqueName: \"kubernetes.io/projected/53ff183f-11d1-4232-ad78-5bdd33a2c180-kube-api-access-2p92v\") pod \"53ff183f-11d1-4232-ad78-5bdd33a2c180\" (UID: \"53ff183f-11d1-4232-ad78-5bdd33a2c180\") " Dec 06 06:38:49 crc kubenswrapper[4809]: I1206 06:38:49.606527 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53ff183f-11d1-4232-ad78-5bdd33a2c180-utilities" (OuterVolumeSpecName: "utilities") pod "53ff183f-11d1-4232-ad78-5bdd33a2c180" (UID: "53ff183f-11d1-4232-ad78-5bdd33a2c180"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:38:49 crc kubenswrapper[4809]: I1206 06:38:49.623507 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53ff183f-11d1-4232-ad78-5bdd33a2c180-kube-api-access-2p92v" (OuterVolumeSpecName: "kube-api-access-2p92v") pod "53ff183f-11d1-4232-ad78-5bdd33a2c180" (UID: "53ff183f-11d1-4232-ad78-5bdd33a2c180"). InnerVolumeSpecName "kube-api-access-2p92v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:38:49 crc kubenswrapper[4809]: I1206 06:38:49.661081 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53ff183f-11d1-4232-ad78-5bdd33a2c180-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "53ff183f-11d1-4232-ad78-5bdd33a2c180" (UID: "53ff183f-11d1-4232-ad78-5bdd33a2c180"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:38:49 crc kubenswrapper[4809]: I1206 06:38:49.686249 4809 generic.go:334] "Generic (PLEG): container finished" podID="53ff183f-11d1-4232-ad78-5bdd33a2c180" containerID="aa97e8ec541b70cdebbe530fdff77e80cf8bc276132aae479b066d6e58fbc1e7" exitCode=0 Dec 06 06:38:49 crc kubenswrapper[4809]: I1206 06:38:49.686324 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vz8vw" Dec 06 06:38:49 crc kubenswrapper[4809]: I1206 06:38:49.686346 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vz8vw" event={"ID":"53ff183f-11d1-4232-ad78-5bdd33a2c180","Type":"ContainerDied","Data":"aa97e8ec541b70cdebbe530fdff77e80cf8bc276132aae479b066d6e58fbc1e7"} Dec 06 06:38:49 crc kubenswrapper[4809]: I1206 06:38:49.686686 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vz8vw" event={"ID":"53ff183f-11d1-4232-ad78-5bdd33a2c180","Type":"ContainerDied","Data":"066de22473e7d338bcfefb8c26deab2167388aef1cd659c5da0f2982c5050316"} Dec 06 06:38:49 crc kubenswrapper[4809]: I1206 06:38:49.686711 4809 scope.go:117] "RemoveContainer" containerID="aa97e8ec541b70cdebbe530fdff77e80cf8bc276132aae479b066d6e58fbc1e7" Dec 06 06:38:49 crc kubenswrapper[4809]: I1206 06:38:49.709383 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2p92v\" (UniqueName: \"kubernetes.io/projected/53ff183f-11d1-4232-ad78-5bdd33a2c180-kube-api-access-2p92v\") on node \"crc\" DevicePath \"\"" Dec 06 06:38:49 crc kubenswrapper[4809]: I1206 06:38:49.709425 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53ff183f-11d1-4232-ad78-5bdd33a2c180-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:38:49 crc kubenswrapper[4809]: I1206 06:38:49.709437 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53ff183f-11d1-4232-ad78-5bdd33a2c180-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:38:49 crc kubenswrapper[4809]: I1206 06:38:49.713237 4809 scope.go:117] "RemoveContainer" containerID="965fa12055c28f45592b6559e68f27128ddf5c297e726a755231961189ec1bac" Dec 06 06:38:49 crc kubenswrapper[4809]: I1206 06:38:49.723756 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vz8vw"] Dec 06 06:38:49 crc kubenswrapper[4809]: I1206 06:38:49.734349 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vz8vw"] Dec 06 06:38:49 crc kubenswrapper[4809]: I1206 06:38:49.773295 4809 scope.go:117] "RemoveContainer" containerID="8e1df3efd957fd03f1bbe3c93e84dbd268f74c04b8f566e341b049858c1518f3" Dec 06 06:38:49 crc kubenswrapper[4809]: I1206 06:38:49.796522 4809 scope.go:117] "RemoveContainer" containerID="aa97e8ec541b70cdebbe530fdff77e80cf8bc276132aae479b066d6e58fbc1e7" Dec 06 06:38:49 crc kubenswrapper[4809]: E1206 06:38:49.796877 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa97e8ec541b70cdebbe530fdff77e80cf8bc276132aae479b066d6e58fbc1e7\": container with ID starting with aa97e8ec541b70cdebbe530fdff77e80cf8bc276132aae479b066d6e58fbc1e7 not found: ID does not exist" containerID="aa97e8ec541b70cdebbe530fdff77e80cf8bc276132aae479b066d6e58fbc1e7" Dec 06 06:38:49 crc kubenswrapper[4809]: I1206 06:38:49.796906 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa97e8ec541b70cdebbe530fdff77e80cf8bc276132aae479b066d6e58fbc1e7"} err="failed to get container status \"aa97e8ec541b70cdebbe530fdff77e80cf8bc276132aae479b066d6e58fbc1e7\": rpc error: code = NotFound desc = could not find container \"aa97e8ec541b70cdebbe530fdff77e80cf8bc276132aae479b066d6e58fbc1e7\": container with ID starting with aa97e8ec541b70cdebbe530fdff77e80cf8bc276132aae479b066d6e58fbc1e7 not found: ID does not exist" Dec 06 06:38:49 crc kubenswrapper[4809]: I1206 06:38:49.796928 4809 scope.go:117] "RemoveContainer" containerID="965fa12055c28f45592b6559e68f27128ddf5c297e726a755231961189ec1bac" Dec 06 06:38:49 crc kubenswrapper[4809]: E1206 06:38:49.797211 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"965fa12055c28f45592b6559e68f27128ddf5c297e726a755231961189ec1bac\": container with ID starting with 965fa12055c28f45592b6559e68f27128ddf5c297e726a755231961189ec1bac not found: ID does not exist" containerID="965fa12055c28f45592b6559e68f27128ddf5c297e726a755231961189ec1bac" Dec 06 06:38:49 crc kubenswrapper[4809]: I1206 06:38:49.797229 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"965fa12055c28f45592b6559e68f27128ddf5c297e726a755231961189ec1bac"} err="failed to get container status \"965fa12055c28f45592b6559e68f27128ddf5c297e726a755231961189ec1bac\": rpc error: code = NotFound desc = could not find container \"965fa12055c28f45592b6559e68f27128ddf5c297e726a755231961189ec1bac\": container with ID starting with 965fa12055c28f45592b6559e68f27128ddf5c297e726a755231961189ec1bac not found: ID does not exist" Dec 06 06:38:49 crc kubenswrapper[4809]: I1206 06:38:49.797243 4809 scope.go:117] "RemoveContainer" containerID="8e1df3efd957fd03f1bbe3c93e84dbd268f74c04b8f566e341b049858c1518f3" Dec 06 06:38:49 crc kubenswrapper[4809]: E1206 06:38:49.797497 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e1df3efd957fd03f1bbe3c93e84dbd268f74c04b8f566e341b049858c1518f3\": container with ID starting with 8e1df3efd957fd03f1bbe3c93e84dbd268f74c04b8f566e341b049858c1518f3 not found: ID does not exist" containerID="8e1df3efd957fd03f1bbe3c93e84dbd268f74c04b8f566e341b049858c1518f3" Dec 06 06:38:49 crc kubenswrapper[4809]: I1206 06:38:49.797521 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e1df3efd957fd03f1bbe3c93e84dbd268f74c04b8f566e341b049858c1518f3"} err="failed to get container status \"8e1df3efd957fd03f1bbe3c93e84dbd268f74c04b8f566e341b049858c1518f3\": rpc error: code = NotFound desc = could not find container \"8e1df3efd957fd03f1bbe3c93e84dbd268f74c04b8f566e341b049858c1518f3\": container with ID starting with 8e1df3efd957fd03f1bbe3c93e84dbd268f74c04b8f566e341b049858c1518f3 not found: ID does not exist" Dec 06 06:38:51 crc kubenswrapper[4809]: I1206 06:38:51.411025 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53ff183f-11d1-4232-ad78-5bdd33a2c180" path="/var/lib/kubelet/pods/53ff183f-11d1-4232-ad78-5bdd33a2c180/volumes" Dec 06 06:39:04 crc kubenswrapper[4809]: I1206 06:39:04.496261 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:39:04 crc kubenswrapper[4809]: I1206 06:39:04.497046 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:39:04 crc kubenswrapper[4809]: I1206 06:39:04.497107 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-npms2" Dec 06 06:39:04 crc kubenswrapper[4809]: I1206 06:39:04.498323 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"96a97222b57fb6887f7526caa62d3152cf10b334fdc68d7a809ceb204d462e92"} pod="openshift-machine-config-operator/machine-config-daemon-npms2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 06:39:04 crc kubenswrapper[4809]: I1206 06:39:04.498430 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" containerID="cri-o://96a97222b57fb6887f7526caa62d3152cf10b334fdc68d7a809ceb204d462e92" gracePeriod=600 Dec 06 06:39:04 crc kubenswrapper[4809]: I1206 06:39:04.864198 4809 generic.go:334] "Generic (PLEG): container finished" podID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerID="96a97222b57fb6887f7526caa62d3152cf10b334fdc68d7a809ceb204d462e92" exitCode=0 Dec 06 06:39:04 crc kubenswrapper[4809]: I1206 06:39:04.864285 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" event={"ID":"cbf4cf62-024e-4703-a8b8-9aecda9cd26a","Type":"ContainerDied","Data":"96a97222b57fb6887f7526caa62d3152cf10b334fdc68d7a809ceb204d462e92"} Dec 06 06:39:04 crc kubenswrapper[4809]: I1206 06:39:04.864640 4809 scope.go:117] "RemoveContainer" containerID="3aa8c940468b111de41fd7720fd216eebfc0bb377ed933226628430658845c32" Dec 06 06:39:05 crc kubenswrapper[4809]: I1206 06:39:05.876227 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" event={"ID":"cbf4cf62-024e-4703-a8b8-9aecda9cd26a","Type":"ContainerStarted","Data":"ca23ae36692bfc6cbf4be3666821469da7cf4aff660aa8bd1e9a06863c4d9e82"} Dec 06 06:39:32 crc kubenswrapper[4809]: I1206 06:39:32.235751 4809 scope.go:117] "RemoveContainer" containerID="69787a1985dcfb0030137c138151e298fbfdc05489ac7353c4687ab3cfa91a2f" Dec 06 06:39:32 crc kubenswrapper[4809]: I1206 06:39:32.281813 4809 scope.go:117] "RemoveContainer" containerID="8c5f23df949dfd147c31ed3da5789f7c989161fffc6357ffe6a89762d95b8805" Dec 06 06:40:32 crc kubenswrapper[4809]: I1206 06:40:32.431016 4809 scope.go:117] "RemoveContainer" containerID="4317db95676b08496917b855ca265932f6ea982bcd35014ee23e80ff78ae2594" Dec 06 06:41:04 crc kubenswrapper[4809]: I1206 06:41:04.496053 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:41:04 crc kubenswrapper[4809]: I1206 06:41:04.496594 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:41:12 crc kubenswrapper[4809]: I1206 06:41:12.504418 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tn4dm"] Dec 06 06:41:12 crc kubenswrapper[4809]: E1206 06:41:12.505648 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53ff183f-11d1-4232-ad78-5bdd33a2c180" containerName="extract-utilities" Dec 06 06:41:12 crc kubenswrapper[4809]: I1206 06:41:12.505665 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="53ff183f-11d1-4232-ad78-5bdd33a2c180" containerName="extract-utilities" Dec 06 06:41:12 crc kubenswrapper[4809]: E1206 06:41:12.505681 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53ff183f-11d1-4232-ad78-5bdd33a2c180" containerName="extract-content" Dec 06 06:41:12 crc kubenswrapper[4809]: I1206 06:41:12.505689 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="53ff183f-11d1-4232-ad78-5bdd33a2c180" containerName="extract-content" Dec 06 06:41:12 crc kubenswrapper[4809]: E1206 06:41:12.505728 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53ff183f-11d1-4232-ad78-5bdd33a2c180" containerName="registry-server" Dec 06 06:41:12 crc kubenswrapper[4809]: I1206 06:41:12.505736 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="53ff183f-11d1-4232-ad78-5bdd33a2c180" containerName="registry-server" Dec 06 06:41:12 crc kubenswrapper[4809]: I1206 06:41:12.506098 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="53ff183f-11d1-4232-ad78-5bdd33a2c180" containerName="registry-server" Dec 06 06:41:12 crc kubenswrapper[4809]: I1206 06:41:12.507841 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tn4dm" Dec 06 06:41:12 crc kubenswrapper[4809]: I1206 06:41:12.516860 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tn4dm"] Dec 06 06:41:12 crc kubenswrapper[4809]: I1206 06:41:12.646759 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c799b4e-2e75-4144-87aa-09fbe9e301f4-catalog-content\") pod \"redhat-marketplace-tn4dm\" (UID: \"3c799b4e-2e75-4144-87aa-09fbe9e301f4\") " pod="openshift-marketplace/redhat-marketplace-tn4dm" Dec 06 06:41:12 crc kubenswrapper[4809]: I1206 06:41:12.647300 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cmjn\" (UniqueName: \"kubernetes.io/projected/3c799b4e-2e75-4144-87aa-09fbe9e301f4-kube-api-access-8cmjn\") pod \"redhat-marketplace-tn4dm\" (UID: \"3c799b4e-2e75-4144-87aa-09fbe9e301f4\") " pod="openshift-marketplace/redhat-marketplace-tn4dm" Dec 06 06:41:12 crc kubenswrapper[4809]: I1206 06:41:12.647582 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c799b4e-2e75-4144-87aa-09fbe9e301f4-utilities\") pod \"redhat-marketplace-tn4dm\" (UID: \"3c799b4e-2e75-4144-87aa-09fbe9e301f4\") " pod="openshift-marketplace/redhat-marketplace-tn4dm" Dec 06 06:41:12 crc kubenswrapper[4809]: I1206 06:41:12.749627 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c799b4e-2e75-4144-87aa-09fbe9e301f4-utilities\") pod \"redhat-marketplace-tn4dm\" (UID: \"3c799b4e-2e75-4144-87aa-09fbe9e301f4\") " pod="openshift-marketplace/redhat-marketplace-tn4dm" Dec 06 06:41:12 crc kubenswrapper[4809]: I1206 06:41:12.749718 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c799b4e-2e75-4144-87aa-09fbe9e301f4-catalog-content\") pod \"redhat-marketplace-tn4dm\" (UID: \"3c799b4e-2e75-4144-87aa-09fbe9e301f4\") " pod="openshift-marketplace/redhat-marketplace-tn4dm" Dec 06 06:41:12 crc kubenswrapper[4809]: I1206 06:41:12.749837 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cmjn\" (UniqueName: \"kubernetes.io/projected/3c799b4e-2e75-4144-87aa-09fbe9e301f4-kube-api-access-8cmjn\") pod \"redhat-marketplace-tn4dm\" (UID: \"3c799b4e-2e75-4144-87aa-09fbe9e301f4\") " pod="openshift-marketplace/redhat-marketplace-tn4dm" Dec 06 06:41:12 crc kubenswrapper[4809]: I1206 06:41:12.750135 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c799b4e-2e75-4144-87aa-09fbe9e301f4-utilities\") pod \"redhat-marketplace-tn4dm\" (UID: \"3c799b4e-2e75-4144-87aa-09fbe9e301f4\") " pod="openshift-marketplace/redhat-marketplace-tn4dm" Dec 06 06:41:12 crc kubenswrapper[4809]: I1206 06:41:12.750677 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c799b4e-2e75-4144-87aa-09fbe9e301f4-catalog-content\") pod \"redhat-marketplace-tn4dm\" (UID: \"3c799b4e-2e75-4144-87aa-09fbe9e301f4\") " pod="openshift-marketplace/redhat-marketplace-tn4dm" Dec 06 06:41:12 crc kubenswrapper[4809]: I1206 06:41:12.775220 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cmjn\" (UniqueName: \"kubernetes.io/projected/3c799b4e-2e75-4144-87aa-09fbe9e301f4-kube-api-access-8cmjn\") pod \"redhat-marketplace-tn4dm\" (UID: \"3c799b4e-2e75-4144-87aa-09fbe9e301f4\") " pod="openshift-marketplace/redhat-marketplace-tn4dm" Dec 06 06:41:12 crc kubenswrapper[4809]: I1206 06:41:12.829156 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tn4dm" Dec 06 06:41:13 crc kubenswrapper[4809]: I1206 06:41:13.387527 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tn4dm"] Dec 06 06:41:14 crc kubenswrapper[4809]: I1206 06:41:14.415875 4809 generic.go:334] "Generic (PLEG): container finished" podID="3c799b4e-2e75-4144-87aa-09fbe9e301f4" containerID="c88a34c1b74325cbd53403ed8b5a6f51c027a3f7b29609ac5cb74eb604b492f1" exitCode=0 Dec 06 06:41:14 crc kubenswrapper[4809]: I1206 06:41:14.416136 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tn4dm" event={"ID":"3c799b4e-2e75-4144-87aa-09fbe9e301f4","Type":"ContainerDied","Data":"c88a34c1b74325cbd53403ed8b5a6f51c027a3f7b29609ac5cb74eb604b492f1"} Dec 06 06:41:14 crc kubenswrapper[4809]: I1206 06:41:14.416362 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tn4dm" event={"ID":"3c799b4e-2e75-4144-87aa-09fbe9e301f4","Type":"ContainerStarted","Data":"4d99648b1c268f5a56b27b433e71841dc2b8b6354f1d9694864ac1ae12515d94"} Dec 06 06:41:14 crc kubenswrapper[4809]: I1206 06:41:14.417899 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 06:41:17 crc kubenswrapper[4809]: I1206 06:41:17.453296 4809 generic.go:334] "Generic (PLEG): container finished" podID="3c799b4e-2e75-4144-87aa-09fbe9e301f4" containerID="37383a3871edf91e8d56e6c06cafd62e57759baa14705ed39865499e31be4705" exitCode=0 Dec 06 06:41:17 crc kubenswrapper[4809]: I1206 06:41:17.453361 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tn4dm" event={"ID":"3c799b4e-2e75-4144-87aa-09fbe9e301f4","Type":"ContainerDied","Data":"37383a3871edf91e8d56e6c06cafd62e57759baa14705ed39865499e31be4705"} Dec 06 06:41:21 crc kubenswrapper[4809]: I1206 06:41:21.513102 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tn4dm" event={"ID":"3c799b4e-2e75-4144-87aa-09fbe9e301f4","Type":"ContainerStarted","Data":"3df09008b7388461b114efec9c42b68b6d4a98863221fcc926ba0706db13183b"} Dec 06 06:41:22 crc kubenswrapper[4809]: I1206 06:41:22.548687 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tn4dm" podStartSLOduration=4.977723176 podStartE2EDuration="10.548667523s" podCreationTimestamp="2025-12-06 06:41:12 +0000 UTC" firstStartedPulling="2025-12-06 06:41:14.417643007 +0000 UTC m=+2999.306625949" lastFinishedPulling="2025-12-06 06:41:19.988587344 +0000 UTC m=+3004.877570296" observedRunningTime="2025-12-06 06:41:22.540674766 +0000 UTC m=+3007.429657728" watchObservedRunningTime="2025-12-06 06:41:22.548667523 +0000 UTC m=+3007.437650465" Dec 06 06:41:22 crc kubenswrapper[4809]: I1206 06:41:22.830297 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tn4dm" Dec 06 06:41:22 crc kubenswrapper[4809]: I1206 06:41:22.830368 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tn4dm" Dec 06 06:41:22 crc kubenswrapper[4809]: I1206 06:41:22.886680 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tn4dm" Dec 06 06:41:32 crc kubenswrapper[4809]: I1206 06:41:32.900581 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tn4dm" Dec 06 06:41:32 crc kubenswrapper[4809]: I1206 06:41:32.973109 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tn4dm"] Dec 06 06:41:33 crc kubenswrapper[4809]: I1206 06:41:33.664524 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-tn4dm" podUID="3c799b4e-2e75-4144-87aa-09fbe9e301f4" containerName="registry-server" containerID="cri-o://3df09008b7388461b114efec9c42b68b6d4a98863221fcc926ba0706db13183b" gracePeriod=2 Dec 06 06:41:34 crc kubenswrapper[4809]: I1206 06:41:34.496152 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:41:34 crc kubenswrapper[4809]: I1206 06:41:34.496210 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:41:34 crc kubenswrapper[4809]: I1206 06:41:34.678415 4809 generic.go:334] "Generic (PLEG): container finished" podID="3c799b4e-2e75-4144-87aa-09fbe9e301f4" containerID="3df09008b7388461b114efec9c42b68b6d4a98863221fcc926ba0706db13183b" exitCode=0 Dec 06 06:41:34 crc kubenswrapper[4809]: I1206 06:41:34.678491 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tn4dm" event={"ID":"3c799b4e-2e75-4144-87aa-09fbe9e301f4","Type":"ContainerDied","Data":"3df09008b7388461b114efec9c42b68b6d4a98863221fcc926ba0706db13183b"} Dec 06 06:41:36 crc kubenswrapper[4809]: I1206 06:41:36.952053 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tn4dm" Dec 06 06:41:37 crc kubenswrapper[4809]: I1206 06:41:37.088459 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8cmjn\" (UniqueName: \"kubernetes.io/projected/3c799b4e-2e75-4144-87aa-09fbe9e301f4-kube-api-access-8cmjn\") pod \"3c799b4e-2e75-4144-87aa-09fbe9e301f4\" (UID: \"3c799b4e-2e75-4144-87aa-09fbe9e301f4\") " Dec 06 06:41:37 crc kubenswrapper[4809]: I1206 06:41:37.088585 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c799b4e-2e75-4144-87aa-09fbe9e301f4-utilities\") pod \"3c799b4e-2e75-4144-87aa-09fbe9e301f4\" (UID: \"3c799b4e-2e75-4144-87aa-09fbe9e301f4\") " Dec 06 06:41:37 crc kubenswrapper[4809]: I1206 06:41:37.088832 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c799b4e-2e75-4144-87aa-09fbe9e301f4-catalog-content\") pod \"3c799b4e-2e75-4144-87aa-09fbe9e301f4\" (UID: \"3c799b4e-2e75-4144-87aa-09fbe9e301f4\") " Dec 06 06:41:37 crc kubenswrapper[4809]: I1206 06:41:37.089543 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c799b4e-2e75-4144-87aa-09fbe9e301f4-utilities" (OuterVolumeSpecName: "utilities") pod "3c799b4e-2e75-4144-87aa-09fbe9e301f4" (UID: "3c799b4e-2e75-4144-87aa-09fbe9e301f4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:41:37 crc kubenswrapper[4809]: I1206 06:41:37.094224 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c799b4e-2e75-4144-87aa-09fbe9e301f4-kube-api-access-8cmjn" (OuterVolumeSpecName: "kube-api-access-8cmjn") pod "3c799b4e-2e75-4144-87aa-09fbe9e301f4" (UID: "3c799b4e-2e75-4144-87aa-09fbe9e301f4"). InnerVolumeSpecName "kube-api-access-8cmjn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:41:37 crc kubenswrapper[4809]: I1206 06:41:37.108473 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c799b4e-2e75-4144-87aa-09fbe9e301f4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3c799b4e-2e75-4144-87aa-09fbe9e301f4" (UID: "3c799b4e-2e75-4144-87aa-09fbe9e301f4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:41:37 crc kubenswrapper[4809]: I1206 06:41:37.192028 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c799b4e-2e75-4144-87aa-09fbe9e301f4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:41:37 crc kubenswrapper[4809]: I1206 06:41:37.192263 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8cmjn\" (UniqueName: \"kubernetes.io/projected/3c799b4e-2e75-4144-87aa-09fbe9e301f4-kube-api-access-8cmjn\") on node \"crc\" DevicePath \"\"" Dec 06 06:41:37 crc kubenswrapper[4809]: I1206 06:41:37.192375 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c799b4e-2e75-4144-87aa-09fbe9e301f4-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:41:37 crc kubenswrapper[4809]: I1206 06:41:37.732504 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tn4dm" event={"ID":"3c799b4e-2e75-4144-87aa-09fbe9e301f4","Type":"ContainerDied","Data":"4d99648b1c268f5a56b27b433e71841dc2b8b6354f1d9694864ac1ae12515d94"} Dec 06 06:41:37 crc kubenswrapper[4809]: I1206 06:41:37.732582 4809 scope.go:117] "RemoveContainer" containerID="3df09008b7388461b114efec9c42b68b6d4a98863221fcc926ba0706db13183b" Dec 06 06:41:37 crc kubenswrapper[4809]: I1206 06:41:37.732629 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tn4dm" Dec 06 06:41:37 crc kubenswrapper[4809]: I1206 06:41:37.777870 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tn4dm"] Dec 06 06:41:37 crc kubenswrapper[4809]: I1206 06:41:37.789736 4809 scope.go:117] "RemoveContainer" containerID="37383a3871edf91e8d56e6c06cafd62e57759baa14705ed39865499e31be4705" Dec 06 06:41:37 crc kubenswrapper[4809]: I1206 06:41:37.795422 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-tn4dm"] Dec 06 06:41:37 crc kubenswrapper[4809]: I1206 06:41:37.822104 4809 scope.go:117] "RemoveContainer" containerID="c88a34c1b74325cbd53403ed8b5a6f51c027a3f7b29609ac5cb74eb604b492f1" Dec 06 06:41:39 crc kubenswrapper[4809]: I1206 06:41:39.404539 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c799b4e-2e75-4144-87aa-09fbe9e301f4" path="/var/lib/kubelet/pods/3c799b4e-2e75-4144-87aa-09fbe9e301f4/volumes" Dec 06 06:42:04 crc kubenswrapper[4809]: I1206 06:42:04.496357 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:42:04 crc kubenswrapper[4809]: I1206 06:42:04.497346 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:42:04 crc kubenswrapper[4809]: I1206 06:42:04.497437 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-npms2" Dec 06 06:42:04 crc kubenswrapper[4809]: I1206 06:42:04.499062 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ca23ae36692bfc6cbf4be3666821469da7cf4aff660aa8bd1e9a06863c4d9e82"} pod="openshift-machine-config-operator/machine-config-daemon-npms2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 06:42:04 crc kubenswrapper[4809]: I1206 06:42:04.499169 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" containerID="cri-o://ca23ae36692bfc6cbf4be3666821469da7cf4aff660aa8bd1e9a06863c4d9e82" gracePeriod=600 Dec 06 06:42:05 crc kubenswrapper[4809]: I1206 06:42:05.112081 4809 generic.go:334] "Generic (PLEG): container finished" podID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerID="ca23ae36692bfc6cbf4be3666821469da7cf4aff660aa8bd1e9a06863c4d9e82" exitCode=0 Dec 06 06:42:05 crc kubenswrapper[4809]: I1206 06:42:05.112148 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" event={"ID":"cbf4cf62-024e-4703-a8b8-9aecda9cd26a","Type":"ContainerDied","Data":"ca23ae36692bfc6cbf4be3666821469da7cf4aff660aa8bd1e9a06863c4d9e82"} Dec 06 06:42:05 crc kubenswrapper[4809]: I1206 06:42:05.112462 4809 scope.go:117] "RemoveContainer" containerID="96a97222b57fb6887f7526caa62d3152cf10b334fdc68d7a809ceb204d462e92" Dec 06 06:42:06 crc kubenswrapper[4809]: E1206 06:42:06.451030 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:42:07 crc kubenswrapper[4809]: I1206 06:42:07.133515 4809 scope.go:117] "RemoveContainer" containerID="ca23ae36692bfc6cbf4be3666821469da7cf4aff660aa8bd1e9a06863c4d9e82" Dec 06 06:42:07 crc kubenswrapper[4809]: E1206 06:42:07.134151 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:42:20 crc kubenswrapper[4809]: I1206 06:42:20.389579 4809 scope.go:117] "RemoveContainer" containerID="ca23ae36692bfc6cbf4be3666821469da7cf4aff660aa8bd1e9a06863c4d9e82" Dec 06 06:42:20 crc kubenswrapper[4809]: E1206 06:42:20.390430 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:42:31 crc kubenswrapper[4809]: I1206 06:42:31.389118 4809 scope.go:117] "RemoveContainer" containerID="ca23ae36692bfc6cbf4be3666821469da7cf4aff660aa8bd1e9a06863c4d9e82" Dec 06 06:42:31 crc kubenswrapper[4809]: E1206 06:42:31.390344 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:42:38 crc kubenswrapper[4809]: I1206 06:42:38.505842 4809 generic.go:334] "Generic (PLEG): container finished" podID="c4b57a32-8e04-4fbe-864d-387273af423f" containerID="3f1ee8401752600c08a780800491544e46437e10a61712ed115a912edb66200a" exitCode=0 Dec 06 06:42:38 crc kubenswrapper[4809]: I1206 06:42:38.505981 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmgn9" event={"ID":"c4b57a32-8e04-4fbe-864d-387273af423f","Type":"ContainerDied","Data":"3f1ee8401752600c08a780800491544e46437e10a61712ed115a912edb66200a"} Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.024599 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmgn9" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.130351 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c4b57a32-8e04-4fbe-864d-387273af423f-ssh-key\") pod \"c4b57a32-8e04-4fbe-864d-387273af423f\" (UID: \"c4b57a32-8e04-4fbe-864d-387273af423f\") " Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.130561 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4b57a32-8e04-4fbe-864d-387273af423f-libvirt-combined-ca-bundle\") pod \"c4b57a32-8e04-4fbe-864d-387273af423f\" (UID: \"c4b57a32-8e04-4fbe-864d-387273af423f\") " Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.130677 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/c4b57a32-8e04-4fbe-864d-387273af423f-libvirt-secret-0\") pod \"c4b57a32-8e04-4fbe-864d-387273af423f\" (UID: \"c4b57a32-8e04-4fbe-864d-387273af423f\") " Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.130766 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6xwnx\" (UniqueName: \"kubernetes.io/projected/c4b57a32-8e04-4fbe-864d-387273af423f-kube-api-access-6xwnx\") pod \"c4b57a32-8e04-4fbe-864d-387273af423f\" (UID: \"c4b57a32-8e04-4fbe-864d-387273af423f\") " Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.131120 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c4b57a32-8e04-4fbe-864d-387273af423f-inventory\") pod \"c4b57a32-8e04-4fbe-864d-387273af423f\" (UID: \"c4b57a32-8e04-4fbe-864d-387273af423f\") " Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.136537 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4b57a32-8e04-4fbe-864d-387273af423f-kube-api-access-6xwnx" (OuterVolumeSpecName: "kube-api-access-6xwnx") pod "c4b57a32-8e04-4fbe-864d-387273af423f" (UID: "c4b57a32-8e04-4fbe-864d-387273af423f"). InnerVolumeSpecName "kube-api-access-6xwnx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.136745 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4b57a32-8e04-4fbe-864d-387273af423f-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "c4b57a32-8e04-4fbe-864d-387273af423f" (UID: "c4b57a32-8e04-4fbe-864d-387273af423f"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.231098 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4b57a32-8e04-4fbe-864d-387273af423f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c4b57a32-8e04-4fbe-864d-387273af423f" (UID: "c4b57a32-8e04-4fbe-864d-387273af423f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.237667 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6xwnx\" (UniqueName: \"kubernetes.io/projected/c4b57a32-8e04-4fbe-864d-387273af423f-kube-api-access-6xwnx\") on node \"crc\" DevicePath \"\"" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.237703 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c4b57a32-8e04-4fbe-864d-387273af423f-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.237714 4809 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4b57a32-8e04-4fbe-864d-387273af423f-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.248118 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4b57a32-8e04-4fbe-864d-387273af423f-inventory" (OuterVolumeSpecName: "inventory") pod "c4b57a32-8e04-4fbe-864d-387273af423f" (UID: "c4b57a32-8e04-4fbe-864d-387273af423f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.261194 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4b57a32-8e04-4fbe-864d-387273af423f-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "c4b57a32-8e04-4fbe-864d-387273af423f" (UID: "c4b57a32-8e04-4fbe-864d-387273af423f"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.340490 4809 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/c4b57a32-8e04-4fbe-864d-387273af423f-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.340532 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c4b57a32-8e04-4fbe-864d-387273af423f-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.529319 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmgn9" event={"ID":"c4b57a32-8e04-4fbe-864d-387273af423f","Type":"ContainerDied","Data":"367cb841eb6e329a056a41254f0f7126c1d029d32bf41987fe37ce6ce1a83f3e"} Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.529357 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="367cb841eb6e329a056a41254f0f7126c1d029d32bf41987fe37ce6ce1a83f3e" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.529413 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmgn9" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.624864 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-njt6f"] Dec 06 06:42:40 crc kubenswrapper[4809]: E1206 06:42:40.625459 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4b57a32-8e04-4fbe-864d-387273af423f" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.625487 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4b57a32-8e04-4fbe-864d-387273af423f" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 06 06:42:40 crc kubenswrapper[4809]: E1206 06:42:40.625513 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c799b4e-2e75-4144-87aa-09fbe9e301f4" containerName="extract-utilities" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.625524 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c799b4e-2e75-4144-87aa-09fbe9e301f4" containerName="extract-utilities" Dec 06 06:42:40 crc kubenswrapper[4809]: E1206 06:42:40.625550 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c799b4e-2e75-4144-87aa-09fbe9e301f4" containerName="registry-server" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.625558 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c799b4e-2e75-4144-87aa-09fbe9e301f4" containerName="registry-server" Dec 06 06:42:40 crc kubenswrapper[4809]: E1206 06:42:40.625585 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c799b4e-2e75-4144-87aa-09fbe9e301f4" containerName="extract-content" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.625592 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c799b4e-2e75-4144-87aa-09fbe9e301f4" containerName="extract-content" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.625883 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4b57a32-8e04-4fbe-864d-387273af423f" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.625905 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c799b4e-2e75-4144-87aa-09fbe9e301f4" containerName="registry-server" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.627148 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njt6f" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.629831 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.630019 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.630235 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.630344 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-bzlf5" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.630557 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.633505 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.638815 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.641868 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-njt6f"] Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.751865 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/ec2bf0dd-c16c-4912-a487-01711c8a0d84-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njt6f\" (UID: \"ec2bf0dd-c16c-4912-a487-01711c8a0d84\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njt6f" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.751974 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/ec2bf0dd-c16c-4912-a487-01711c8a0d84-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njt6f\" (UID: \"ec2bf0dd-c16c-4912-a487-01711c8a0d84\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njt6f" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.752023 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec2bf0dd-c16c-4912-a487-01711c8a0d84-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njt6f\" (UID: \"ec2bf0dd-c16c-4912-a487-01711c8a0d84\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njt6f" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.752049 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ec2bf0dd-c16c-4912-a487-01711c8a0d84-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njt6f\" (UID: \"ec2bf0dd-c16c-4912-a487-01711c8a0d84\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njt6f" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.752083 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/ec2bf0dd-c16c-4912-a487-01711c8a0d84-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njt6f\" (UID: \"ec2bf0dd-c16c-4912-a487-01711c8a0d84\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njt6f" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.752113 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/ec2bf0dd-c16c-4912-a487-01711c8a0d84-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njt6f\" (UID: \"ec2bf0dd-c16c-4912-a487-01711c8a0d84\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njt6f" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.752130 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ec2bf0dd-c16c-4912-a487-01711c8a0d84-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njt6f\" (UID: \"ec2bf0dd-c16c-4912-a487-01711c8a0d84\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njt6f" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.752163 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbsqt\" (UniqueName: \"kubernetes.io/projected/ec2bf0dd-c16c-4912-a487-01711c8a0d84-kube-api-access-hbsqt\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njt6f\" (UID: \"ec2bf0dd-c16c-4912-a487-01711c8a0d84\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njt6f" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.752259 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/ec2bf0dd-c16c-4912-a487-01711c8a0d84-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njt6f\" (UID: \"ec2bf0dd-c16c-4912-a487-01711c8a0d84\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njt6f" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.854162 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/ec2bf0dd-c16c-4912-a487-01711c8a0d84-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njt6f\" (UID: \"ec2bf0dd-c16c-4912-a487-01711c8a0d84\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njt6f" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.854591 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/ec2bf0dd-c16c-4912-a487-01711c8a0d84-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njt6f\" (UID: \"ec2bf0dd-c16c-4912-a487-01711c8a0d84\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njt6f" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.854718 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec2bf0dd-c16c-4912-a487-01711c8a0d84-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njt6f\" (UID: \"ec2bf0dd-c16c-4912-a487-01711c8a0d84\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njt6f" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.854798 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ec2bf0dd-c16c-4912-a487-01711c8a0d84-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njt6f\" (UID: \"ec2bf0dd-c16c-4912-a487-01711c8a0d84\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njt6f" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.854900 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/ec2bf0dd-c16c-4912-a487-01711c8a0d84-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njt6f\" (UID: \"ec2bf0dd-c16c-4912-a487-01711c8a0d84\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njt6f" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.855003 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/ec2bf0dd-c16c-4912-a487-01711c8a0d84-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njt6f\" (UID: \"ec2bf0dd-c16c-4912-a487-01711c8a0d84\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njt6f" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.855126 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ec2bf0dd-c16c-4912-a487-01711c8a0d84-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njt6f\" (UID: \"ec2bf0dd-c16c-4912-a487-01711c8a0d84\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njt6f" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.855228 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbsqt\" (UniqueName: \"kubernetes.io/projected/ec2bf0dd-c16c-4912-a487-01711c8a0d84-kube-api-access-hbsqt\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njt6f\" (UID: \"ec2bf0dd-c16c-4912-a487-01711c8a0d84\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njt6f" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.855426 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/ec2bf0dd-c16c-4912-a487-01711c8a0d84-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njt6f\" (UID: \"ec2bf0dd-c16c-4912-a487-01711c8a0d84\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njt6f" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.856392 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/ec2bf0dd-c16c-4912-a487-01711c8a0d84-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njt6f\" (UID: \"ec2bf0dd-c16c-4912-a487-01711c8a0d84\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njt6f" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.858907 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ec2bf0dd-c16c-4912-a487-01711c8a0d84-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njt6f\" (UID: \"ec2bf0dd-c16c-4912-a487-01711c8a0d84\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njt6f" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.859045 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/ec2bf0dd-c16c-4912-a487-01711c8a0d84-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njt6f\" (UID: \"ec2bf0dd-c16c-4912-a487-01711c8a0d84\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njt6f" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.859401 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/ec2bf0dd-c16c-4912-a487-01711c8a0d84-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njt6f\" (UID: \"ec2bf0dd-c16c-4912-a487-01711c8a0d84\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njt6f" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.860480 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ec2bf0dd-c16c-4912-a487-01711c8a0d84-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njt6f\" (UID: \"ec2bf0dd-c16c-4912-a487-01711c8a0d84\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njt6f" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.860667 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/ec2bf0dd-c16c-4912-a487-01711c8a0d84-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njt6f\" (UID: \"ec2bf0dd-c16c-4912-a487-01711c8a0d84\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njt6f" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.861148 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/ec2bf0dd-c16c-4912-a487-01711c8a0d84-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njt6f\" (UID: \"ec2bf0dd-c16c-4912-a487-01711c8a0d84\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njt6f" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.861738 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec2bf0dd-c16c-4912-a487-01711c8a0d84-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njt6f\" (UID: \"ec2bf0dd-c16c-4912-a487-01711c8a0d84\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njt6f" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.870340 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbsqt\" (UniqueName: \"kubernetes.io/projected/ec2bf0dd-c16c-4912-a487-01711c8a0d84-kube-api-access-hbsqt\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njt6f\" (UID: \"ec2bf0dd-c16c-4912-a487-01711c8a0d84\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njt6f" Dec 06 06:42:40 crc kubenswrapper[4809]: I1206 06:42:40.962860 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njt6f" Dec 06 06:42:41 crc kubenswrapper[4809]: I1206 06:42:41.501303 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-njt6f"] Dec 06 06:42:41 crc kubenswrapper[4809]: I1206 06:42:41.539118 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njt6f" event={"ID":"ec2bf0dd-c16c-4912-a487-01711c8a0d84","Type":"ContainerStarted","Data":"ffeafa35fd78205f77e78bb6e6c083936fed0e4b738ded903487b4601771b1e8"} Dec 06 06:42:42 crc kubenswrapper[4809]: I1206 06:42:42.550687 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njt6f" event={"ID":"ec2bf0dd-c16c-4912-a487-01711c8a0d84","Type":"ContainerStarted","Data":"af6f9c6f55fbe2016111be1403c9b1dca947455e5deb54d47f0440ec5ac1252a"} Dec 06 06:42:42 crc kubenswrapper[4809]: I1206 06:42:42.575247 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njt6f" podStartSLOduration=2.033828516 podStartE2EDuration="2.575223299s" podCreationTimestamp="2025-12-06 06:42:40 +0000 UTC" firstStartedPulling="2025-12-06 06:42:41.499908793 +0000 UTC m=+3086.388891755" lastFinishedPulling="2025-12-06 06:42:42.041303596 +0000 UTC m=+3086.930286538" observedRunningTime="2025-12-06 06:42:42.568001753 +0000 UTC m=+3087.456984705" watchObservedRunningTime="2025-12-06 06:42:42.575223299 +0000 UTC m=+3087.464206261" Dec 06 06:42:46 crc kubenswrapper[4809]: I1206 06:42:46.388920 4809 scope.go:117] "RemoveContainer" containerID="ca23ae36692bfc6cbf4be3666821469da7cf4aff660aa8bd1e9a06863c4d9e82" Dec 06 06:42:46 crc kubenswrapper[4809]: E1206 06:42:46.389926 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:43:00 crc kubenswrapper[4809]: I1206 06:43:00.389796 4809 scope.go:117] "RemoveContainer" containerID="ca23ae36692bfc6cbf4be3666821469da7cf4aff660aa8bd1e9a06863c4d9e82" Dec 06 06:43:00 crc kubenswrapper[4809]: E1206 06:43:00.391161 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:43:13 crc kubenswrapper[4809]: I1206 06:43:13.388348 4809 scope.go:117] "RemoveContainer" containerID="ca23ae36692bfc6cbf4be3666821469da7cf4aff660aa8bd1e9a06863c4d9e82" Dec 06 06:43:13 crc kubenswrapper[4809]: E1206 06:43:13.389128 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:43:25 crc kubenswrapper[4809]: I1206 06:43:25.417730 4809 scope.go:117] "RemoveContainer" containerID="ca23ae36692bfc6cbf4be3666821469da7cf4aff660aa8bd1e9a06863c4d9e82" Dec 06 06:43:25 crc kubenswrapper[4809]: E1206 06:43:25.419259 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:43:36 crc kubenswrapper[4809]: I1206 06:43:36.388722 4809 scope.go:117] "RemoveContainer" containerID="ca23ae36692bfc6cbf4be3666821469da7cf4aff660aa8bd1e9a06863c4d9e82" Dec 06 06:43:36 crc kubenswrapper[4809]: E1206 06:43:36.390372 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:43:47 crc kubenswrapper[4809]: I1206 06:43:47.389109 4809 scope.go:117] "RemoveContainer" containerID="ca23ae36692bfc6cbf4be3666821469da7cf4aff660aa8bd1e9a06863c4d9e82" Dec 06 06:43:47 crc kubenswrapper[4809]: E1206 06:43:47.390029 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:44:01 crc kubenswrapper[4809]: I1206 06:44:01.389247 4809 scope.go:117] "RemoveContainer" containerID="ca23ae36692bfc6cbf4be3666821469da7cf4aff660aa8bd1e9a06863c4d9e82" Dec 06 06:44:01 crc kubenswrapper[4809]: E1206 06:44:01.390392 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:44:09 crc kubenswrapper[4809]: I1206 06:44:09.151650 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5kgd4"] Dec 06 06:44:09 crc kubenswrapper[4809]: I1206 06:44:09.154737 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5kgd4" Dec 06 06:44:09 crc kubenswrapper[4809]: I1206 06:44:09.171153 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5kgd4"] Dec 06 06:44:09 crc kubenswrapper[4809]: I1206 06:44:09.293067 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jt5fx\" (UniqueName: \"kubernetes.io/projected/a0239592-0787-44e3-823d-53101aa589df-kube-api-access-jt5fx\") pod \"redhat-operators-5kgd4\" (UID: \"a0239592-0787-44e3-823d-53101aa589df\") " pod="openshift-marketplace/redhat-operators-5kgd4" Dec 06 06:44:09 crc kubenswrapper[4809]: I1206 06:44:09.293710 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0239592-0787-44e3-823d-53101aa589df-catalog-content\") pod \"redhat-operators-5kgd4\" (UID: \"a0239592-0787-44e3-823d-53101aa589df\") " pod="openshift-marketplace/redhat-operators-5kgd4" Dec 06 06:44:09 crc kubenswrapper[4809]: I1206 06:44:09.293964 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0239592-0787-44e3-823d-53101aa589df-utilities\") pod \"redhat-operators-5kgd4\" (UID: \"a0239592-0787-44e3-823d-53101aa589df\") " pod="openshift-marketplace/redhat-operators-5kgd4" Dec 06 06:44:09 crc kubenswrapper[4809]: I1206 06:44:09.396007 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0239592-0787-44e3-823d-53101aa589df-catalog-content\") pod \"redhat-operators-5kgd4\" (UID: \"a0239592-0787-44e3-823d-53101aa589df\") " pod="openshift-marketplace/redhat-operators-5kgd4" Dec 06 06:44:09 crc kubenswrapper[4809]: I1206 06:44:09.396096 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0239592-0787-44e3-823d-53101aa589df-utilities\") pod \"redhat-operators-5kgd4\" (UID: \"a0239592-0787-44e3-823d-53101aa589df\") " pod="openshift-marketplace/redhat-operators-5kgd4" Dec 06 06:44:09 crc kubenswrapper[4809]: I1206 06:44:09.396164 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jt5fx\" (UniqueName: \"kubernetes.io/projected/a0239592-0787-44e3-823d-53101aa589df-kube-api-access-jt5fx\") pod \"redhat-operators-5kgd4\" (UID: \"a0239592-0787-44e3-823d-53101aa589df\") " pod="openshift-marketplace/redhat-operators-5kgd4" Dec 06 06:44:09 crc kubenswrapper[4809]: I1206 06:44:09.396627 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0239592-0787-44e3-823d-53101aa589df-catalog-content\") pod \"redhat-operators-5kgd4\" (UID: \"a0239592-0787-44e3-823d-53101aa589df\") " pod="openshift-marketplace/redhat-operators-5kgd4" Dec 06 06:44:09 crc kubenswrapper[4809]: I1206 06:44:09.398268 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0239592-0787-44e3-823d-53101aa589df-utilities\") pod \"redhat-operators-5kgd4\" (UID: \"a0239592-0787-44e3-823d-53101aa589df\") " pod="openshift-marketplace/redhat-operators-5kgd4" Dec 06 06:44:09 crc kubenswrapper[4809]: I1206 06:44:09.428755 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jt5fx\" (UniqueName: \"kubernetes.io/projected/a0239592-0787-44e3-823d-53101aa589df-kube-api-access-jt5fx\") pod \"redhat-operators-5kgd4\" (UID: \"a0239592-0787-44e3-823d-53101aa589df\") " pod="openshift-marketplace/redhat-operators-5kgd4" Dec 06 06:44:09 crc kubenswrapper[4809]: I1206 06:44:09.498681 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5kgd4" Dec 06 06:44:10 crc kubenswrapper[4809]: I1206 06:44:10.017020 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5kgd4"] Dec 06 06:44:10 crc kubenswrapper[4809]: I1206 06:44:10.585015 4809 generic.go:334] "Generic (PLEG): container finished" podID="a0239592-0787-44e3-823d-53101aa589df" containerID="d4311b3d67cfc79eb4d58ab21e01750238b2e65d997ded719f775ab7b3c7509f" exitCode=0 Dec 06 06:44:10 crc kubenswrapper[4809]: I1206 06:44:10.586340 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5kgd4" event={"ID":"a0239592-0787-44e3-823d-53101aa589df","Type":"ContainerDied","Data":"d4311b3d67cfc79eb4d58ab21e01750238b2e65d997ded719f775ab7b3c7509f"} Dec 06 06:44:10 crc kubenswrapper[4809]: I1206 06:44:10.586473 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5kgd4" event={"ID":"a0239592-0787-44e3-823d-53101aa589df","Type":"ContainerStarted","Data":"e99eb98505786049bbc1012933628492d8bd7a960f4e7c3af1ee450eec697840"} Dec 06 06:44:12 crc kubenswrapper[4809]: I1206 06:44:12.618313 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5kgd4" event={"ID":"a0239592-0787-44e3-823d-53101aa589df","Type":"ContainerStarted","Data":"f485d604b1baa4d876ec6396eaf010f8360d43ab96e36af8902d5fd3bb4bf7d9"} Dec 06 06:44:13 crc kubenswrapper[4809]: I1206 06:44:13.389321 4809 scope.go:117] "RemoveContainer" containerID="ca23ae36692bfc6cbf4be3666821469da7cf4aff660aa8bd1e9a06863c4d9e82" Dec 06 06:44:13 crc kubenswrapper[4809]: E1206 06:44:13.389911 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:44:13 crc kubenswrapper[4809]: I1206 06:44:13.636139 4809 generic.go:334] "Generic (PLEG): container finished" podID="a0239592-0787-44e3-823d-53101aa589df" containerID="f485d604b1baa4d876ec6396eaf010f8360d43ab96e36af8902d5fd3bb4bf7d9" exitCode=0 Dec 06 06:44:13 crc kubenswrapper[4809]: I1206 06:44:13.636233 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5kgd4" event={"ID":"a0239592-0787-44e3-823d-53101aa589df","Type":"ContainerDied","Data":"f485d604b1baa4d876ec6396eaf010f8360d43ab96e36af8902d5fd3bb4bf7d9"} Dec 06 06:44:17 crc kubenswrapper[4809]: I1206 06:44:17.682312 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5kgd4" event={"ID":"a0239592-0787-44e3-823d-53101aa589df","Type":"ContainerStarted","Data":"e137ec7add300f3bb117dc43fbf30efd8c3cd0da1891a5963997d40da66f366b"} Dec 06 06:44:17 crc kubenswrapper[4809]: I1206 06:44:17.715541 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5kgd4" podStartSLOduration=1.9039027800000001 podStartE2EDuration="8.715516478s" podCreationTimestamp="2025-12-06 06:44:09 +0000 UTC" firstStartedPulling="2025-12-06 06:44:10.589084588 +0000 UTC m=+3175.478067540" lastFinishedPulling="2025-12-06 06:44:17.400698296 +0000 UTC m=+3182.289681238" observedRunningTime="2025-12-06 06:44:17.704350576 +0000 UTC m=+3182.593333518" watchObservedRunningTime="2025-12-06 06:44:17.715516478 +0000 UTC m=+3182.604499420" Dec 06 06:44:19 crc kubenswrapper[4809]: I1206 06:44:19.499838 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5kgd4" Dec 06 06:44:19 crc kubenswrapper[4809]: I1206 06:44:19.500211 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5kgd4" Dec 06 06:44:20 crc kubenswrapper[4809]: I1206 06:44:20.550075 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5kgd4" podUID="a0239592-0787-44e3-823d-53101aa589df" containerName="registry-server" probeResult="failure" output=< Dec 06 06:44:20 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Dec 06 06:44:20 crc kubenswrapper[4809]: > Dec 06 06:44:25 crc kubenswrapper[4809]: I1206 06:44:25.400275 4809 scope.go:117] "RemoveContainer" containerID="ca23ae36692bfc6cbf4be3666821469da7cf4aff660aa8bd1e9a06863c4d9e82" Dec 06 06:44:25 crc kubenswrapper[4809]: E1206 06:44:25.401194 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:44:29 crc kubenswrapper[4809]: I1206 06:44:29.556115 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5kgd4" Dec 06 06:44:29 crc kubenswrapper[4809]: I1206 06:44:29.610838 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5kgd4" Dec 06 06:44:29 crc kubenswrapper[4809]: I1206 06:44:29.804601 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5kgd4"] Dec 06 06:44:30 crc kubenswrapper[4809]: I1206 06:44:30.829469 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5kgd4" podUID="a0239592-0787-44e3-823d-53101aa589df" containerName="registry-server" containerID="cri-o://e137ec7add300f3bb117dc43fbf30efd8c3cd0da1891a5963997d40da66f366b" gracePeriod=2 Dec 06 06:44:31 crc kubenswrapper[4809]: I1206 06:44:31.844857 4809 generic.go:334] "Generic (PLEG): container finished" podID="a0239592-0787-44e3-823d-53101aa589df" containerID="e137ec7add300f3bb117dc43fbf30efd8c3cd0da1891a5963997d40da66f366b" exitCode=0 Dec 06 06:44:31 crc kubenswrapper[4809]: I1206 06:44:31.845174 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5kgd4" event={"ID":"a0239592-0787-44e3-823d-53101aa589df","Type":"ContainerDied","Data":"e137ec7add300f3bb117dc43fbf30efd8c3cd0da1891a5963997d40da66f366b"} Dec 06 06:44:31 crc kubenswrapper[4809]: I1206 06:44:31.845201 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5kgd4" event={"ID":"a0239592-0787-44e3-823d-53101aa589df","Type":"ContainerDied","Data":"e99eb98505786049bbc1012933628492d8bd7a960f4e7c3af1ee450eec697840"} Dec 06 06:44:31 crc kubenswrapper[4809]: I1206 06:44:31.845212 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e99eb98505786049bbc1012933628492d8bd7a960f4e7c3af1ee450eec697840" Dec 06 06:44:31 crc kubenswrapper[4809]: I1206 06:44:31.946563 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5kgd4" Dec 06 06:44:31 crc kubenswrapper[4809]: I1206 06:44:31.991521 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0239592-0787-44e3-823d-53101aa589df-utilities\") pod \"a0239592-0787-44e3-823d-53101aa589df\" (UID: \"a0239592-0787-44e3-823d-53101aa589df\") " Dec 06 06:44:31 crc kubenswrapper[4809]: I1206 06:44:31.991907 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0239592-0787-44e3-823d-53101aa589df-catalog-content\") pod \"a0239592-0787-44e3-823d-53101aa589df\" (UID: \"a0239592-0787-44e3-823d-53101aa589df\") " Dec 06 06:44:31 crc kubenswrapper[4809]: I1206 06:44:31.992160 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jt5fx\" (UniqueName: \"kubernetes.io/projected/a0239592-0787-44e3-823d-53101aa589df-kube-api-access-jt5fx\") pod \"a0239592-0787-44e3-823d-53101aa589df\" (UID: \"a0239592-0787-44e3-823d-53101aa589df\") " Dec 06 06:44:31 crc kubenswrapper[4809]: I1206 06:44:31.992541 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0239592-0787-44e3-823d-53101aa589df-utilities" (OuterVolumeSpecName: "utilities") pod "a0239592-0787-44e3-823d-53101aa589df" (UID: "a0239592-0787-44e3-823d-53101aa589df"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:44:31 crc kubenswrapper[4809]: I1206 06:44:31.993312 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0239592-0787-44e3-823d-53101aa589df-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:44:32 crc kubenswrapper[4809]: I1206 06:44:32.000689 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0239592-0787-44e3-823d-53101aa589df-kube-api-access-jt5fx" (OuterVolumeSpecName: "kube-api-access-jt5fx") pod "a0239592-0787-44e3-823d-53101aa589df" (UID: "a0239592-0787-44e3-823d-53101aa589df"). InnerVolumeSpecName "kube-api-access-jt5fx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:44:32 crc kubenswrapper[4809]: I1206 06:44:32.095875 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jt5fx\" (UniqueName: \"kubernetes.io/projected/a0239592-0787-44e3-823d-53101aa589df-kube-api-access-jt5fx\") on node \"crc\" DevicePath \"\"" Dec 06 06:44:32 crc kubenswrapper[4809]: I1206 06:44:32.108796 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0239592-0787-44e3-823d-53101aa589df-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a0239592-0787-44e3-823d-53101aa589df" (UID: "a0239592-0787-44e3-823d-53101aa589df"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:44:32 crc kubenswrapper[4809]: I1206 06:44:32.198504 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0239592-0787-44e3-823d-53101aa589df-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:44:32 crc kubenswrapper[4809]: I1206 06:44:32.855215 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5kgd4" Dec 06 06:44:32 crc kubenswrapper[4809]: I1206 06:44:32.890314 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5kgd4"] Dec 06 06:44:32 crc kubenswrapper[4809]: I1206 06:44:32.901188 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5kgd4"] Dec 06 06:44:33 crc kubenswrapper[4809]: I1206 06:44:33.434368 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0239592-0787-44e3-823d-53101aa589df" path="/var/lib/kubelet/pods/a0239592-0787-44e3-823d-53101aa589df/volumes" Dec 06 06:44:38 crc kubenswrapper[4809]: I1206 06:44:38.388733 4809 scope.go:117] "RemoveContainer" containerID="ca23ae36692bfc6cbf4be3666821469da7cf4aff660aa8bd1e9a06863c4d9e82" Dec 06 06:44:38 crc kubenswrapper[4809]: E1206 06:44:38.389716 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:44:53 crc kubenswrapper[4809]: I1206 06:44:53.389492 4809 scope.go:117] "RemoveContainer" containerID="ca23ae36692bfc6cbf4be3666821469da7cf4aff660aa8bd1e9a06863c4d9e82" Dec 06 06:44:53 crc kubenswrapper[4809]: E1206 06:44:53.390573 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:45:00 crc kubenswrapper[4809]: I1206 06:45:00.169728 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416725-kq2pl"] Dec 06 06:45:00 crc kubenswrapper[4809]: E1206 06:45:00.170875 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0239592-0787-44e3-823d-53101aa589df" containerName="extract-utilities" Dec 06 06:45:00 crc kubenswrapper[4809]: I1206 06:45:00.170890 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0239592-0787-44e3-823d-53101aa589df" containerName="extract-utilities" Dec 06 06:45:00 crc kubenswrapper[4809]: E1206 06:45:00.170958 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0239592-0787-44e3-823d-53101aa589df" containerName="registry-server" Dec 06 06:45:00 crc kubenswrapper[4809]: I1206 06:45:00.170965 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0239592-0787-44e3-823d-53101aa589df" containerName="registry-server" Dec 06 06:45:00 crc kubenswrapper[4809]: E1206 06:45:00.170984 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0239592-0787-44e3-823d-53101aa589df" containerName="extract-content" Dec 06 06:45:00 crc kubenswrapper[4809]: I1206 06:45:00.170990 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0239592-0787-44e3-823d-53101aa589df" containerName="extract-content" Dec 06 06:45:00 crc kubenswrapper[4809]: I1206 06:45:00.171205 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0239592-0787-44e3-823d-53101aa589df" containerName="registry-server" Dec 06 06:45:00 crc kubenswrapper[4809]: I1206 06:45:00.176178 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-kq2pl" Dec 06 06:45:00 crc kubenswrapper[4809]: I1206 06:45:00.180211 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 06:45:00 crc kubenswrapper[4809]: I1206 06:45:00.185116 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 06:45:00 crc kubenswrapper[4809]: I1206 06:45:00.186647 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416725-kq2pl"] Dec 06 06:45:00 crc kubenswrapper[4809]: I1206 06:45:00.253770 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9bf15846-fb7f-45fc-9146-02092049e864-config-volume\") pod \"collect-profiles-29416725-kq2pl\" (UID: \"9bf15846-fb7f-45fc-9146-02092049e864\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-kq2pl" Dec 06 06:45:00 crc kubenswrapper[4809]: I1206 06:45:00.254094 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pp7w\" (UniqueName: \"kubernetes.io/projected/9bf15846-fb7f-45fc-9146-02092049e864-kube-api-access-7pp7w\") pod \"collect-profiles-29416725-kq2pl\" (UID: \"9bf15846-fb7f-45fc-9146-02092049e864\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-kq2pl" Dec 06 06:45:00 crc kubenswrapper[4809]: I1206 06:45:00.254279 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9bf15846-fb7f-45fc-9146-02092049e864-secret-volume\") pod \"collect-profiles-29416725-kq2pl\" (UID: \"9bf15846-fb7f-45fc-9146-02092049e864\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-kq2pl" Dec 06 06:45:00 crc kubenswrapper[4809]: I1206 06:45:00.356459 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9bf15846-fb7f-45fc-9146-02092049e864-config-volume\") pod \"collect-profiles-29416725-kq2pl\" (UID: \"9bf15846-fb7f-45fc-9146-02092049e864\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-kq2pl" Dec 06 06:45:00 crc kubenswrapper[4809]: I1206 06:45:00.356515 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pp7w\" (UniqueName: \"kubernetes.io/projected/9bf15846-fb7f-45fc-9146-02092049e864-kube-api-access-7pp7w\") pod \"collect-profiles-29416725-kq2pl\" (UID: \"9bf15846-fb7f-45fc-9146-02092049e864\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-kq2pl" Dec 06 06:45:00 crc kubenswrapper[4809]: I1206 06:45:00.356590 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9bf15846-fb7f-45fc-9146-02092049e864-secret-volume\") pod \"collect-profiles-29416725-kq2pl\" (UID: \"9bf15846-fb7f-45fc-9146-02092049e864\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-kq2pl" Dec 06 06:45:00 crc kubenswrapper[4809]: I1206 06:45:00.358168 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9bf15846-fb7f-45fc-9146-02092049e864-config-volume\") pod \"collect-profiles-29416725-kq2pl\" (UID: \"9bf15846-fb7f-45fc-9146-02092049e864\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-kq2pl" Dec 06 06:45:00 crc kubenswrapper[4809]: I1206 06:45:00.369271 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9bf15846-fb7f-45fc-9146-02092049e864-secret-volume\") pod \"collect-profiles-29416725-kq2pl\" (UID: \"9bf15846-fb7f-45fc-9146-02092049e864\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-kq2pl" Dec 06 06:45:00 crc kubenswrapper[4809]: I1206 06:45:00.378160 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pp7w\" (UniqueName: \"kubernetes.io/projected/9bf15846-fb7f-45fc-9146-02092049e864-kube-api-access-7pp7w\") pod \"collect-profiles-29416725-kq2pl\" (UID: \"9bf15846-fb7f-45fc-9146-02092049e864\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-kq2pl" Dec 06 06:45:00 crc kubenswrapper[4809]: I1206 06:45:00.509019 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-kq2pl" Dec 06 06:45:01 crc kubenswrapper[4809]: I1206 06:45:01.014478 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416725-kq2pl"] Dec 06 06:45:01 crc kubenswrapper[4809]: I1206 06:45:01.211822 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-kq2pl" event={"ID":"9bf15846-fb7f-45fc-9146-02092049e864","Type":"ContainerStarted","Data":"24063d21fc3de5c944c7d9087664622b245d18e8d0a66964fa6a88d8ec87ff70"} Dec 06 06:45:02 crc kubenswrapper[4809]: I1206 06:45:02.226484 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-kq2pl" event={"ID":"9bf15846-fb7f-45fc-9146-02092049e864","Type":"ContainerStarted","Data":"154f2d114d6b997794c52348b060d97bce9656c84765b90e42eee6e301a8d492"} Dec 06 06:45:02 crc kubenswrapper[4809]: I1206 06:45:02.248255 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-kq2pl" podStartSLOduration=2.248226814 podStartE2EDuration="2.248226814s" podCreationTimestamp="2025-12-06 06:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:45:02.242225962 +0000 UTC m=+3227.131208914" watchObservedRunningTime="2025-12-06 06:45:02.248226814 +0000 UTC m=+3227.137209756" Dec 06 06:45:03 crc kubenswrapper[4809]: I1206 06:45:03.243375 4809 generic.go:334] "Generic (PLEG): container finished" podID="9bf15846-fb7f-45fc-9146-02092049e864" containerID="154f2d114d6b997794c52348b060d97bce9656c84765b90e42eee6e301a8d492" exitCode=0 Dec 06 06:45:03 crc kubenswrapper[4809]: I1206 06:45:03.243491 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-kq2pl" event={"ID":"9bf15846-fb7f-45fc-9146-02092049e864","Type":"ContainerDied","Data":"154f2d114d6b997794c52348b060d97bce9656c84765b90e42eee6e301a8d492"} Dec 06 06:45:04 crc kubenswrapper[4809]: I1206 06:45:04.388663 4809 scope.go:117] "RemoveContainer" containerID="ca23ae36692bfc6cbf4be3666821469da7cf4aff660aa8bd1e9a06863c4d9e82" Dec 06 06:45:04 crc kubenswrapper[4809]: E1206 06:45:04.389081 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:45:04 crc kubenswrapper[4809]: I1206 06:45:04.654745 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-kq2pl" Dec 06 06:45:04 crc kubenswrapper[4809]: I1206 06:45:04.776572 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9bf15846-fb7f-45fc-9146-02092049e864-config-volume\") pod \"9bf15846-fb7f-45fc-9146-02092049e864\" (UID: \"9bf15846-fb7f-45fc-9146-02092049e864\") " Dec 06 06:45:04 crc kubenswrapper[4809]: I1206 06:45:04.777035 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7pp7w\" (UniqueName: \"kubernetes.io/projected/9bf15846-fb7f-45fc-9146-02092049e864-kube-api-access-7pp7w\") pod \"9bf15846-fb7f-45fc-9146-02092049e864\" (UID: \"9bf15846-fb7f-45fc-9146-02092049e864\") " Dec 06 06:45:04 crc kubenswrapper[4809]: I1206 06:45:04.777288 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9bf15846-fb7f-45fc-9146-02092049e864-secret-volume\") pod \"9bf15846-fb7f-45fc-9146-02092049e864\" (UID: \"9bf15846-fb7f-45fc-9146-02092049e864\") " Dec 06 06:45:04 crc kubenswrapper[4809]: I1206 06:45:04.777436 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9bf15846-fb7f-45fc-9146-02092049e864-config-volume" (OuterVolumeSpecName: "config-volume") pod "9bf15846-fb7f-45fc-9146-02092049e864" (UID: "9bf15846-fb7f-45fc-9146-02092049e864"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:45:04 crc kubenswrapper[4809]: I1206 06:45:04.778285 4809 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9bf15846-fb7f-45fc-9146-02092049e864-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 06:45:04 crc kubenswrapper[4809]: I1206 06:45:04.783196 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9bf15846-fb7f-45fc-9146-02092049e864-kube-api-access-7pp7w" (OuterVolumeSpecName: "kube-api-access-7pp7w") pod "9bf15846-fb7f-45fc-9146-02092049e864" (UID: "9bf15846-fb7f-45fc-9146-02092049e864"). InnerVolumeSpecName "kube-api-access-7pp7w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:45:04 crc kubenswrapper[4809]: I1206 06:45:04.789397 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9bf15846-fb7f-45fc-9146-02092049e864-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "9bf15846-fb7f-45fc-9146-02092049e864" (UID: "9bf15846-fb7f-45fc-9146-02092049e864"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:45:04 crc kubenswrapper[4809]: I1206 06:45:04.880620 4809 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9bf15846-fb7f-45fc-9146-02092049e864-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 06:45:04 crc kubenswrapper[4809]: I1206 06:45:04.880668 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7pp7w\" (UniqueName: \"kubernetes.io/projected/9bf15846-fb7f-45fc-9146-02092049e864-kube-api-access-7pp7w\") on node \"crc\" DevicePath \"\"" Dec 06 06:45:05 crc kubenswrapper[4809]: I1206 06:45:05.268008 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-kq2pl" event={"ID":"9bf15846-fb7f-45fc-9146-02092049e864","Type":"ContainerDied","Data":"24063d21fc3de5c944c7d9087664622b245d18e8d0a66964fa6a88d8ec87ff70"} Dec 06 06:45:05 crc kubenswrapper[4809]: I1206 06:45:05.268061 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="24063d21fc3de5c944c7d9087664622b245d18e8d0a66964fa6a88d8ec87ff70" Dec 06 06:45:05 crc kubenswrapper[4809]: I1206 06:45:05.268078 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-kq2pl" Dec 06 06:45:05 crc kubenswrapper[4809]: I1206 06:45:05.332364 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416680-97j2f"] Dec 06 06:45:05 crc kubenswrapper[4809]: I1206 06:45:05.345369 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416680-97j2f"] Dec 06 06:45:05 crc kubenswrapper[4809]: I1206 06:45:05.406665 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c52c53b2-f8ba-4272-863b-f8ba48c1b3cc" path="/var/lib/kubelet/pods/c52c53b2-f8ba-4272-863b-f8ba48c1b3cc/volumes" Dec 06 06:45:18 crc kubenswrapper[4809]: I1206 06:45:18.389156 4809 scope.go:117] "RemoveContainer" containerID="ca23ae36692bfc6cbf4be3666821469da7cf4aff660aa8bd1e9a06863c4d9e82" Dec 06 06:45:18 crc kubenswrapper[4809]: E1206 06:45:18.390070 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:45:32 crc kubenswrapper[4809]: I1206 06:45:32.598517 4809 scope.go:117] "RemoveContainer" containerID="34e04c3ffeae84899ea75dfbf59cbd3770a5c9f7104321b0bf1f6a4bad7da62f" Dec 06 06:45:33 crc kubenswrapper[4809]: I1206 06:45:33.389201 4809 scope.go:117] "RemoveContainer" containerID="ca23ae36692bfc6cbf4be3666821469da7cf4aff660aa8bd1e9a06863c4d9e82" Dec 06 06:45:33 crc kubenswrapper[4809]: E1206 06:45:33.389658 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:45:33 crc kubenswrapper[4809]: I1206 06:45:33.587086 4809 generic.go:334] "Generic (PLEG): container finished" podID="ec2bf0dd-c16c-4912-a487-01711c8a0d84" containerID="af6f9c6f55fbe2016111be1403c9b1dca947455e5deb54d47f0440ec5ac1252a" exitCode=0 Dec 06 06:45:33 crc kubenswrapper[4809]: I1206 06:45:33.587156 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njt6f" event={"ID":"ec2bf0dd-c16c-4912-a487-01711c8a0d84","Type":"ContainerDied","Data":"af6f9c6f55fbe2016111be1403c9b1dca947455e5deb54d47f0440ec5ac1252a"} Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.079369 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njt6f" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.200792 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/ec2bf0dd-c16c-4912-a487-01711c8a0d84-nova-migration-ssh-key-1\") pod \"ec2bf0dd-c16c-4912-a487-01711c8a0d84\" (UID: \"ec2bf0dd-c16c-4912-a487-01711c8a0d84\") " Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.200871 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ec2bf0dd-c16c-4912-a487-01711c8a0d84-ssh-key\") pod \"ec2bf0dd-c16c-4912-a487-01711c8a0d84\" (UID: \"ec2bf0dd-c16c-4912-a487-01711c8a0d84\") " Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.200910 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ec2bf0dd-c16c-4912-a487-01711c8a0d84-inventory\") pod \"ec2bf0dd-c16c-4912-a487-01711c8a0d84\" (UID: \"ec2bf0dd-c16c-4912-a487-01711c8a0d84\") " Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.201049 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbsqt\" (UniqueName: \"kubernetes.io/projected/ec2bf0dd-c16c-4912-a487-01711c8a0d84-kube-api-access-hbsqt\") pod \"ec2bf0dd-c16c-4912-a487-01711c8a0d84\" (UID: \"ec2bf0dd-c16c-4912-a487-01711c8a0d84\") " Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.201086 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/ec2bf0dd-c16c-4912-a487-01711c8a0d84-nova-cell1-compute-config-1\") pod \"ec2bf0dd-c16c-4912-a487-01711c8a0d84\" (UID: \"ec2bf0dd-c16c-4912-a487-01711c8a0d84\") " Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.201129 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/ec2bf0dd-c16c-4912-a487-01711c8a0d84-nova-migration-ssh-key-0\") pod \"ec2bf0dd-c16c-4912-a487-01711c8a0d84\" (UID: \"ec2bf0dd-c16c-4912-a487-01711c8a0d84\") " Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.201173 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec2bf0dd-c16c-4912-a487-01711c8a0d84-nova-combined-ca-bundle\") pod \"ec2bf0dd-c16c-4912-a487-01711c8a0d84\" (UID: \"ec2bf0dd-c16c-4912-a487-01711c8a0d84\") " Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.201275 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/ec2bf0dd-c16c-4912-a487-01711c8a0d84-nova-extra-config-0\") pod \"ec2bf0dd-c16c-4912-a487-01711c8a0d84\" (UID: \"ec2bf0dd-c16c-4912-a487-01711c8a0d84\") " Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.201390 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/ec2bf0dd-c16c-4912-a487-01711c8a0d84-nova-cell1-compute-config-0\") pod \"ec2bf0dd-c16c-4912-a487-01711c8a0d84\" (UID: \"ec2bf0dd-c16c-4912-a487-01711c8a0d84\") " Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.206973 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec2bf0dd-c16c-4912-a487-01711c8a0d84-kube-api-access-hbsqt" (OuterVolumeSpecName: "kube-api-access-hbsqt") pod "ec2bf0dd-c16c-4912-a487-01711c8a0d84" (UID: "ec2bf0dd-c16c-4912-a487-01711c8a0d84"). InnerVolumeSpecName "kube-api-access-hbsqt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.217072 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec2bf0dd-c16c-4912-a487-01711c8a0d84-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "ec2bf0dd-c16c-4912-a487-01711c8a0d84" (UID: "ec2bf0dd-c16c-4912-a487-01711c8a0d84"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.241970 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec2bf0dd-c16c-4912-a487-01711c8a0d84-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "ec2bf0dd-c16c-4912-a487-01711c8a0d84" (UID: "ec2bf0dd-c16c-4912-a487-01711c8a0d84"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.242525 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec2bf0dd-c16c-4912-a487-01711c8a0d84-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "ec2bf0dd-c16c-4912-a487-01711c8a0d84" (UID: "ec2bf0dd-c16c-4912-a487-01711c8a0d84"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.245827 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec2bf0dd-c16c-4912-a487-01711c8a0d84-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "ec2bf0dd-c16c-4912-a487-01711c8a0d84" (UID: "ec2bf0dd-c16c-4912-a487-01711c8a0d84"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.249815 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec2bf0dd-c16c-4912-a487-01711c8a0d84-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ec2bf0dd-c16c-4912-a487-01711c8a0d84" (UID: "ec2bf0dd-c16c-4912-a487-01711c8a0d84"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.260708 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec2bf0dd-c16c-4912-a487-01711c8a0d84-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "ec2bf0dd-c16c-4912-a487-01711c8a0d84" (UID: "ec2bf0dd-c16c-4912-a487-01711c8a0d84"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.269719 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec2bf0dd-c16c-4912-a487-01711c8a0d84-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "ec2bf0dd-c16c-4912-a487-01711c8a0d84" (UID: "ec2bf0dd-c16c-4912-a487-01711c8a0d84"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.283887 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec2bf0dd-c16c-4912-a487-01711c8a0d84-inventory" (OuterVolumeSpecName: "inventory") pod "ec2bf0dd-c16c-4912-a487-01711c8a0d84" (UID: "ec2bf0dd-c16c-4912-a487-01711c8a0d84"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.304741 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ec2bf0dd-c16c-4912-a487-01711c8a0d84-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.304786 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ec2bf0dd-c16c-4912-a487-01711c8a0d84-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.304800 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbsqt\" (UniqueName: \"kubernetes.io/projected/ec2bf0dd-c16c-4912-a487-01711c8a0d84-kube-api-access-hbsqt\") on node \"crc\" DevicePath \"\"" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.304810 4809 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/ec2bf0dd-c16c-4912-a487-01711c8a0d84-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.304819 4809 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/ec2bf0dd-c16c-4912-a487-01711c8a0d84-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.304828 4809 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec2bf0dd-c16c-4912-a487-01711c8a0d84-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.304837 4809 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/ec2bf0dd-c16c-4912-a487-01711c8a0d84-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.304848 4809 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/ec2bf0dd-c16c-4912-a487-01711c8a0d84-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.304856 4809 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/ec2bf0dd-c16c-4912-a487-01711c8a0d84-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.608984 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njt6f" event={"ID":"ec2bf0dd-c16c-4912-a487-01711c8a0d84","Type":"ContainerDied","Data":"ffeafa35fd78205f77e78bb6e6c083936fed0e4b738ded903487b4601771b1e8"} Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.609283 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ffeafa35fd78205f77e78bb6e6c083936fed0e4b738ded903487b4601771b1e8" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.609048 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njt6f" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.728009 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mswz"] Dec 06 06:45:35 crc kubenswrapper[4809]: E1206 06:45:35.728510 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bf15846-fb7f-45fc-9146-02092049e864" containerName="collect-profiles" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.728527 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bf15846-fb7f-45fc-9146-02092049e864" containerName="collect-profiles" Dec 06 06:45:35 crc kubenswrapper[4809]: E1206 06:45:35.728546 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec2bf0dd-c16c-4912-a487-01711c8a0d84" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.728552 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec2bf0dd-c16c-4912-a487-01711c8a0d84" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.728776 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="9bf15846-fb7f-45fc-9146-02092049e864" containerName="collect-profiles" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.728795 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec2bf0dd-c16c-4912-a487-01711c8a0d84" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.729883 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mswz" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.736085 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.736087 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.736250 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.736690 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-bzlf5" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.740182 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mswz"] Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.744810 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.827453 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/44cad7cb-41c6-453b-895d-7c560e19f84f-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2mswz\" (UID: \"44cad7cb-41c6-453b-895d-7c560e19f84f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mswz" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.827506 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/44cad7cb-41c6-453b-895d-7c560e19f84f-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2mswz\" (UID: \"44cad7cb-41c6-453b-895d-7c560e19f84f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mswz" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.827677 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/44cad7cb-41c6-453b-895d-7c560e19f84f-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2mswz\" (UID: \"44cad7cb-41c6-453b-895d-7c560e19f84f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mswz" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.827869 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44cad7cb-41c6-453b-895d-7c560e19f84f-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2mswz\" (UID: \"44cad7cb-41c6-453b-895d-7c560e19f84f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mswz" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.827996 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vls2l\" (UniqueName: \"kubernetes.io/projected/44cad7cb-41c6-453b-895d-7c560e19f84f-kube-api-access-vls2l\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2mswz\" (UID: \"44cad7cb-41c6-453b-895d-7c560e19f84f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mswz" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.828047 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/44cad7cb-41c6-453b-895d-7c560e19f84f-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2mswz\" (UID: \"44cad7cb-41c6-453b-895d-7c560e19f84f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mswz" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.828080 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/44cad7cb-41c6-453b-895d-7c560e19f84f-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2mswz\" (UID: \"44cad7cb-41c6-453b-895d-7c560e19f84f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mswz" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.930354 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/44cad7cb-41c6-453b-895d-7c560e19f84f-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2mswz\" (UID: \"44cad7cb-41c6-453b-895d-7c560e19f84f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mswz" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.930450 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44cad7cb-41c6-453b-895d-7c560e19f84f-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2mswz\" (UID: \"44cad7cb-41c6-453b-895d-7c560e19f84f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mswz" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.930516 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vls2l\" (UniqueName: \"kubernetes.io/projected/44cad7cb-41c6-453b-895d-7c560e19f84f-kube-api-access-vls2l\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2mswz\" (UID: \"44cad7cb-41c6-453b-895d-7c560e19f84f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mswz" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.930561 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/44cad7cb-41c6-453b-895d-7c560e19f84f-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2mswz\" (UID: \"44cad7cb-41c6-453b-895d-7c560e19f84f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mswz" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.930610 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/44cad7cb-41c6-453b-895d-7c560e19f84f-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2mswz\" (UID: \"44cad7cb-41c6-453b-895d-7c560e19f84f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mswz" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.930687 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/44cad7cb-41c6-453b-895d-7c560e19f84f-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2mswz\" (UID: \"44cad7cb-41c6-453b-895d-7c560e19f84f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mswz" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.930722 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/44cad7cb-41c6-453b-895d-7c560e19f84f-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2mswz\" (UID: \"44cad7cb-41c6-453b-895d-7c560e19f84f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mswz" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.934655 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/44cad7cb-41c6-453b-895d-7c560e19f84f-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2mswz\" (UID: \"44cad7cb-41c6-453b-895d-7c560e19f84f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mswz" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.935723 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/44cad7cb-41c6-453b-895d-7c560e19f84f-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2mswz\" (UID: \"44cad7cb-41c6-453b-895d-7c560e19f84f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mswz" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.935794 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/44cad7cb-41c6-453b-895d-7c560e19f84f-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2mswz\" (UID: \"44cad7cb-41c6-453b-895d-7c560e19f84f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mswz" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.935907 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/44cad7cb-41c6-453b-895d-7c560e19f84f-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2mswz\" (UID: \"44cad7cb-41c6-453b-895d-7c560e19f84f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mswz" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.940825 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44cad7cb-41c6-453b-895d-7c560e19f84f-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2mswz\" (UID: \"44cad7cb-41c6-453b-895d-7c560e19f84f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mswz" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.949318 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/44cad7cb-41c6-453b-895d-7c560e19f84f-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2mswz\" (UID: \"44cad7cb-41c6-453b-895d-7c560e19f84f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mswz" Dec 06 06:45:35 crc kubenswrapper[4809]: I1206 06:45:35.951832 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vls2l\" (UniqueName: \"kubernetes.io/projected/44cad7cb-41c6-453b-895d-7c560e19f84f-kube-api-access-vls2l\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2mswz\" (UID: \"44cad7cb-41c6-453b-895d-7c560e19f84f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mswz" Dec 06 06:45:36 crc kubenswrapper[4809]: I1206 06:45:36.053489 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mswz" Dec 06 06:45:36 crc kubenswrapper[4809]: I1206 06:45:36.665053 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mswz"] Dec 06 06:45:37 crc kubenswrapper[4809]: I1206 06:45:37.627781 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mswz" event={"ID":"44cad7cb-41c6-453b-895d-7c560e19f84f","Type":"ContainerStarted","Data":"ca0571bb800aabd65397a72977a92651bde7e62b114fc7497fb7887bdedb07e3"} Dec 06 06:45:40 crc kubenswrapper[4809]: I1206 06:45:40.708445 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mswz" event={"ID":"44cad7cb-41c6-453b-895d-7c560e19f84f","Type":"ContainerStarted","Data":"48ae1e0a34a6164dd1081399bd3a368d9ab007ea430c7e5332d1eb65b24b7d64"} Dec 06 06:45:43 crc kubenswrapper[4809]: I1206 06:45:43.778992 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mswz" podStartSLOduration=6.157291838 podStartE2EDuration="8.778974197s" podCreationTimestamp="2025-12-06 06:45:35 +0000 UTC" firstStartedPulling="2025-12-06 06:45:36.663328518 +0000 UTC m=+3261.552311460" lastFinishedPulling="2025-12-06 06:45:39.285010867 +0000 UTC m=+3264.173993819" observedRunningTime="2025-12-06 06:45:43.775331528 +0000 UTC m=+3268.664314470" watchObservedRunningTime="2025-12-06 06:45:43.778974197 +0000 UTC m=+3268.667957139" Dec 06 06:45:48 crc kubenswrapper[4809]: I1206 06:45:48.389188 4809 scope.go:117] "RemoveContainer" containerID="ca23ae36692bfc6cbf4be3666821469da7cf4aff660aa8bd1e9a06863c4d9e82" Dec 06 06:45:48 crc kubenswrapper[4809]: E1206 06:45:48.389983 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:45:59 crc kubenswrapper[4809]: I1206 06:45:59.388739 4809 scope.go:117] "RemoveContainer" containerID="ca23ae36692bfc6cbf4be3666821469da7cf4aff660aa8bd1e9a06863c4d9e82" Dec 06 06:45:59 crc kubenswrapper[4809]: E1206 06:45:59.389557 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:46:05 crc kubenswrapper[4809]: I1206 06:46:05.473237 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-t4skg"] Dec 06 06:46:05 crc kubenswrapper[4809]: I1206 06:46:05.476552 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t4skg" Dec 06 06:46:05 crc kubenswrapper[4809]: I1206 06:46:05.486304 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t4skg"] Dec 06 06:46:05 crc kubenswrapper[4809]: I1206 06:46:05.627298 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsnrm\" (UniqueName: \"kubernetes.io/projected/b80a4f13-776f-42db-9da2-17df1a676fa9-kube-api-access-qsnrm\") pod \"certified-operators-t4skg\" (UID: \"b80a4f13-776f-42db-9da2-17df1a676fa9\") " pod="openshift-marketplace/certified-operators-t4skg" Dec 06 06:46:05 crc kubenswrapper[4809]: I1206 06:46:05.627591 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b80a4f13-776f-42db-9da2-17df1a676fa9-catalog-content\") pod \"certified-operators-t4skg\" (UID: \"b80a4f13-776f-42db-9da2-17df1a676fa9\") " pod="openshift-marketplace/certified-operators-t4skg" Dec 06 06:46:05 crc kubenswrapper[4809]: I1206 06:46:05.627664 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b80a4f13-776f-42db-9da2-17df1a676fa9-utilities\") pod \"certified-operators-t4skg\" (UID: \"b80a4f13-776f-42db-9da2-17df1a676fa9\") " pod="openshift-marketplace/certified-operators-t4skg" Dec 06 06:46:05 crc kubenswrapper[4809]: I1206 06:46:05.729908 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b80a4f13-776f-42db-9da2-17df1a676fa9-catalog-content\") pod \"certified-operators-t4skg\" (UID: \"b80a4f13-776f-42db-9da2-17df1a676fa9\") " pod="openshift-marketplace/certified-operators-t4skg" Dec 06 06:46:05 crc kubenswrapper[4809]: I1206 06:46:05.730012 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b80a4f13-776f-42db-9da2-17df1a676fa9-utilities\") pod \"certified-operators-t4skg\" (UID: \"b80a4f13-776f-42db-9da2-17df1a676fa9\") " pod="openshift-marketplace/certified-operators-t4skg" Dec 06 06:46:05 crc kubenswrapper[4809]: I1206 06:46:05.730105 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsnrm\" (UniqueName: \"kubernetes.io/projected/b80a4f13-776f-42db-9da2-17df1a676fa9-kube-api-access-qsnrm\") pod \"certified-operators-t4skg\" (UID: \"b80a4f13-776f-42db-9da2-17df1a676fa9\") " pod="openshift-marketplace/certified-operators-t4skg" Dec 06 06:46:05 crc kubenswrapper[4809]: I1206 06:46:05.730608 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b80a4f13-776f-42db-9da2-17df1a676fa9-catalog-content\") pod \"certified-operators-t4skg\" (UID: \"b80a4f13-776f-42db-9da2-17df1a676fa9\") " pod="openshift-marketplace/certified-operators-t4skg" Dec 06 06:46:05 crc kubenswrapper[4809]: I1206 06:46:05.730615 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b80a4f13-776f-42db-9da2-17df1a676fa9-utilities\") pod \"certified-operators-t4skg\" (UID: \"b80a4f13-776f-42db-9da2-17df1a676fa9\") " pod="openshift-marketplace/certified-operators-t4skg" Dec 06 06:46:05 crc kubenswrapper[4809]: I1206 06:46:05.752288 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsnrm\" (UniqueName: \"kubernetes.io/projected/b80a4f13-776f-42db-9da2-17df1a676fa9-kube-api-access-qsnrm\") pod \"certified-operators-t4skg\" (UID: \"b80a4f13-776f-42db-9da2-17df1a676fa9\") " pod="openshift-marketplace/certified-operators-t4skg" Dec 06 06:46:05 crc kubenswrapper[4809]: I1206 06:46:05.805258 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t4skg" Dec 06 06:46:07 crc kubenswrapper[4809]: I1206 06:46:07.415271 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t4skg"] Dec 06 06:46:08 crc kubenswrapper[4809]: I1206 06:46:08.047460 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t4skg" event={"ID":"b80a4f13-776f-42db-9da2-17df1a676fa9","Type":"ContainerStarted","Data":"ff69d1996d94743d9640423bd25fdeb71bd1b7ff63239f8680f604f65d22d7f6"} Dec 06 06:46:09 crc kubenswrapper[4809]: I1206 06:46:09.059889 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t4skg" event={"ID":"b80a4f13-776f-42db-9da2-17df1a676fa9","Type":"ContainerStarted","Data":"0e936aa2931c18819ae90c5e32baa42206dbc8181c3a1e2d0f3babbbcf28dab6"} Dec 06 06:46:10 crc kubenswrapper[4809]: I1206 06:46:10.078224 4809 generic.go:334] "Generic (PLEG): container finished" podID="b80a4f13-776f-42db-9da2-17df1a676fa9" containerID="0e936aa2931c18819ae90c5e32baa42206dbc8181c3a1e2d0f3babbbcf28dab6" exitCode=0 Dec 06 06:46:10 crc kubenswrapper[4809]: I1206 06:46:10.078280 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t4skg" event={"ID":"b80a4f13-776f-42db-9da2-17df1a676fa9","Type":"ContainerDied","Data":"0e936aa2931c18819ae90c5e32baa42206dbc8181c3a1e2d0f3babbbcf28dab6"} Dec 06 06:46:12 crc kubenswrapper[4809]: I1206 06:46:12.388717 4809 scope.go:117] "RemoveContainer" containerID="ca23ae36692bfc6cbf4be3666821469da7cf4aff660aa8bd1e9a06863c4d9e82" Dec 06 06:46:12 crc kubenswrapper[4809]: E1206 06:46:12.389486 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:46:23 crc kubenswrapper[4809]: I1206 06:46:23.259088 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t4skg" event={"ID":"b80a4f13-776f-42db-9da2-17df1a676fa9","Type":"ContainerStarted","Data":"775eb54497c807cea5cc28330a3d52220b0acccedd0b180c74f253851c5bc18b"} Dec 06 06:46:26 crc kubenswrapper[4809]: I1206 06:46:26.298280 4809 generic.go:334] "Generic (PLEG): container finished" podID="b80a4f13-776f-42db-9da2-17df1a676fa9" containerID="775eb54497c807cea5cc28330a3d52220b0acccedd0b180c74f253851c5bc18b" exitCode=0 Dec 06 06:46:26 crc kubenswrapper[4809]: I1206 06:46:26.298373 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t4skg" event={"ID":"b80a4f13-776f-42db-9da2-17df1a676fa9","Type":"ContainerDied","Data":"775eb54497c807cea5cc28330a3d52220b0acccedd0b180c74f253851c5bc18b"} Dec 06 06:46:26 crc kubenswrapper[4809]: I1206 06:46:26.301851 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 06:46:26 crc kubenswrapper[4809]: I1206 06:46:26.389947 4809 scope.go:117] "RemoveContainer" containerID="ca23ae36692bfc6cbf4be3666821469da7cf4aff660aa8bd1e9a06863c4d9e82" Dec 06 06:46:26 crc kubenswrapper[4809]: E1206 06:46:26.390398 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:46:41 crc kubenswrapper[4809]: I1206 06:46:41.390062 4809 scope.go:117] "RemoveContainer" containerID="ca23ae36692bfc6cbf4be3666821469da7cf4aff660aa8bd1e9a06863c4d9e82" Dec 06 06:46:41 crc kubenswrapper[4809]: E1206 06:46:41.391584 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:46:42 crc kubenswrapper[4809]: I1206 06:46:42.527296 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t4skg" event={"ID":"b80a4f13-776f-42db-9da2-17df1a676fa9","Type":"ContainerStarted","Data":"baad2a4994b17da114ebe7a2b96ec8de685b89d9ef5d3a22fd1704471cf60ff9"} Dec 06 06:46:43 crc kubenswrapper[4809]: I1206 06:46:43.559099 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-t4skg" podStartSLOduration=7.099450562 podStartE2EDuration="38.559078834s" podCreationTimestamp="2025-12-06 06:46:05 +0000 UTC" firstStartedPulling="2025-12-06 06:46:10.081121457 +0000 UTC m=+3294.970104399" lastFinishedPulling="2025-12-06 06:46:41.540749729 +0000 UTC m=+3326.429732671" observedRunningTime="2025-12-06 06:46:43.552367992 +0000 UTC m=+3328.441350934" watchObservedRunningTime="2025-12-06 06:46:43.559078834 +0000 UTC m=+3328.448061786" Dec 06 06:46:45 crc kubenswrapper[4809]: I1206 06:46:45.805449 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-t4skg" Dec 06 06:46:45 crc kubenswrapper[4809]: I1206 06:46:45.806731 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-t4skg" Dec 06 06:46:45 crc kubenswrapper[4809]: I1206 06:46:45.853283 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-t4skg" Dec 06 06:46:47 crc kubenswrapper[4809]: I1206 06:46:47.628056 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-t4skg" Dec 06 06:46:47 crc kubenswrapper[4809]: I1206 06:46:47.680406 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t4skg"] Dec 06 06:46:49 crc kubenswrapper[4809]: I1206 06:46:49.606409 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-t4skg" podUID="b80a4f13-776f-42db-9da2-17df1a676fa9" containerName="registry-server" containerID="cri-o://baad2a4994b17da114ebe7a2b96ec8de685b89d9ef5d3a22fd1704471cf60ff9" gracePeriod=2 Dec 06 06:46:50 crc kubenswrapper[4809]: I1206 06:46:50.628460 4809 generic.go:334] "Generic (PLEG): container finished" podID="b80a4f13-776f-42db-9da2-17df1a676fa9" containerID="baad2a4994b17da114ebe7a2b96ec8de685b89d9ef5d3a22fd1704471cf60ff9" exitCode=0 Dec 06 06:46:50 crc kubenswrapper[4809]: I1206 06:46:50.629083 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t4skg" event={"ID":"b80a4f13-776f-42db-9da2-17df1a676fa9","Type":"ContainerDied","Data":"baad2a4994b17da114ebe7a2b96ec8de685b89d9ef5d3a22fd1704471cf60ff9"} Dec 06 06:46:50 crc kubenswrapper[4809]: I1206 06:46:50.801578 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t4skg" Dec 06 06:46:50 crc kubenswrapper[4809]: I1206 06:46:50.868808 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qsnrm\" (UniqueName: \"kubernetes.io/projected/b80a4f13-776f-42db-9da2-17df1a676fa9-kube-api-access-qsnrm\") pod \"b80a4f13-776f-42db-9da2-17df1a676fa9\" (UID: \"b80a4f13-776f-42db-9da2-17df1a676fa9\") " Dec 06 06:46:50 crc kubenswrapper[4809]: I1206 06:46:50.869007 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b80a4f13-776f-42db-9da2-17df1a676fa9-utilities\") pod \"b80a4f13-776f-42db-9da2-17df1a676fa9\" (UID: \"b80a4f13-776f-42db-9da2-17df1a676fa9\") " Dec 06 06:46:50 crc kubenswrapper[4809]: I1206 06:46:50.869204 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b80a4f13-776f-42db-9da2-17df1a676fa9-catalog-content\") pod \"b80a4f13-776f-42db-9da2-17df1a676fa9\" (UID: \"b80a4f13-776f-42db-9da2-17df1a676fa9\") " Dec 06 06:46:50 crc kubenswrapper[4809]: I1206 06:46:50.870010 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b80a4f13-776f-42db-9da2-17df1a676fa9-utilities" (OuterVolumeSpecName: "utilities") pod "b80a4f13-776f-42db-9da2-17df1a676fa9" (UID: "b80a4f13-776f-42db-9da2-17df1a676fa9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:46:50 crc kubenswrapper[4809]: I1206 06:46:50.873548 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b80a4f13-776f-42db-9da2-17df1a676fa9-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:46:50 crc kubenswrapper[4809]: I1206 06:46:50.874571 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b80a4f13-776f-42db-9da2-17df1a676fa9-kube-api-access-qsnrm" (OuterVolumeSpecName: "kube-api-access-qsnrm") pod "b80a4f13-776f-42db-9da2-17df1a676fa9" (UID: "b80a4f13-776f-42db-9da2-17df1a676fa9"). InnerVolumeSpecName "kube-api-access-qsnrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:46:50 crc kubenswrapper[4809]: I1206 06:46:50.925447 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b80a4f13-776f-42db-9da2-17df1a676fa9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b80a4f13-776f-42db-9da2-17df1a676fa9" (UID: "b80a4f13-776f-42db-9da2-17df1a676fa9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:46:50 crc kubenswrapper[4809]: I1206 06:46:50.975322 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qsnrm\" (UniqueName: \"kubernetes.io/projected/b80a4f13-776f-42db-9da2-17df1a676fa9-kube-api-access-qsnrm\") on node \"crc\" DevicePath \"\"" Dec 06 06:46:50 crc kubenswrapper[4809]: I1206 06:46:50.975365 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b80a4f13-776f-42db-9da2-17df1a676fa9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:46:51 crc kubenswrapper[4809]: I1206 06:46:51.689247 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t4skg" event={"ID":"b80a4f13-776f-42db-9da2-17df1a676fa9","Type":"ContainerDied","Data":"ff69d1996d94743d9640423bd25fdeb71bd1b7ff63239f8680f604f65d22d7f6"} Dec 06 06:46:51 crc kubenswrapper[4809]: I1206 06:46:51.689664 4809 scope.go:117] "RemoveContainer" containerID="baad2a4994b17da114ebe7a2b96ec8de685b89d9ef5d3a22fd1704471cf60ff9" Dec 06 06:46:51 crc kubenswrapper[4809]: I1206 06:46:51.689315 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t4skg" Dec 06 06:46:51 crc kubenswrapper[4809]: I1206 06:46:51.725446 4809 scope.go:117] "RemoveContainer" containerID="775eb54497c807cea5cc28330a3d52220b0acccedd0b180c74f253851c5bc18b" Dec 06 06:46:51 crc kubenswrapper[4809]: I1206 06:46:51.727991 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t4skg"] Dec 06 06:46:51 crc kubenswrapper[4809]: I1206 06:46:51.749222 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-t4skg"] Dec 06 06:46:51 crc kubenswrapper[4809]: I1206 06:46:51.769620 4809 scope.go:117] "RemoveContainer" containerID="0e936aa2931c18819ae90c5e32baa42206dbc8181c3a1e2d0f3babbbcf28dab6" Dec 06 06:46:52 crc kubenswrapper[4809]: I1206 06:46:52.388830 4809 scope.go:117] "RemoveContainer" containerID="ca23ae36692bfc6cbf4be3666821469da7cf4aff660aa8bd1e9a06863c4d9e82" Dec 06 06:46:52 crc kubenswrapper[4809]: E1206 06:46:52.389118 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:46:53 crc kubenswrapper[4809]: I1206 06:46:53.408515 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b80a4f13-776f-42db-9da2-17df1a676fa9" path="/var/lib/kubelet/pods/b80a4f13-776f-42db-9da2-17df1a676fa9/volumes" Dec 06 06:47:03 crc kubenswrapper[4809]: I1206 06:47:03.391818 4809 scope.go:117] "RemoveContainer" containerID="ca23ae36692bfc6cbf4be3666821469da7cf4aff660aa8bd1e9a06863c4d9e82" Dec 06 06:47:03 crc kubenswrapper[4809]: E1206 06:47:03.393007 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:47:18 crc kubenswrapper[4809]: I1206 06:47:18.389286 4809 scope.go:117] "RemoveContainer" containerID="ca23ae36692bfc6cbf4be3666821469da7cf4aff660aa8bd1e9a06863c4d9e82" Dec 06 06:47:21 crc kubenswrapper[4809]: I1206 06:47:21.034235 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" event={"ID":"cbf4cf62-024e-4703-a8b8-9aecda9cd26a","Type":"ContainerStarted","Data":"083a8276457cb2db9f64bc581dbeeb57fcdb1feabe656331716e6ce96734d239"} Dec 06 06:47:53 crc kubenswrapper[4809]: I1206 06:47:53.550137 4809 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-r2lg6 container/marketplace-operator namespace/openshift-marketplace: Liveness probe status=failure output="Get \"http://10.217.0.63:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 06:47:53 crc kubenswrapper[4809]: I1206 06:47:53.550674 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/marketplace-operator-79b997595-r2lg6" podUID="286c8de4-0ea9-43ad-bb4f-970319e0f4d3" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.63:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 06:48:20 crc kubenswrapper[4809]: I1206 06:48:20.735784 4809 generic.go:334] "Generic (PLEG): container finished" podID="44cad7cb-41c6-453b-895d-7c560e19f84f" containerID="48ae1e0a34a6164dd1081399bd3a368d9ab007ea430c7e5332d1eb65b24b7d64" exitCode=0 Dec 06 06:48:20 crc kubenswrapper[4809]: I1206 06:48:20.735850 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mswz" event={"ID":"44cad7cb-41c6-453b-895d-7c560e19f84f","Type":"ContainerDied","Data":"48ae1e0a34a6164dd1081399bd3a368d9ab007ea430c7e5332d1eb65b24b7d64"} Dec 06 06:48:22 crc kubenswrapper[4809]: I1206 06:48:22.243400 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mswz" Dec 06 06:48:22 crc kubenswrapper[4809]: I1206 06:48:22.348683 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vls2l\" (UniqueName: \"kubernetes.io/projected/44cad7cb-41c6-453b-895d-7c560e19f84f-kube-api-access-vls2l\") pod \"44cad7cb-41c6-453b-895d-7c560e19f84f\" (UID: \"44cad7cb-41c6-453b-895d-7c560e19f84f\") " Dec 06 06:48:22 crc kubenswrapper[4809]: I1206 06:48:22.348749 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/44cad7cb-41c6-453b-895d-7c560e19f84f-ceilometer-compute-config-data-1\") pod \"44cad7cb-41c6-453b-895d-7c560e19f84f\" (UID: \"44cad7cb-41c6-453b-895d-7c560e19f84f\") " Dec 06 06:48:22 crc kubenswrapper[4809]: I1206 06:48:22.348811 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/44cad7cb-41c6-453b-895d-7c560e19f84f-inventory\") pod \"44cad7cb-41c6-453b-895d-7c560e19f84f\" (UID: \"44cad7cb-41c6-453b-895d-7c560e19f84f\") " Dec 06 06:48:22 crc kubenswrapper[4809]: I1206 06:48:22.348868 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/44cad7cb-41c6-453b-895d-7c560e19f84f-ceilometer-compute-config-data-0\") pod \"44cad7cb-41c6-453b-895d-7c560e19f84f\" (UID: \"44cad7cb-41c6-453b-895d-7c560e19f84f\") " Dec 06 06:48:22 crc kubenswrapper[4809]: I1206 06:48:22.348924 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/44cad7cb-41c6-453b-895d-7c560e19f84f-ssh-key\") pod \"44cad7cb-41c6-453b-895d-7c560e19f84f\" (UID: \"44cad7cb-41c6-453b-895d-7c560e19f84f\") " Dec 06 06:48:22 crc kubenswrapper[4809]: I1206 06:48:22.349002 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/44cad7cb-41c6-453b-895d-7c560e19f84f-ceilometer-compute-config-data-2\") pod \"44cad7cb-41c6-453b-895d-7c560e19f84f\" (UID: \"44cad7cb-41c6-453b-895d-7c560e19f84f\") " Dec 06 06:48:22 crc kubenswrapper[4809]: I1206 06:48:22.349213 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44cad7cb-41c6-453b-895d-7c560e19f84f-telemetry-combined-ca-bundle\") pod \"44cad7cb-41c6-453b-895d-7c560e19f84f\" (UID: \"44cad7cb-41c6-453b-895d-7c560e19f84f\") " Dec 06 06:48:22 crc kubenswrapper[4809]: I1206 06:48:22.356158 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44cad7cb-41c6-453b-895d-7c560e19f84f-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "44cad7cb-41c6-453b-895d-7c560e19f84f" (UID: "44cad7cb-41c6-453b-895d-7c560e19f84f"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:48:22 crc kubenswrapper[4809]: I1206 06:48:22.357784 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44cad7cb-41c6-453b-895d-7c560e19f84f-kube-api-access-vls2l" (OuterVolumeSpecName: "kube-api-access-vls2l") pod "44cad7cb-41c6-453b-895d-7c560e19f84f" (UID: "44cad7cb-41c6-453b-895d-7c560e19f84f"). InnerVolumeSpecName "kube-api-access-vls2l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:48:22 crc kubenswrapper[4809]: I1206 06:48:22.385797 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44cad7cb-41c6-453b-895d-7c560e19f84f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "44cad7cb-41c6-453b-895d-7c560e19f84f" (UID: "44cad7cb-41c6-453b-895d-7c560e19f84f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:48:22 crc kubenswrapper[4809]: I1206 06:48:22.386918 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44cad7cb-41c6-453b-895d-7c560e19f84f-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "44cad7cb-41c6-453b-895d-7c560e19f84f" (UID: "44cad7cb-41c6-453b-895d-7c560e19f84f"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:48:22 crc kubenswrapper[4809]: I1206 06:48:22.390718 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44cad7cb-41c6-453b-895d-7c560e19f84f-inventory" (OuterVolumeSpecName: "inventory") pod "44cad7cb-41c6-453b-895d-7c560e19f84f" (UID: "44cad7cb-41c6-453b-895d-7c560e19f84f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:48:22 crc kubenswrapper[4809]: I1206 06:48:22.391761 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44cad7cb-41c6-453b-895d-7c560e19f84f-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "44cad7cb-41c6-453b-895d-7c560e19f84f" (UID: "44cad7cb-41c6-453b-895d-7c560e19f84f"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:48:22 crc kubenswrapper[4809]: I1206 06:48:22.398338 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44cad7cb-41c6-453b-895d-7c560e19f84f-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "44cad7cb-41c6-453b-895d-7c560e19f84f" (UID: "44cad7cb-41c6-453b-895d-7c560e19f84f"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:48:22 crc kubenswrapper[4809]: I1206 06:48:22.453183 4809 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44cad7cb-41c6-453b-895d-7c560e19f84f-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:48:22 crc kubenswrapper[4809]: I1206 06:48:22.453230 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vls2l\" (UniqueName: \"kubernetes.io/projected/44cad7cb-41c6-453b-895d-7c560e19f84f-kube-api-access-vls2l\") on node \"crc\" DevicePath \"\"" Dec 06 06:48:22 crc kubenswrapper[4809]: I1206 06:48:22.453245 4809 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/44cad7cb-41c6-453b-895d-7c560e19f84f-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 06 06:48:22 crc kubenswrapper[4809]: I1206 06:48:22.453259 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/44cad7cb-41c6-453b-895d-7c560e19f84f-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 06:48:22 crc kubenswrapper[4809]: I1206 06:48:22.453275 4809 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/44cad7cb-41c6-453b-895d-7c560e19f84f-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:48:22 crc kubenswrapper[4809]: I1206 06:48:22.453288 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/44cad7cb-41c6-453b-895d-7c560e19f84f-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 06:48:22 crc kubenswrapper[4809]: I1206 06:48:22.453300 4809 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/44cad7cb-41c6-453b-895d-7c560e19f84f-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 06 06:48:22 crc kubenswrapper[4809]: I1206 06:48:22.761437 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mswz" event={"ID":"44cad7cb-41c6-453b-895d-7c560e19f84f","Type":"ContainerDied","Data":"ca0571bb800aabd65397a72977a92651bde7e62b114fc7497fb7887bdedb07e3"} Dec 06 06:48:22 crc kubenswrapper[4809]: I1206 06:48:22.761484 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ca0571bb800aabd65397a72977a92651bde7e62b114fc7497fb7887bdedb07e3" Dec 06 06:48:22 crc kubenswrapper[4809]: I1206 06:48:22.761512 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mswz" Dec 06 06:48:22 crc kubenswrapper[4809]: I1206 06:48:22.879103 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw"] Dec 06 06:48:22 crc kubenswrapper[4809]: E1206 06:48:22.879975 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44cad7cb-41c6-453b-895d-7c560e19f84f" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 06 06:48:22 crc kubenswrapper[4809]: I1206 06:48:22.880003 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="44cad7cb-41c6-453b-895d-7c560e19f84f" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 06 06:48:22 crc kubenswrapper[4809]: E1206 06:48:22.880022 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b80a4f13-776f-42db-9da2-17df1a676fa9" containerName="registry-server" Dec 06 06:48:22 crc kubenswrapper[4809]: I1206 06:48:22.880047 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b80a4f13-776f-42db-9da2-17df1a676fa9" containerName="registry-server" Dec 06 06:48:22 crc kubenswrapper[4809]: E1206 06:48:22.880113 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b80a4f13-776f-42db-9da2-17df1a676fa9" containerName="extract-content" Dec 06 06:48:22 crc kubenswrapper[4809]: I1206 06:48:22.880122 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b80a4f13-776f-42db-9da2-17df1a676fa9" containerName="extract-content" Dec 06 06:48:22 crc kubenswrapper[4809]: E1206 06:48:22.880148 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b80a4f13-776f-42db-9da2-17df1a676fa9" containerName="extract-utilities" Dec 06 06:48:22 crc kubenswrapper[4809]: I1206 06:48:22.880157 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b80a4f13-776f-42db-9da2-17df1a676fa9" containerName="extract-utilities" Dec 06 06:48:22 crc kubenswrapper[4809]: I1206 06:48:22.880467 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b80a4f13-776f-42db-9da2-17df1a676fa9" containerName="registry-server" Dec 06 06:48:22 crc kubenswrapper[4809]: I1206 06:48:22.880497 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="44cad7cb-41c6-453b-895d-7c560e19f84f" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 06 06:48:22 crc kubenswrapper[4809]: I1206 06:48:22.881821 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw" Dec 06 06:48:22 crc kubenswrapper[4809]: I1206 06:48:22.884553 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 06:48:22 crc kubenswrapper[4809]: I1206 06:48:22.884654 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 06:48:22 crc kubenswrapper[4809]: I1206 06:48:22.885220 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-bzlf5" Dec 06 06:48:22 crc kubenswrapper[4809]: I1206 06:48:22.888105 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 06:48:22 crc kubenswrapper[4809]: I1206 06:48:22.897413 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-ipmi-config-data" Dec 06 06:48:22 crc kubenswrapper[4809]: I1206 06:48:22.898378 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw"] Dec 06 06:48:22 crc kubenswrapper[4809]: I1206 06:48:22.965724 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0a499909-59b6-42ec-a9f7-75218a100ead-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw\" (UID: \"0a499909-59b6-42ec-a9f7-75218a100ead\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw" Dec 06 06:48:22 crc kubenswrapper[4809]: I1206 06:48:22.965948 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/0a499909-59b6-42ec-a9f7-75218a100ead-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw\" (UID: \"0a499909-59b6-42ec-a9f7-75218a100ead\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw" Dec 06 06:48:22 crc kubenswrapper[4809]: I1206 06:48:22.966001 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/0a499909-59b6-42ec-a9f7-75218a100ead-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw\" (UID: \"0a499909-59b6-42ec-a9f7-75218a100ead\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw" Dec 06 06:48:22 crc kubenswrapper[4809]: I1206 06:48:22.966274 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/0a499909-59b6-42ec-a9f7-75218a100ead-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw\" (UID: \"0a499909-59b6-42ec-a9f7-75218a100ead\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw" Dec 06 06:48:22 crc kubenswrapper[4809]: I1206 06:48:22.966777 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a499909-59b6-42ec-a9f7-75218a100ead-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw\" (UID: \"0a499909-59b6-42ec-a9f7-75218a100ead\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw" Dec 06 06:48:22 crc kubenswrapper[4809]: I1206 06:48:22.967254 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhzsc\" (UniqueName: \"kubernetes.io/projected/0a499909-59b6-42ec-a9f7-75218a100ead-kube-api-access-xhzsc\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw\" (UID: \"0a499909-59b6-42ec-a9f7-75218a100ead\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw" Dec 06 06:48:22 crc kubenswrapper[4809]: I1206 06:48:22.967485 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0a499909-59b6-42ec-a9f7-75218a100ead-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw\" (UID: \"0a499909-59b6-42ec-a9f7-75218a100ead\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw" Dec 06 06:48:23 crc kubenswrapper[4809]: I1206 06:48:23.070202 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/0a499909-59b6-42ec-a9f7-75218a100ead-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw\" (UID: \"0a499909-59b6-42ec-a9f7-75218a100ead\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw" Dec 06 06:48:23 crc kubenswrapper[4809]: I1206 06:48:23.070293 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/0a499909-59b6-42ec-a9f7-75218a100ead-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw\" (UID: \"0a499909-59b6-42ec-a9f7-75218a100ead\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw" Dec 06 06:48:23 crc kubenswrapper[4809]: I1206 06:48:23.070409 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/0a499909-59b6-42ec-a9f7-75218a100ead-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw\" (UID: \"0a499909-59b6-42ec-a9f7-75218a100ead\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw" Dec 06 06:48:23 crc kubenswrapper[4809]: I1206 06:48:23.070569 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a499909-59b6-42ec-a9f7-75218a100ead-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw\" (UID: \"0a499909-59b6-42ec-a9f7-75218a100ead\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw" Dec 06 06:48:23 crc kubenswrapper[4809]: I1206 06:48:23.070673 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhzsc\" (UniqueName: \"kubernetes.io/projected/0a499909-59b6-42ec-a9f7-75218a100ead-kube-api-access-xhzsc\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw\" (UID: \"0a499909-59b6-42ec-a9f7-75218a100ead\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw" Dec 06 06:48:23 crc kubenswrapper[4809]: I1206 06:48:23.070778 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0a499909-59b6-42ec-a9f7-75218a100ead-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw\" (UID: \"0a499909-59b6-42ec-a9f7-75218a100ead\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw" Dec 06 06:48:23 crc kubenswrapper[4809]: I1206 06:48:23.070834 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0a499909-59b6-42ec-a9f7-75218a100ead-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw\" (UID: \"0a499909-59b6-42ec-a9f7-75218a100ead\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw" Dec 06 06:48:23 crc kubenswrapper[4809]: I1206 06:48:23.075625 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a499909-59b6-42ec-a9f7-75218a100ead-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw\" (UID: \"0a499909-59b6-42ec-a9f7-75218a100ead\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw" Dec 06 06:48:23 crc kubenswrapper[4809]: I1206 06:48:23.077139 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/0a499909-59b6-42ec-a9f7-75218a100ead-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw\" (UID: \"0a499909-59b6-42ec-a9f7-75218a100ead\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw" Dec 06 06:48:23 crc kubenswrapper[4809]: I1206 06:48:23.077618 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/0a499909-59b6-42ec-a9f7-75218a100ead-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw\" (UID: \"0a499909-59b6-42ec-a9f7-75218a100ead\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw" Dec 06 06:48:23 crc kubenswrapper[4809]: I1206 06:48:23.078621 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/0a499909-59b6-42ec-a9f7-75218a100ead-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw\" (UID: \"0a499909-59b6-42ec-a9f7-75218a100ead\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw" Dec 06 06:48:23 crc kubenswrapper[4809]: I1206 06:48:23.078759 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0a499909-59b6-42ec-a9f7-75218a100ead-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw\" (UID: \"0a499909-59b6-42ec-a9f7-75218a100ead\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw" Dec 06 06:48:23 crc kubenswrapper[4809]: I1206 06:48:23.083274 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0a499909-59b6-42ec-a9f7-75218a100ead-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw\" (UID: \"0a499909-59b6-42ec-a9f7-75218a100ead\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw" Dec 06 06:48:23 crc kubenswrapper[4809]: I1206 06:48:23.094538 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhzsc\" (UniqueName: \"kubernetes.io/projected/0a499909-59b6-42ec-a9f7-75218a100ead-kube-api-access-xhzsc\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw\" (UID: \"0a499909-59b6-42ec-a9f7-75218a100ead\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw" Dec 06 06:48:23 crc kubenswrapper[4809]: I1206 06:48:23.213966 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw" Dec 06 06:48:23 crc kubenswrapper[4809]: I1206 06:48:23.795727 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw"] Dec 06 06:48:23 crc kubenswrapper[4809]: W1206 06:48:23.800420 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0a499909_59b6_42ec_a9f7_75218a100ead.slice/crio-27bf1dbfd78ed4f30e5bfb237bb3b3295bc11dd15c395ea2066c50cf49eaff2e WatchSource:0}: Error finding container 27bf1dbfd78ed4f30e5bfb237bb3b3295bc11dd15c395ea2066c50cf49eaff2e: Status 404 returned error can't find the container with id 27bf1dbfd78ed4f30e5bfb237bb3b3295bc11dd15c395ea2066c50cf49eaff2e Dec 06 06:48:24 crc kubenswrapper[4809]: I1206 06:48:24.786503 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw" event={"ID":"0a499909-59b6-42ec-a9f7-75218a100ead","Type":"ContainerStarted","Data":"24a8bd65bcf0b39b4ebbd77eac15afb4c3b67605f2d3c41ce441271fe3fe9ed9"} Dec 06 06:48:24 crc kubenswrapper[4809]: I1206 06:48:24.787151 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw" event={"ID":"0a499909-59b6-42ec-a9f7-75218a100ead","Type":"ContainerStarted","Data":"27bf1dbfd78ed4f30e5bfb237bb3b3295bc11dd15c395ea2066c50cf49eaff2e"} Dec 06 06:48:24 crc kubenswrapper[4809]: I1206 06:48:24.807651 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw" podStartSLOduration=2.361860322 podStartE2EDuration="2.807629259s" podCreationTimestamp="2025-12-06 06:48:22 +0000 UTC" firstStartedPulling="2025-12-06 06:48:23.803895458 +0000 UTC m=+3428.692878400" lastFinishedPulling="2025-12-06 06:48:24.249664395 +0000 UTC m=+3429.138647337" observedRunningTime="2025-12-06 06:48:24.804611438 +0000 UTC m=+3429.693594390" watchObservedRunningTime="2025-12-06 06:48:24.807629259 +0000 UTC m=+3429.696612211" Dec 06 06:48:58 crc kubenswrapper[4809]: I1206 06:48:58.738140 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-msjdt"] Dec 06 06:48:58 crc kubenswrapper[4809]: I1206 06:48:58.742184 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-msjdt" Dec 06 06:48:58 crc kubenswrapper[4809]: I1206 06:48:58.748452 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-msjdt"] Dec 06 06:48:58 crc kubenswrapper[4809]: I1206 06:48:58.931382 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdbs5\" (UniqueName: \"kubernetes.io/projected/a442e1cb-3f09-4644-a666-db3dd7bce027-kube-api-access-xdbs5\") pod \"community-operators-msjdt\" (UID: \"a442e1cb-3f09-4644-a666-db3dd7bce027\") " pod="openshift-marketplace/community-operators-msjdt" Dec 06 06:48:58 crc kubenswrapper[4809]: I1206 06:48:58.931697 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a442e1cb-3f09-4644-a666-db3dd7bce027-utilities\") pod \"community-operators-msjdt\" (UID: \"a442e1cb-3f09-4644-a666-db3dd7bce027\") " pod="openshift-marketplace/community-operators-msjdt" Dec 06 06:48:58 crc kubenswrapper[4809]: I1206 06:48:58.931854 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a442e1cb-3f09-4644-a666-db3dd7bce027-catalog-content\") pod \"community-operators-msjdt\" (UID: \"a442e1cb-3f09-4644-a666-db3dd7bce027\") " pod="openshift-marketplace/community-operators-msjdt" Dec 06 06:48:59 crc kubenswrapper[4809]: I1206 06:48:59.033798 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a442e1cb-3f09-4644-a666-db3dd7bce027-utilities\") pod \"community-operators-msjdt\" (UID: \"a442e1cb-3f09-4644-a666-db3dd7bce027\") " pod="openshift-marketplace/community-operators-msjdt" Dec 06 06:48:59 crc kubenswrapper[4809]: I1206 06:48:59.033893 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a442e1cb-3f09-4644-a666-db3dd7bce027-catalog-content\") pod \"community-operators-msjdt\" (UID: \"a442e1cb-3f09-4644-a666-db3dd7bce027\") " pod="openshift-marketplace/community-operators-msjdt" Dec 06 06:48:59 crc kubenswrapper[4809]: I1206 06:48:59.034044 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdbs5\" (UniqueName: \"kubernetes.io/projected/a442e1cb-3f09-4644-a666-db3dd7bce027-kube-api-access-xdbs5\") pod \"community-operators-msjdt\" (UID: \"a442e1cb-3f09-4644-a666-db3dd7bce027\") " pod="openshift-marketplace/community-operators-msjdt" Dec 06 06:48:59 crc kubenswrapper[4809]: I1206 06:48:59.034479 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a442e1cb-3f09-4644-a666-db3dd7bce027-catalog-content\") pod \"community-operators-msjdt\" (UID: \"a442e1cb-3f09-4644-a666-db3dd7bce027\") " pod="openshift-marketplace/community-operators-msjdt" Dec 06 06:48:59 crc kubenswrapper[4809]: I1206 06:48:59.034495 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a442e1cb-3f09-4644-a666-db3dd7bce027-utilities\") pod \"community-operators-msjdt\" (UID: \"a442e1cb-3f09-4644-a666-db3dd7bce027\") " pod="openshift-marketplace/community-operators-msjdt" Dec 06 06:48:59 crc kubenswrapper[4809]: I1206 06:48:59.053756 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdbs5\" (UniqueName: \"kubernetes.io/projected/a442e1cb-3f09-4644-a666-db3dd7bce027-kube-api-access-xdbs5\") pod \"community-operators-msjdt\" (UID: \"a442e1cb-3f09-4644-a666-db3dd7bce027\") " pod="openshift-marketplace/community-operators-msjdt" Dec 06 06:48:59 crc kubenswrapper[4809]: I1206 06:48:59.074501 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-msjdt" Dec 06 06:48:59 crc kubenswrapper[4809]: I1206 06:48:59.619494 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-msjdt"] Dec 06 06:49:00 crc kubenswrapper[4809]: I1206 06:49:00.192181 4809 generic.go:334] "Generic (PLEG): container finished" podID="a442e1cb-3f09-4644-a666-db3dd7bce027" containerID="1014a9b8e4ad90b81dd44dc2115026f89b95172e68d7d44608f8ac4ba980a089" exitCode=0 Dec 06 06:49:00 crc kubenswrapper[4809]: I1206 06:49:00.192261 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-msjdt" event={"ID":"a442e1cb-3f09-4644-a666-db3dd7bce027","Type":"ContainerDied","Data":"1014a9b8e4ad90b81dd44dc2115026f89b95172e68d7d44608f8ac4ba980a089"} Dec 06 06:49:00 crc kubenswrapper[4809]: I1206 06:49:00.192314 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-msjdt" event={"ID":"a442e1cb-3f09-4644-a666-db3dd7bce027","Type":"ContainerStarted","Data":"ead72aaf9943fc3f7473ec516b20049605bc3f139875212e602e6f0abdaf7813"} Dec 06 06:49:05 crc kubenswrapper[4809]: I1206 06:49:05.252151 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-msjdt" event={"ID":"a442e1cb-3f09-4644-a666-db3dd7bce027","Type":"ContainerStarted","Data":"3c8a6e2159b0a295ea852c7555d48c401c081c47e49280a5a65b27a3ae642bec"} Dec 06 06:49:09 crc kubenswrapper[4809]: I1206 06:49:09.300681 4809 generic.go:334] "Generic (PLEG): container finished" podID="a442e1cb-3f09-4644-a666-db3dd7bce027" containerID="3c8a6e2159b0a295ea852c7555d48c401c081c47e49280a5a65b27a3ae642bec" exitCode=0 Dec 06 06:49:09 crc kubenswrapper[4809]: I1206 06:49:09.300735 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-msjdt" event={"ID":"a442e1cb-3f09-4644-a666-db3dd7bce027","Type":"ContainerDied","Data":"3c8a6e2159b0a295ea852c7555d48c401c081c47e49280a5a65b27a3ae642bec"} Dec 06 06:49:13 crc kubenswrapper[4809]: I1206 06:49:13.358082 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-msjdt" event={"ID":"a442e1cb-3f09-4644-a666-db3dd7bce027","Type":"ContainerStarted","Data":"f77359eff7b1deeadb98f0df5b9824ca1190184fbd26f9296f5d691ecd2613a5"} Dec 06 06:49:13 crc kubenswrapper[4809]: I1206 06:49:13.392015 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-msjdt" podStartSLOduration=2.6707770809999998 podStartE2EDuration="15.3919944s" podCreationTimestamp="2025-12-06 06:48:58 +0000 UTC" firstStartedPulling="2025-12-06 06:49:00.194500913 +0000 UTC m=+3465.083483855" lastFinishedPulling="2025-12-06 06:49:12.915718232 +0000 UTC m=+3477.804701174" observedRunningTime="2025-12-06 06:49:13.381478255 +0000 UTC m=+3478.270461207" watchObservedRunningTime="2025-12-06 06:49:13.3919944 +0000 UTC m=+3478.280977342" Dec 06 06:49:19 crc kubenswrapper[4809]: I1206 06:49:19.075262 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-msjdt" Dec 06 06:49:19 crc kubenswrapper[4809]: I1206 06:49:19.076970 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-msjdt" Dec 06 06:49:19 crc kubenswrapper[4809]: I1206 06:49:19.131883 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-msjdt" Dec 06 06:49:19 crc kubenswrapper[4809]: I1206 06:49:19.509723 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-msjdt" Dec 06 06:49:19 crc kubenswrapper[4809]: I1206 06:49:19.561963 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-msjdt"] Dec 06 06:49:21 crc kubenswrapper[4809]: I1206 06:49:21.465709 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-msjdt" podUID="a442e1cb-3f09-4644-a666-db3dd7bce027" containerName="registry-server" containerID="cri-o://f77359eff7b1deeadb98f0df5b9824ca1190184fbd26f9296f5d691ecd2613a5" gracePeriod=2 Dec 06 06:49:22 crc kubenswrapper[4809]: I1206 06:49:22.018465 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-msjdt" Dec 06 06:49:22 crc kubenswrapper[4809]: I1206 06:49:22.148147 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xdbs5\" (UniqueName: \"kubernetes.io/projected/a442e1cb-3f09-4644-a666-db3dd7bce027-kube-api-access-xdbs5\") pod \"a442e1cb-3f09-4644-a666-db3dd7bce027\" (UID: \"a442e1cb-3f09-4644-a666-db3dd7bce027\") " Dec 06 06:49:22 crc kubenswrapper[4809]: I1206 06:49:22.148277 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a442e1cb-3f09-4644-a666-db3dd7bce027-catalog-content\") pod \"a442e1cb-3f09-4644-a666-db3dd7bce027\" (UID: \"a442e1cb-3f09-4644-a666-db3dd7bce027\") " Dec 06 06:49:22 crc kubenswrapper[4809]: I1206 06:49:22.148558 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a442e1cb-3f09-4644-a666-db3dd7bce027-utilities\") pod \"a442e1cb-3f09-4644-a666-db3dd7bce027\" (UID: \"a442e1cb-3f09-4644-a666-db3dd7bce027\") " Dec 06 06:49:22 crc kubenswrapper[4809]: I1206 06:49:22.149344 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a442e1cb-3f09-4644-a666-db3dd7bce027-utilities" (OuterVolumeSpecName: "utilities") pod "a442e1cb-3f09-4644-a666-db3dd7bce027" (UID: "a442e1cb-3f09-4644-a666-db3dd7bce027"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:49:22 crc kubenswrapper[4809]: I1206 06:49:22.150280 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a442e1cb-3f09-4644-a666-db3dd7bce027-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:49:22 crc kubenswrapper[4809]: I1206 06:49:22.154454 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a442e1cb-3f09-4644-a666-db3dd7bce027-kube-api-access-xdbs5" (OuterVolumeSpecName: "kube-api-access-xdbs5") pod "a442e1cb-3f09-4644-a666-db3dd7bce027" (UID: "a442e1cb-3f09-4644-a666-db3dd7bce027"). InnerVolumeSpecName "kube-api-access-xdbs5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:49:22 crc kubenswrapper[4809]: I1206 06:49:22.203382 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a442e1cb-3f09-4644-a666-db3dd7bce027-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a442e1cb-3f09-4644-a666-db3dd7bce027" (UID: "a442e1cb-3f09-4644-a666-db3dd7bce027"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:49:22 crc kubenswrapper[4809]: I1206 06:49:22.251520 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a442e1cb-3f09-4644-a666-db3dd7bce027-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:49:22 crc kubenswrapper[4809]: I1206 06:49:22.251568 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xdbs5\" (UniqueName: \"kubernetes.io/projected/a442e1cb-3f09-4644-a666-db3dd7bce027-kube-api-access-xdbs5\") on node \"crc\" DevicePath \"\"" Dec 06 06:49:22 crc kubenswrapper[4809]: I1206 06:49:22.483287 4809 generic.go:334] "Generic (PLEG): container finished" podID="a442e1cb-3f09-4644-a666-db3dd7bce027" containerID="f77359eff7b1deeadb98f0df5b9824ca1190184fbd26f9296f5d691ecd2613a5" exitCode=0 Dec 06 06:49:22 crc kubenswrapper[4809]: I1206 06:49:22.483339 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-msjdt" event={"ID":"a442e1cb-3f09-4644-a666-db3dd7bce027","Type":"ContainerDied","Data":"f77359eff7b1deeadb98f0df5b9824ca1190184fbd26f9296f5d691ecd2613a5"} Dec 06 06:49:22 crc kubenswrapper[4809]: I1206 06:49:22.483372 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-msjdt" event={"ID":"a442e1cb-3f09-4644-a666-db3dd7bce027","Type":"ContainerDied","Data":"ead72aaf9943fc3f7473ec516b20049605bc3f139875212e602e6f0abdaf7813"} Dec 06 06:49:22 crc kubenswrapper[4809]: I1206 06:49:22.483391 4809 scope.go:117] "RemoveContainer" containerID="f77359eff7b1deeadb98f0df5b9824ca1190184fbd26f9296f5d691ecd2613a5" Dec 06 06:49:22 crc kubenswrapper[4809]: I1206 06:49:22.483565 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-msjdt" Dec 06 06:49:22 crc kubenswrapper[4809]: I1206 06:49:22.526568 4809 scope.go:117] "RemoveContainer" containerID="3c8a6e2159b0a295ea852c7555d48c401c081c47e49280a5a65b27a3ae642bec" Dec 06 06:49:22 crc kubenswrapper[4809]: I1206 06:49:22.537465 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-msjdt"] Dec 06 06:49:22 crc kubenswrapper[4809]: I1206 06:49:22.553127 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-msjdt"] Dec 06 06:49:22 crc kubenswrapper[4809]: I1206 06:49:22.556174 4809 scope.go:117] "RemoveContainer" containerID="1014a9b8e4ad90b81dd44dc2115026f89b95172e68d7d44608f8ac4ba980a089" Dec 06 06:49:22 crc kubenswrapper[4809]: I1206 06:49:22.606790 4809 scope.go:117] "RemoveContainer" containerID="f77359eff7b1deeadb98f0df5b9824ca1190184fbd26f9296f5d691ecd2613a5" Dec 06 06:49:22 crc kubenswrapper[4809]: E1206 06:49:22.607407 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f77359eff7b1deeadb98f0df5b9824ca1190184fbd26f9296f5d691ecd2613a5\": container with ID starting with f77359eff7b1deeadb98f0df5b9824ca1190184fbd26f9296f5d691ecd2613a5 not found: ID does not exist" containerID="f77359eff7b1deeadb98f0df5b9824ca1190184fbd26f9296f5d691ecd2613a5" Dec 06 06:49:22 crc kubenswrapper[4809]: I1206 06:49:22.607508 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f77359eff7b1deeadb98f0df5b9824ca1190184fbd26f9296f5d691ecd2613a5"} err="failed to get container status \"f77359eff7b1deeadb98f0df5b9824ca1190184fbd26f9296f5d691ecd2613a5\": rpc error: code = NotFound desc = could not find container \"f77359eff7b1deeadb98f0df5b9824ca1190184fbd26f9296f5d691ecd2613a5\": container with ID starting with f77359eff7b1deeadb98f0df5b9824ca1190184fbd26f9296f5d691ecd2613a5 not found: ID does not exist" Dec 06 06:49:22 crc kubenswrapper[4809]: I1206 06:49:22.607584 4809 scope.go:117] "RemoveContainer" containerID="3c8a6e2159b0a295ea852c7555d48c401c081c47e49280a5a65b27a3ae642bec" Dec 06 06:49:22 crc kubenswrapper[4809]: E1206 06:49:22.607974 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c8a6e2159b0a295ea852c7555d48c401c081c47e49280a5a65b27a3ae642bec\": container with ID starting with 3c8a6e2159b0a295ea852c7555d48c401c081c47e49280a5a65b27a3ae642bec not found: ID does not exist" containerID="3c8a6e2159b0a295ea852c7555d48c401c081c47e49280a5a65b27a3ae642bec" Dec 06 06:49:22 crc kubenswrapper[4809]: I1206 06:49:22.608079 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c8a6e2159b0a295ea852c7555d48c401c081c47e49280a5a65b27a3ae642bec"} err="failed to get container status \"3c8a6e2159b0a295ea852c7555d48c401c081c47e49280a5a65b27a3ae642bec\": rpc error: code = NotFound desc = could not find container \"3c8a6e2159b0a295ea852c7555d48c401c081c47e49280a5a65b27a3ae642bec\": container with ID starting with 3c8a6e2159b0a295ea852c7555d48c401c081c47e49280a5a65b27a3ae642bec not found: ID does not exist" Dec 06 06:49:22 crc kubenswrapper[4809]: I1206 06:49:22.608182 4809 scope.go:117] "RemoveContainer" containerID="1014a9b8e4ad90b81dd44dc2115026f89b95172e68d7d44608f8ac4ba980a089" Dec 06 06:49:22 crc kubenswrapper[4809]: E1206 06:49:22.608581 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1014a9b8e4ad90b81dd44dc2115026f89b95172e68d7d44608f8ac4ba980a089\": container with ID starting with 1014a9b8e4ad90b81dd44dc2115026f89b95172e68d7d44608f8ac4ba980a089 not found: ID does not exist" containerID="1014a9b8e4ad90b81dd44dc2115026f89b95172e68d7d44608f8ac4ba980a089" Dec 06 06:49:22 crc kubenswrapper[4809]: I1206 06:49:22.608626 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1014a9b8e4ad90b81dd44dc2115026f89b95172e68d7d44608f8ac4ba980a089"} err="failed to get container status \"1014a9b8e4ad90b81dd44dc2115026f89b95172e68d7d44608f8ac4ba980a089\": rpc error: code = NotFound desc = could not find container \"1014a9b8e4ad90b81dd44dc2115026f89b95172e68d7d44608f8ac4ba980a089\": container with ID starting with 1014a9b8e4ad90b81dd44dc2115026f89b95172e68d7d44608f8ac4ba980a089 not found: ID does not exist" Dec 06 06:49:23 crc kubenswrapper[4809]: I1206 06:49:23.403676 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a442e1cb-3f09-4644-a666-db3dd7bce027" path="/var/lib/kubelet/pods/a442e1cb-3f09-4644-a666-db3dd7bce027/volumes" Dec 06 06:49:34 crc kubenswrapper[4809]: I1206 06:49:34.496335 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:49:34 crc kubenswrapper[4809]: I1206 06:49:34.496957 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:50:04 crc kubenswrapper[4809]: I1206 06:50:04.496826 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:50:04 crc kubenswrapper[4809]: I1206 06:50:04.497340 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:50:33 crc kubenswrapper[4809]: I1206 06:50:33.267763 4809 scope.go:117] "RemoveContainer" containerID="d4311b3d67cfc79eb4d58ab21e01750238b2e65d997ded719f775ab7b3c7509f" Dec 06 06:50:33 crc kubenswrapper[4809]: I1206 06:50:33.290437 4809 scope.go:117] "RemoveContainer" containerID="f485d604b1baa4d876ec6396eaf010f8360d43ab96e36af8902d5fd3bb4bf7d9" Dec 06 06:50:33 crc kubenswrapper[4809]: I1206 06:50:33.374199 4809 scope.go:117] "RemoveContainer" containerID="e137ec7add300f3bb117dc43fbf30efd8c3cd0da1891a5963997d40da66f366b" Dec 06 06:50:34 crc kubenswrapper[4809]: I1206 06:50:34.496892 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:50:34 crc kubenswrapper[4809]: I1206 06:50:34.497274 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:50:34 crc kubenswrapper[4809]: I1206 06:50:34.497328 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-npms2" Dec 06 06:50:34 crc kubenswrapper[4809]: I1206 06:50:34.498330 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"083a8276457cb2db9f64bc581dbeeb57fcdb1feabe656331716e6ce96734d239"} pod="openshift-machine-config-operator/machine-config-daemon-npms2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 06:50:34 crc kubenswrapper[4809]: I1206 06:50:34.498401 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" containerID="cri-o://083a8276457cb2db9f64bc581dbeeb57fcdb1feabe656331716e6ce96734d239" gracePeriod=600 Dec 06 06:50:35 crc kubenswrapper[4809]: I1206 06:50:35.264674 4809 generic.go:334] "Generic (PLEG): container finished" podID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerID="083a8276457cb2db9f64bc581dbeeb57fcdb1feabe656331716e6ce96734d239" exitCode=0 Dec 06 06:50:35 crc kubenswrapper[4809]: I1206 06:50:35.264758 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" event={"ID":"cbf4cf62-024e-4703-a8b8-9aecda9cd26a","Type":"ContainerDied","Data":"083a8276457cb2db9f64bc581dbeeb57fcdb1feabe656331716e6ce96734d239"} Dec 06 06:50:35 crc kubenswrapper[4809]: I1206 06:50:35.265281 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" event={"ID":"cbf4cf62-024e-4703-a8b8-9aecda9cd26a","Type":"ContainerStarted","Data":"030a0d8b374914264276796b77b60278c64c06d896d2041127b7a240ba65b24c"} Dec 06 06:50:35 crc kubenswrapper[4809]: I1206 06:50:35.265310 4809 scope.go:117] "RemoveContainer" containerID="ca23ae36692bfc6cbf4be3666821469da7cf4aff660aa8bd1e9a06863c4d9e82" Dec 06 06:50:40 crc kubenswrapper[4809]: I1206 06:50:40.325602 4809 generic.go:334] "Generic (PLEG): container finished" podID="0a499909-59b6-42ec-a9f7-75218a100ead" containerID="24a8bd65bcf0b39b4ebbd77eac15afb4c3b67605f2d3c41ce441271fe3fe9ed9" exitCode=0 Dec 06 06:50:40 crc kubenswrapper[4809]: I1206 06:50:40.325697 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw" event={"ID":"0a499909-59b6-42ec-a9f7-75218a100ead","Type":"ContainerDied","Data":"24a8bd65bcf0b39b4ebbd77eac15afb4c3b67605f2d3c41ce441271fe3fe9ed9"} Dec 06 06:50:41 crc kubenswrapper[4809]: I1206 06:50:41.828127 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw" Dec 06 06:50:41 crc kubenswrapper[4809]: I1206 06:50:41.995513 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/0a499909-59b6-42ec-a9f7-75218a100ead-ceilometer-ipmi-config-data-2\") pod \"0a499909-59b6-42ec-a9f7-75218a100ead\" (UID: \"0a499909-59b6-42ec-a9f7-75218a100ead\") " Dec 06 06:50:41 crc kubenswrapper[4809]: I1206 06:50:41.995799 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0a499909-59b6-42ec-a9f7-75218a100ead-inventory\") pod \"0a499909-59b6-42ec-a9f7-75218a100ead\" (UID: \"0a499909-59b6-42ec-a9f7-75218a100ead\") " Dec 06 06:50:41 crc kubenswrapper[4809]: I1206 06:50:41.995972 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0a499909-59b6-42ec-a9f7-75218a100ead-ssh-key\") pod \"0a499909-59b6-42ec-a9f7-75218a100ead\" (UID: \"0a499909-59b6-42ec-a9f7-75218a100ead\") " Dec 06 06:50:41 crc kubenswrapper[4809]: I1206 06:50:41.996102 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/0a499909-59b6-42ec-a9f7-75218a100ead-ceilometer-ipmi-config-data-0\") pod \"0a499909-59b6-42ec-a9f7-75218a100ead\" (UID: \"0a499909-59b6-42ec-a9f7-75218a100ead\") " Dec 06 06:50:41 crc kubenswrapper[4809]: I1206 06:50:41.996319 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a499909-59b6-42ec-a9f7-75218a100ead-telemetry-power-monitoring-combined-ca-bundle\") pod \"0a499909-59b6-42ec-a9f7-75218a100ead\" (UID: \"0a499909-59b6-42ec-a9f7-75218a100ead\") " Dec 06 06:50:41 crc kubenswrapper[4809]: I1206 06:50:41.996504 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/0a499909-59b6-42ec-a9f7-75218a100ead-ceilometer-ipmi-config-data-1\") pod \"0a499909-59b6-42ec-a9f7-75218a100ead\" (UID: \"0a499909-59b6-42ec-a9f7-75218a100ead\") " Dec 06 06:50:41 crc kubenswrapper[4809]: I1206 06:50:41.996656 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhzsc\" (UniqueName: \"kubernetes.io/projected/0a499909-59b6-42ec-a9f7-75218a100ead-kube-api-access-xhzsc\") pod \"0a499909-59b6-42ec-a9f7-75218a100ead\" (UID: \"0a499909-59b6-42ec-a9f7-75218a100ead\") " Dec 06 06:50:42 crc kubenswrapper[4809]: I1206 06:50:42.002604 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a499909-59b6-42ec-a9f7-75218a100ead-kube-api-access-xhzsc" (OuterVolumeSpecName: "kube-api-access-xhzsc") pod "0a499909-59b6-42ec-a9f7-75218a100ead" (UID: "0a499909-59b6-42ec-a9f7-75218a100ead"). InnerVolumeSpecName "kube-api-access-xhzsc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:50:42 crc kubenswrapper[4809]: I1206 06:50:42.002739 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a499909-59b6-42ec-a9f7-75218a100ead-telemetry-power-monitoring-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-power-monitoring-combined-ca-bundle") pod "0a499909-59b6-42ec-a9f7-75218a100ead" (UID: "0a499909-59b6-42ec-a9f7-75218a100ead"). InnerVolumeSpecName "telemetry-power-monitoring-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:50:42 crc kubenswrapper[4809]: I1206 06:50:42.031502 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a499909-59b6-42ec-a9f7-75218a100ead-ceilometer-ipmi-config-data-2" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-2") pod "0a499909-59b6-42ec-a9f7-75218a100ead" (UID: "0a499909-59b6-42ec-a9f7-75218a100ead"). InnerVolumeSpecName "ceilometer-ipmi-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:50:42 crc kubenswrapper[4809]: I1206 06:50:42.033435 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a499909-59b6-42ec-a9f7-75218a100ead-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0a499909-59b6-42ec-a9f7-75218a100ead" (UID: "0a499909-59b6-42ec-a9f7-75218a100ead"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:50:42 crc kubenswrapper[4809]: I1206 06:50:42.033837 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a499909-59b6-42ec-a9f7-75218a100ead-ceilometer-ipmi-config-data-0" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-0") pod "0a499909-59b6-42ec-a9f7-75218a100ead" (UID: "0a499909-59b6-42ec-a9f7-75218a100ead"). InnerVolumeSpecName "ceilometer-ipmi-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:50:42 crc kubenswrapper[4809]: I1206 06:50:42.035183 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a499909-59b6-42ec-a9f7-75218a100ead-ceilometer-ipmi-config-data-1" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-1") pod "0a499909-59b6-42ec-a9f7-75218a100ead" (UID: "0a499909-59b6-42ec-a9f7-75218a100ead"). InnerVolumeSpecName "ceilometer-ipmi-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:50:42 crc kubenswrapper[4809]: I1206 06:50:42.043294 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a499909-59b6-42ec-a9f7-75218a100ead-inventory" (OuterVolumeSpecName: "inventory") pod "0a499909-59b6-42ec-a9f7-75218a100ead" (UID: "0a499909-59b6-42ec-a9f7-75218a100ead"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:50:42 crc kubenswrapper[4809]: I1206 06:50:42.107426 4809 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/0a499909-59b6-42ec-a9f7-75218a100ead-ceilometer-ipmi-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 06 06:50:42 crc kubenswrapper[4809]: I1206 06:50:42.107477 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhzsc\" (UniqueName: \"kubernetes.io/projected/0a499909-59b6-42ec-a9f7-75218a100ead-kube-api-access-xhzsc\") on node \"crc\" DevicePath \"\"" Dec 06 06:50:42 crc kubenswrapper[4809]: I1206 06:50:42.107488 4809 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/0a499909-59b6-42ec-a9f7-75218a100ead-ceilometer-ipmi-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 06 06:50:42 crc kubenswrapper[4809]: I1206 06:50:42.107500 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0a499909-59b6-42ec-a9f7-75218a100ead-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 06:50:42 crc kubenswrapper[4809]: I1206 06:50:42.107513 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0a499909-59b6-42ec-a9f7-75218a100ead-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 06:50:42 crc kubenswrapper[4809]: I1206 06:50:42.107532 4809 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/0a499909-59b6-42ec-a9f7-75218a100ead-ceilometer-ipmi-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:50:42 crc kubenswrapper[4809]: I1206 06:50:42.107550 4809 reconciler_common.go:293] "Volume detached for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a499909-59b6-42ec-a9f7-75218a100ead-telemetry-power-monitoring-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:50:42 crc kubenswrapper[4809]: I1206 06:50:42.349256 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw" event={"ID":"0a499909-59b6-42ec-a9f7-75218a100ead","Type":"ContainerDied","Data":"27bf1dbfd78ed4f30e5bfb237bb3b3295bc11dd15c395ea2066c50cf49eaff2e"} Dec 06 06:50:42 crc kubenswrapper[4809]: I1206 06:50:42.349662 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="27bf1dbfd78ed4f30e5bfb237bb3b3295bc11dd15c395ea2066c50cf49eaff2e" Dec 06 06:50:42 crc kubenswrapper[4809]: I1206 06:50:42.349298 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw" Dec 06 06:50:42 crc kubenswrapper[4809]: I1206 06:50:42.436686 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-crjgz"] Dec 06 06:50:42 crc kubenswrapper[4809]: E1206 06:50:42.437397 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a499909-59b6-42ec-a9f7-75218a100ead" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Dec 06 06:50:42 crc kubenswrapper[4809]: I1206 06:50:42.437529 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a499909-59b6-42ec-a9f7-75218a100ead" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Dec 06 06:50:42 crc kubenswrapper[4809]: E1206 06:50:42.437604 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a442e1cb-3f09-4644-a666-db3dd7bce027" containerName="extract-utilities" Dec 06 06:50:42 crc kubenswrapper[4809]: I1206 06:50:42.437655 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a442e1cb-3f09-4644-a666-db3dd7bce027" containerName="extract-utilities" Dec 06 06:50:42 crc kubenswrapper[4809]: E1206 06:50:42.437719 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a442e1cb-3f09-4644-a666-db3dd7bce027" containerName="registry-server" Dec 06 06:50:42 crc kubenswrapper[4809]: I1206 06:50:42.437766 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a442e1cb-3f09-4644-a666-db3dd7bce027" containerName="registry-server" Dec 06 06:50:42 crc kubenswrapper[4809]: E1206 06:50:42.437826 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a442e1cb-3f09-4644-a666-db3dd7bce027" containerName="extract-content" Dec 06 06:50:42 crc kubenswrapper[4809]: I1206 06:50:42.437878 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a442e1cb-3f09-4644-a666-db3dd7bce027" containerName="extract-content" Dec 06 06:50:42 crc kubenswrapper[4809]: I1206 06:50:42.438141 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="a442e1cb-3f09-4644-a666-db3dd7bce027" containerName="registry-server" Dec 06 06:50:42 crc kubenswrapper[4809]: I1206 06:50:42.438220 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a499909-59b6-42ec-a9f7-75218a100ead" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Dec 06 06:50:42 crc kubenswrapper[4809]: I1206 06:50:42.439102 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-crjgz" Dec 06 06:50:42 crc kubenswrapper[4809]: I1206 06:50:42.443120 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-bzlf5" Dec 06 06:50:42 crc kubenswrapper[4809]: I1206 06:50:42.443120 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 06:50:42 crc kubenswrapper[4809]: I1206 06:50:42.443637 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"logging-compute-config-data" Dec 06 06:50:42 crc kubenswrapper[4809]: I1206 06:50:42.443814 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 06:50:42 crc kubenswrapper[4809]: I1206 06:50:42.444005 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 06:50:42 crc kubenswrapper[4809]: I1206 06:50:42.460343 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-crjgz"] Dec 06 06:50:42 crc kubenswrapper[4809]: I1206 06:50:42.518365 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45v7s\" (UniqueName: \"kubernetes.io/projected/756d1854-be5a-421f-b620-ef10b83e7a61-kube-api-access-45v7s\") pod \"logging-edpm-deployment-openstack-edpm-ipam-crjgz\" (UID: \"756d1854-be5a-421f-b620-ef10b83e7a61\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-crjgz" Dec 06 06:50:42 crc kubenswrapper[4809]: I1206 06:50:42.518452 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/756d1854-be5a-421f-b620-ef10b83e7a61-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-crjgz\" (UID: \"756d1854-be5a-421f-b620-ef10b83e7a61\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-crjgz" Dec 06 06:50:42 crc kubenswrapper[4809]: I1206 06:50:42.518500 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/756d1854-be5a-421f-b620-ef10b83e7a61-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-crjgz\" (UID: \"756d1854-be5a-421f-b620-ef10b83e7a61\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-crjgz" Dec 06 06:50:42 crc kubenswrapper[4809]: I1206 06:50:42.518566 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/756d1854-be5a-421f-b620-ef10b83e7a61-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-crjgz\" (UID: \"756d1854-be5a-421f-b620-ef10b83e7a61\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-crjgz" Dec 06 06:50:42 crc kubenswrapper[4809]: I1206 06:50:42.518586 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/756d1854-be5a-421f-b620-ef10b83e7a61-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-crjgz\" (UID: \"756d1854-be5a-421f-b620-ef10b83e7a61\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-crjgz" Dec 06 06:50:42 crc kubenswrapper[4809]: I1206 06:50:42.620820 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45v7s\" (UniqueName: \"kubernetes.io/projected/756d1854-be5a-421f-b620-ef10b83e7a61-kube-api-access-45v7s\") pod \"logging-edpm-deployment-openstack-edpm-ipam-crjgz\" (UID: \"756d1854-be5a-421f-b620-ef10b83e7a61\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-crjgz" Dec 06 06:50:42 crc kubenswrapper[4809]: I1206 06:50:42.620979 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/756d1854-be5a-421f-b620-ef10b83e7a61-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-crjgz\" (UID: \"756d1854-be5a-421f-b620-ef10b83e7a61\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-crjgz" Dec 06 06:50:42 crc kubenswrapper[4809]: I1206 06:50:42.621058 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/756d1854-be5a-421f-b620-ef10b83e7a61-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-crjgz\" (UID: \"756d1854-be5a-421f-b620-ef10b83e7a61\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-crjgz" Dec 06 06:50:42 crc kubenswrapper[4809]: I1206 06:50:42.621157 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/756d1854-be5a-421f-b620-ef10b83e7a61-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-crjgz\" (UID: \"756d1854-be5a-421f-b620-ef10b83e7a61\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-crjgz" Dec 06 06:50:42 crc kubenswrapper[4809]: I1206 06:50:42.621190 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/756d1854-be5a-421f-b620-ef10b83e7a61-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-crjgz\" (UID: \"756d1854-be5a-421f-b620-ef10b83e7a61\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-crjgz" Dec 06 06:50:42 crc kubenswrapper[4809]: I1206 06:50:42.625426 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/756d1854-be5a-421f-b620-ef10b83e7a61-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-crjgz\" (UID: \"756d1854-be5a-421f-b620-ef10b83e7a61\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-crjgz" Dec 06 06:50:42 crc kubenswrapper[4809]: I1206 06:50:42.625813 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/756d1854-be5a-421f-b620-ef10b83e7a61-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-crjgz\" (UID: \"756d1854-be5a-421f-b620-ef10b83e7a61\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-crjgz" Dec 06 06:50:42 crc kubenswrapper[4809]: I1206 06:50:42.626438 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/756d1854-be5a-421f-b620-ef10b83e7a61-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-crjgz\" (UID: \"756d1854-be5a-421f-b620-ef10b83e7a61\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-crjgz" Dec 06 06:50:42 crc kubenswrapper[4809]: I1206 06:50:42.630379 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/756d1854-be5a-421f-b620-ef10b83e7a61-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-crjgz\" (UID: \"756d1854-be5a-421f-b620-ef10b83e7a61\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-crjgz" Dec 06 06:50:42 crc kubenswrapper[4809]: I1206 06:50:42.644304 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45v7s\" (UniqueName: \"kubernetes.io/projected/756d1854-be5a-421f-b620-ef10b83e7a61-kube-api-access-45v7s\") pod \"logging-edpm-deployment-openstack-edpm-ipam-crjgz\" (UID: \"756d1854-be5a-421f-b620-ef10b83e7a61\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-crjgz" Dec 06 06:50:42 crc kubenswrapper[4809]: I1206 06:50:42.758731 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-crjgz" Dec 06 06:50:43 crc kubenswrapper[4809]: I1206 06:50:43.359198 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-crjgz"] Dec 06 06:50:43 crc kubenswrapper[4809]: W1206 06:50:43.360972 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod756d1854_be5a_421f_b620_ef10b83e7a61.slice/crio-a7611b8b14ad53a8f0663c1005f0660bf8e405d542542883f2be4e343460eff3 WatchSource:0}: Error finding container a7611b8b14ad53a8f0663c1005f0660bf8e405d542542883f2be4e343460eff3: Status 404 returned error can't find the container with id a7611b8b14ad53a8f0663c1005f0660bf8e405d542542883f2be4e343460eff3 Dec 06 06:50:44 crc kubenswrapper[4809]: I1206 06:50:44.381981 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-crjgz" event={"ID":"756d1854-be5a-421f-b620-ef10b83e7a61","Type":"ContainerStarted","Data":"a7611b8b14ad53a8f0663c1005f0660bf8e405d542542883f2be4e343460eff3"} Dec 06 06:50:45 crc kubenswrapper[4809]: I1206 06:50:45.414150 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-crjgz" event={"ID":"756d1854-be5a-421f-b620-ef10b83e7a61","Type":"ContainerStarted","Data":"36db1eea89b94867800b46ad0c0bdfdb7257f02affc596c59e1ade8b00c5f5e7"} Dec 06 06:50:45 crc kubenswrapper[4809]: I1206 06:50:45.447855 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-crjgz" podStartSLOduration=2.443702476 podStartE2EDuration="3.4478302s" podCreationTimestamp="2025-12-06 06:50:42 +0000 UTC" firstStartedPulling="2025-12-06 06:50:43.363316136 +0000 UTC m=+3568.252299078" lastFinishedPulling="2025-12-06 06:50:44.36744384 +0000 UTC m=+3569.256426802" observedRunningTime="2025-12-06 06:50:45.440641475 +0000 UTC m=+3570.329624447" watchObservedRunningTime="2025-12-06 06:50:45.4478302 +0000 UTC m=+3570.336813142" Dec 06 06:51:01 crc kubenswrapper[4809]: I1206 06:51:01.619631 4809 generic.go:334] "Generic (PLEG): container finished" podID="756d1854-be5a-421f-b620-ef10b83e7a61" containerID="36db1eea89b94867800b46ad0c0bdfdb7257f02affc596c59e1ade8b00c5f5e7" exitCode=0 Dec 06 06:51:01 crc kubenswrapper[4809]: I1206 06:51:01.619751 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-crjgz" event={"ID":"756d1854-be5a-421f-b620-ef10b83e7a61","Type":"ContainerDied","Data":"36db1eea89b94867800b46ad0c0bdfdb7257f02affc596c59e1ade8b00c5f5e7"} Dec 06 06:51:03 crc kubenswrapper[4809]: I1206 06:51:03.184785 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-crjgz" Dec 06 06:51:03 crc kubenswrapper[4809]: I1206 06:51:03.369871 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-45v7s\" (UniqueName: \"kubernetes.io/projected/756d1854-be5a-421f-b620-ef10b83e7a61-kube-api-access-45v7s\") pod \"756d1854-be5a-421f-b620-ef10b83e7a61\" (UID: \"756d1854-be5a-421f-b620-ef10b83e7a61\") " Dec 06 06:51:03 crc kubenswrapper[4809]: I1206 06:51:03.369984 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/756d1854-be5a-421f-b620-ef10b83e7a61-logging-compute-config-data-1\") pod \"756d1854-be5a-421f-b620-ef10b83e7a61\" (UID: \"756d1854-be5a-421f-b620-ef10b83e7a61\") " Dec 06 06:51:03 crc kubenswrapper[4809]: I1206 06:51:03.370049 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/756d1854-be5a-421f-b620-ef10b83e7a61-ssh-key\") pod \"756d1854-be5a-421f-b620-ef10b83e7a61\" (UID: \"756d1854-be5a-421f-b620-ef10b83e7a61\") " Dec 06 06:51:03 crc kubenswrapper[4809]: I1206 06:51:03.370071 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/756d1854-be5a-421f-b620-ef10b83e7a61-inventory\") pod \"756d1854-be5a-421f-b620-ef10b83e7a61\" (UID: \"756d1854-be5a-421f-b620-ef10b83e7a61\") " Dec 06 06:51:03 crc kubenswrapper[4809]: I1206 06:51:03.370148 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/756d1854-be5a-421f-b620-ef10b83e7a61-logging-compute-config-data-0\") pod \"756d1854-be5a-421f-b620-ef10b83e7a61\" (UID: \"756d1854-be5a-421f-b620-ef10b83e7a61\") " Dec 06 06:51:03 crc kubenswrapper[4809]: I1206 06:51:03.419430 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/756d1854-be5a-421f-b620-ef10b83e7a61-kube-api-access-45v7s" (OuterVolumeSpecName: "kube-api-access-45v7s") pod "756d1854-be5a-421f-b620-ef10b83e7a61" (UID: "756d1854-be5a-421f-b620-ef10b83e7a61"). InnerVolumeSpecName "kube-api-access-45v7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:51:03 crc kubenswrapper[4809]: I1206 06:51:03.450796 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/756d1854-be5a-421f-b620-ef10b83e7a61-logging-compute-config-data-0" (OuterVolumeSpecName: "logging-compute-config-data-0") pod "756d1854-be5a-421f-b620-ef10b83e7a61" (UID: "756d1854-be5a-421f-b620-ef10b83e7a61"). InnerVolumeSpecName "logging-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:51:03 crc kubenswrapper[4809]: I1206 06:51:03.473321 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-45v7s\" (UniqueName: \"kubernetes.io/projected/756d1854-be5a-421f-b620-ef10b83e7a61-kube-api-access-45v7s\") on node \"crc\" DevicePath \"\"" Dec 06 06:51:03 crc kubenswrapper[4809]: I1206 06:51:03.473367 4809 reconciler_common.go:293] "Volume detached for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/756d1854-be5a-421f-b620-ef10b83e7a61-logging-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:51:03 crc kubenswrapper[4809]: I1206 06:51:03.471278 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/756d1854-be5a-421f-b620-ef10b83e7a61-logging-compute-config-data-1" (OuterVolumeSpecName: "logging-compute-config-data-1") pod "756d1854-be5a-421f-b620-ef10b83e7a61" (UID: "756d1854-be5a-421f-b620-ef10b83e7a61"). InnerVolumeSpecName "logging-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:51:03 crc kubenswrapper[4809]: I1206 06:51:03.487130 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/756d1854-be5a-421f-b620-ef10b83e7a61-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "756d1854-be5a-421f-b620-ef10b83e7a61" (UID: "756d1854-be5a-421f-b620-ef10b83e7a61"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:51:03 crc kubenswrapper[4809]: I1206 06:51:03.489852 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/756d1854-be5a-421f-b620-ef10b83e7a61-inventory" (OuterVolumeSpecName: "inventory") pod "756d1854-be5a-421f-b620-ef10b83e7a61" (UID: "756d1854-be5a-421f-b620-ef10b83e7a61"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:51:03 crc kubenswrapper[4809]: I1206 06:51:03.583282 4809 reconciler_common.go:293] "Volume detached for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/756d1854-be5a-421f-b620-ef10b83e7a61-logging-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 06 06:51:03 crc kubenswrapper[4809]: I1206 06:51:03.583600 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/756d1854-be5a-421f-b620-ef10b83e7a61-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 06:51:03 crc kubenswrapper[4809]: I1206 06:51:03.583613 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/756d1854-be5a-421f-b620-ef10b83e7a61-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 06:51:03 crc kubenswrapper[4809]: I1206 06:51:03.645299 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-crjgz" event={"ID":"756d1854-be5a-421f-b620-ef10b83e7a61","Type":"ContainerDied","Data":"a7611b8b14ad53a8f0663c1005f0660bf8e405d542542883f2be4e343460eff3"} Dec 06 06:51:03 crc kubenswrapper[4809]: I1206 06:51:03.645337 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7611b8b14ad53a8f0663c1005f0660bf8e405d542542883f2be4e343460eff3" Dec 06 06:51:03 crc kubenswrapper[4809]: I1206 06:51:03.645396 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-crjgz" Dec 06 06:52:27 crc kubenswrapper[4809]: I1206 06:52:27.017386 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kd6pj"] Dec 06 06:52:27 crc kubenswrapper[4809]: E1206 06:52:27.018857 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="756d1854-be5a-421f-b620-ef10b83e7a61" containerName="logging-edpm-deployment-openstack-edpm-ipam" Dec 06 06:52:27 crc kubenswrapper[4809]: I1206 06:52:27.018882 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="756d1854-be5a-421f-b620-ef10b83e7a61" containerName="logging-edpm-deployment-openstack-edpm-ipam" Dec 06 06:52:27 crc kubenswrapper[4809]: I1206 06:52:27.026700 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="756d1854-be5a-421f-b620-ef10b83e7a61" containerName="logging-edpm-deployment-openstack-edpm-ipam" Dec 06 06:52:27 crc kubenswrapper[4809]: I1206 06:52:27.029148 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kd6pj" Dec 06 06:52:27 crc kubenswrapper[4809]: I1206 06:52:27.032004 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kd6pj"] Dec 06 06:52:27 crc kubenswrapper[4809]: I1206 06:52:27.097497 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd7ad127-708d-4846-8b67-b14323c4de29-utilities\") pod \"redhat-marketplace-kd6pj\" (UID: \"cd7ad127-708d-4846-8b67-b14323c4de29\") " pod="openshift-marketplace/redhat-marketplace-kd6pj" Dec 06 06:52:27 crc kubenswrapper[4809]: I1206 06:52:27.097680 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd7ad127-708d-4846-8b67-b14323c4de29-catalog-content\") pod \"redhat-marketplace-kd6pj\" (UID: \"cd7ad127-708d-4846-8b67-b14323c4de29\") " pod="openshift-marketplace/redhat-marketplace-kd6pj" Dec 06 06:52:27 crc kubenswrapper[4809]: I1206 06:52:27.097776 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgpcl\" (UniqueName: \"kubernetes.io/projected/cd7ad127-708d-4846-8b67-b14323c4de29-kube-api-access-rgpcl\") pod \"redhat-marketplace-kd6pj\" (UID: \"cd7ad127-708d-4846-8b67-b14323c4de29\") " pod="openshift-marketplace/redhat-marketplace-kd6pj" Dec 06 06:52:27 crc kubenswrapper[4809]: I1206 06:52:27.200691 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd7ad127-708d-4846-8b67-b14323c4de29-utilities\") pod \"redhat-marketplace-kd6pj\" (UID: \"cd7ad127-708d-4846-8b67-b14323c4de29\") " pod="openshift-marketplace/redhat-marketplace-kd6pj" Dec 06 06:52:27 crc kubenswrapper[4809]: I1206 06:52:27.201212 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd7ad127-708d-4846-8b67-b14323c4de29-catalog-content\") pod \"redhat-marketplace-kd6pj\" (UID: \"cd7ad127-708d-4846-8b67-b14323c4de29\") " pod="openshift-marketplace/redhat-marketplace-kd6pj" Dec 06 06:52:27 crc kubenswrapper[4809]: I1206 06:52:27.201302 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgpcl\" (UniqueName: \"kubernetes.io/projected/cd7ad127-708d-4846-8b67-b14323c4de29-kube-api-access-rgpcl\") pod \"redhat-marketplace-kd6pj\" (UID: \"cd7ad127-708d-4846-8b67-b14323c4de29\") " pod="openshift-marketplace/redhat-marketplace-kd6pj" Dec 06 06:52:27 crc kubenswrapper[4809]: I1206 06:52:27.201499 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd7ad127-708d-4846-8b67-b14323c4de29-utilities\") pod \"redhat-marketplace-kd6pj\" (UID: \"cd7ad127-708d-4846-8b67-b14323c4de29\") " pod="openshift-marketplace/redhat-marketplace-kd6pj" Dec 06 06:52:27 crc kubenswrapper[4809]: I1206 06:52:27.201843 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd7ad127-708d-4846-8b67-b14323c4de29-catalog-content\") pod \"redhat-marketplace-kd6pj\" (UID: \"cd7ad127-708d-4846-8b67-b14323c4de29\") " pod="openshift-marketplace/redhat-marketplace-kd6pj" Dec 06 06:52:27 crc kubenswrapper[4809]: I1206 06:52:27.225193 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgpcl\" (UniqueName: \"kubernetes.io/projected/cd7ad127-708d-4846-8b67-b14323c4de29-kube-api-access-rgpcl\") pod \"redhat-marketplace-kd6pj\" (UID: \"cd7ad127-708d-4846-8b67-b14323c4de29\") " pod="openshift-marketplace/redhat-marketplace-kd6pj" Dec 06 06:52:27 crc kubenswrapper[4809]: I1206 06:52:27.358997 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kd6pj" Dec 06 06:52:27 crc kubenswrapper[4809]: I1206 06:52:27.903309 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kd6pj"] Dec 06 06:52:28 crc kubenswrapper[4809]: I1206 06:52:28.658205 4809 generic.go:334] "Generic (PLEG): container finished" podID="cd7ad127-708d-4846-8b67-b14323c4de29" containerID="cc07f102bc6ea1a48ba19d4361c37da1dc26dcdd02bf40a26ee0845ce5f213bc" exitCode=0 Dec 06 06:52:28 crc kubenswrapper[4809]: I1206 06:52:28.658321 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kd6pj" event={"ID":"cd7ad127-708d-4846-8b67-b14323c4de29","Type":"ContainerDied","Data":"cc07f102bc6ea1a48ba19d4361c37da1dc26dcdd02bf40a26ee0845ce5f213bc"} Dec 06 06:52:28 crc kubenswrapper[4809]: I1206 06:52:28.659083 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kd6pj" event={"ID":"cd7ad127-708d-4846-8b67-b14323c4de29","Type":"ContainerStarted","Data":"ba538dfaa68fe03e079801dd28e21afc39ae3b7a7e151e2498798ed40f99cd32"} Dec 06 06:52:28 crc kubenswrapper[4809]: I1206 06:52:28.661241 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 06:52:29 crc kubenswrapper[4809]: I1206 06:52:29.670972 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kd6pj" event={"ID":"cd7ad127-708d-4846-8b67-b14323c4de29","Type":"ContainerStarted","Data":"f40360bc8ec962053a0d0586558ef9ec254d1f076358b0e03fd0dbcdebd10d49"} Dec 06 06:52:30 crc kubenswrapper[4809]: E1206 06:52:30.902114 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcd7ad127_708d_4846_8b67_b14323c4de29.slice/crio-conmon-f40360bc8ec962053a0d0586558ef9ec254d1f076358b0e03fd0dbcdebd10d49.scope\": RecentStats: unable to find data in memory cache]" Dec 06 06:52:31 crc kubenswrapper[4809]: I1206 06:52:31.708065 4809 generic.go:334] "Generic (PLEG): container finished" podID="cd7ad127-708d-4846-8b67-b14323c4de29" containerID="f40360bc8ec962053a0d0586558ef9ec254d1f076358b0e03fd0dbcdebd10d49" exitCode=0 Dec 06 06:52:31 crc kubenswrapper[4809]: I1206 06:52:31.708108 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kd6pj" event={"ID":"cd7ad127-708d-4846-8b67-b14323c4de29","Type":"ContainerDied","Data":"f40360bc8ec962053a0d0586558ef9ec254d1f076358b0e03fd0dbcdebd10d49"} Dec 06 06:52:32 crc kubenswrapper[4809]: I1206 06:52:32.722599 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kd6pj" event={"ID":"cd7ad127-708d-4846-8b67-b14323c4de29","Type":"ContainerStarted","Data":"70e727fbaf1db14b208c41ecd1c52b16be3c574e45abaaa829062402f3aefbc8"} Dec 06 06:52:32 crc kubenswrapper[4809]: I1206 06:52:32.753292 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kd6pj" podStartSLOduration=3.26297164 podStartE2EDuration="6.753259904s" podCreationTimestamp="2025-12-06 06:52:26 +0000 UTC" firstStartedPulling="2025-12-06 06:52:28.661017488 +0000 UTC m=+3673.550000430" lastFinishedPulling="2025-12-06 06:52:32.151305752 +0000 UTC m=+3677.040288694" observedRunningTime="2025-12-06 06:52:32.74426001 +0000 UTC m=+3677.633242952" watchObservedRunningTime="2025-12-06 06:52:32.753259904 +0000 UTC m=+3677.642242846" Dec 06 06:52:34 crc kubenswrapper[4809]: I1206 06:52:34.496478 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:52:34 crc kubenswrapper[4809]: I1206 06:52:34.496793 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:52:37 crc kubenswrapper[4809]: I1206 06:52:37.359501 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kd6pj" Dec 06 06:52:37 crc kubenswrapper[4809]: I1206 06:52:37.360225 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kd6pj" Dec 06 06:52:37 crc kubenswrapper[4809]: I1206 06:52:37.424666 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kd6pj" Dec 06 06:52:37 crc kubenswrapper[4809]: I1206 06:52:37.856802 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kd6pj" Dec 06 06:52:38 crc kubenswrapper[4809]: I1206 06:52:38.589345 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kd6pj"] Dec 06 06:52:39 crc kubenswrapper[4809]: I1206 06:52:39.797440 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kd6pj" podUID="cd7ad127-708d-4846-8b67-b14323c4de29" containerName="registry-server" containerID="cri-o://70e727fbaf1db14b208c41ecd1c52b16be3c574e45abaaa829062402f3aefbc8" gracePeriod=2 Dec 06 06:52:40 crc kubenswrapper[4809]: I1206 06:52:40.344344 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kd6pj" Dec 06 06:52:40 crc kubenswrapper[4809]: I1206 06:52:40.521105 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd7ad127-708d-4846-8b67-b14323c4de29-catalog-content\") pod \"cd7ad127-708d-4846-8b67-b14323c4de29\" (UID: \"cd7ad127-708d-4846-8b67-b14323c4de29\") " Dec 06 06:52:40 crc kubenswrapper[4809]: I1206 06:52:40.522419 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd7ad127-708d-4846-8b67-b14323c4de29-utilities\") pod \"cd7ad127-708d-4846-8b67-b14323c4de29\" (UID: \"cd7ad127-708d-4846-8b67-b14323c4de29\") " Dec 06 06:52:40 crc kubenswrapper[4809]: I1206 06:52:40.522519 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rgpcl\" (UniqueName: \"kubernetes.io/projected/cd7ad127-708d-4846-8b67-b14323c4de29-kube-api-access-rgpcl\") pod \"cd7ad127-708d-4846-8b67-b14323c4de29\" (UID: \"cd7ad127-708d-4846-8b67-b14323c4de29\") " Dec 06 06:52:40 crc kubenswrapper[4809]: I1206 06:52:40.523019 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd7ad127-708d-4846-8b67-b14323c4de29-utilities" (OuterVolumeSpecName: "utilities") pod "cd7ad127-708d-4846-8b67-b14323c4de29" (UID: "cd7ad127-708d-4846-8b67-b14323c4de29"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:52:40 crc kubenswrapper[4809]: I1206 06:52:40.525610 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd7ad127-708d-4846-8b67-b14323c4de29-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:52:40 crc kubenswrapper[4809]: I1206 06:52:40.529242 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd7ad127-708d-4846-8b67-b14323c4de29-kube-api-access-rgpcl" (OuterVolumeSpecName: "kube-api-access-rgpcl") pod "cd7ad127-708d-4846-8b67-b14323c4de29" (UID: "cd7ad127-708d-4846-8b67-b14323c4de29"). InnerVolumeSpecName "kube-api-access-rgpcl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:52:40 crc kubenswrapper[4809]: I1206 06:52:40.552332 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd7ad127-708d-4846-8b67-b14323c4de29-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cd7ad127-708d-4846-8b67-b14323c4de29" (UID: "cd7ad127-708d-4846-8b67-b14323c4de29"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:52:40 crc kubenswrapper[4809]: I1206 06:52:40.628169 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd7ad127-708d-4846-8b67-b14323c4de29-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:52:40 crc kubenswrapper[4809]: I1206 06:52:40.628206 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rgpcl\" (UniqueName: \"kubernetes.io/projected/cd7ad127-708d-4846-8b67-b14323c4de29-kube-api-access-rgpcl\") on node \"crc\" DevicePath \"\"" Dec 06 06:52:40 crc kubenswrapper[4809]: I1206 06:52:40.812754 4809 generic.go:334] "Generic (PLEG): container finished" podID="cd7ad127-708d-4846-8b67-b14323c4de29" containerID="70e727fbaf1db14b208c41ecd1c52b16be3c574e45abaaa829062402f3aefbc8" exitCode=0 Dec 06 06:52:40 crc kubenswrapper[4809]: I1206 06:52:40.812816 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kd6pj" Dec 06 06:52:40 crc kubenswrapper[4809]: I1206 06:52:40.812835 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kd6pj" event={"ID":"cd7ad127-708d-4846-8b67-b14323c4de29","Type":"ContainerDied","Data":"70e727fbaf1db14b208c41ecd1c52b16be3c574e45abaaa829062402f3aefbc8"} Dec 06 06:52:40 crc kubenswrapper[4809]: I1206 06:52:40.813235 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kd6pj" event={"ID":"cd7ad127-708d-4846-8b67-b14323c4de29","Type":"ContainerDied","Data":"ba538dfaa68fe03e079801dd28e21afc39ae3b7a7e151e2498798ed40f99cd32"} Dec 06 06:52:40 crc kubenswrapper[4809]: I1206 06:52:40.813260 4809 scope.go:117] "RemoveContainer" containerID="70e727fbaf1db14b208c41ecd1c52b16be3c574e45abaaa829062402f3aefbc8" Dec 06 06:52:40 crc kubenswrapper[4809]: I1206 06:52:40.837707 4809 scope.go:117] "RemoveContainer" containerID="f40360bc8ec962053a0d0586558ef9ec254d1f076358b0e03fd0dbcdebd10d49" Dec 06 06:52:40 crc kubenswrapper[4809]: I1206 06:52:40.863750 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kd6pj"] Dec 06 06:52:40 crc kubenswrapper[4809]: I1206 06:52:40.873725 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kd6pj"] Dec 06 06:52:40 crc kubenswrapper[4809]: I1206 06:52:40.886242 4809 scope.go:117] "RemoveContainer" containerID="cc07f102bc6ea1a48ba19d4361c37da1dc26dcdd02bf40a26ee0845ce5f213bc" Dec 06 06:52:40 crc kubenswrapper[4809]: I1206 06:52:40.934726 4809 scope.go:117] "RemoveContainer" containerID="70e727fbaf1db14b208c41ecd1c52b16be3c574e45abaaa829062402f3aefbc8" Dec 06 06:52:40 crc kubenswrapper[4809]: E1206 06:52:40.935401 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70e727fbaf1db14b208c41ecd1c52b16be3c574e45abaaa829062402f3aefbc8\": container with ID starting with 70e727fbaf1db14b208c41ecd1c52b16be3c574e45abaaa829062402f3aefbc8 not found: ID does not exist" containerID="70e727fbaf1db14b208c41ecd1c52b16be3c574e45abaaa829062402f3aefbc8" Dec 06 06:52:40 crc kubenswrapper[4809]: I1206 06:52:40.935442 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70e727fbaf1db14b208c41ecd1c52b16be3c574e45abaaa829062402f3aefbc8"} err="failed to get container status \"70e727fbaf1db14b208c41ecd1c52b16be3c574e45abaaa829062402f3aefbc8\": rpc error: code = NotFound desc = could not find container \"70e727fbaf1db14b208c41ecd1c52b16be3c574e45abaaa829062402f3aefbc8\": container with ID starting with 70e727fbaf1db14b208c41ecd1c52b16be3c574e45abaaa829062402f3aefbc8 not found: ID does not exist" Dec 06 06:52:40 crc kubenswrapper[4809]: I1206 06:52:40.935468 4809 scope.go:117] "RemoveContainer" containerID="f40360bc8ec962053a0d0586558ef9ec254d1f076358b0e03fd0dbcdebd10d49" Dec 06 06:52:40 crc kubenswrapper[4809]: E1206 06:52:40.935835 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f40360bc8ec962053a0d0586558ef9ec254d1f076358b0e03fd0dbcdebd10d49\": container with ID starting with f40360bc8ec962053a0d0586558ef9ec254d1f076358b0e03fd0dbcdebd10d49 not found: ID does not exist" containerID="f40360bc8ec962053a0d0586558ef9ec254d1f076358b0e03fd0dbcdebd10d49" Dec 06 06:52:40 crc kubenswrapper[4809]: I1206 06:52:40.935864 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f40360bc8ec962053a0d0586558ef9ec254d1f076358b0e03fd0dbcdebd10d49"} err="failed to get container status \"f40360bc8ec962053a0d0586558ef9ec254d1f076358b0e03fd0dbcdebd10d49\": rpc error: code = NotFound desc = could not find container \"f40360bc8ec962053a0d0586558ef9ec254d1f076358b0e03fd0dbcdebd10d49\": container with ID starting with f40360bc8ec962053a0d0586558ef9ec254d1f076358b0e03fd0dbcdebd10d49 not found: ID does not exist" Dec 06 06:52:40 crc kubenswrapper[4809]: I1206 06:52:40.935883 4809 scope.go:117] "RemoveContainer" containerID="cc07f102bc6ea1a48ba19d4361c37da1dc26dcdd02bf40a26ee0845ce5f213bc" Dec 06 06:52:40 crc kubenswrapper[4809]: E1206 06:52:40.936115 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc07f102bc6ea1a48ba19d4361c37da1dc26dcdd02bf40a26ee0845ce5f213bc\": container with ID starting with cc07f102bc6ea1a48ba19d4361c37da1dc26dcdd02bf40a26ee0845ce5f213bc not found: ID does not exist" containerID="cc07f102bc6ea1a48ba19d4361c37da1dc26dcdd02bf40a26ee0845ce5f213bc" Dec 06 06:52:40 crc kubenswrapper[4809]: I1206 06:52:40.936145 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc07f102bc6ea1a48ba19d4361c37da1dc26dcdd02bf40a26ee0845ce5f213bc"} err="failed to get container status \"cc07f102bc6ea1a48ba19d4361c37da1dc26dcdd02bf40a26ee0845ce5f213bc\": rpc error: code = NotFound desc = could not find container \"cc07f102bc6ea1a48ba19d4361c37da1dc26dcdd02bf40a26ee0845ce5f213bc\": container with ID starting with cc07f102bc6ea1a48ba19d4361c37da1dc26dcdd02bf40a26ee0845ce5f213bc not found: ID does not exist" Dec 06 06:52:41 crc kubenswrapper[4809]: I1206 06:52:41.404761 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd7ad127-708d-4846-8b67-b14323c4de29" path="/var/lib/kubelet/pods/cd7ad127-708d-4846-8b67-b14323c4de29/volumes" Dec 06 06:53:04 crc kubenswrapper[4809]: I1206 06:53:04.496118 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:53:04 crc kubenswrapper[4809]: I1206 06:53:04.496763 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:53:34 crc kubenswrapper[4809]: I1206 06:53:34.496916 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:53:34 crc kubenswrapper[4809]: I1206 06:53:34.497501 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:53:34 crc kubenswrapper[4809]: I1206 06:53:34.497556 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-npms2" Dec 06 06:53:34 crc kubenswrapper[4809]: I1206 06:53:34.498197 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"030a0d8b374914264276796b77b60278c64c06d896d2041127b7a240ba65b24c"} pod="openshift-machine-config-operator/machine-config-daemon-npms2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 06:53:34 crc kubenswrapper[4809]: I1206 06:53:34.498251 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" containerID="cri-o://030a0d8b374914264276796b77b60278c64c06d896d2041127b7a240ba65b24c" gracePeriod=600 Dec 06 06:53:34 crc kubenswrapper[4809]: E1206 06:53:34.615826 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:53:35 crc kubenswrapper[4809]: I1206 06:53:35.447721 4809 generic.go:334] "Generic (PLEG): container finished" podID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerID="030a0d8b374914264276796b77b60278c64c06d896d2041127b7a240ba65b24c" exitCode=0 Dec 06 06:53:35 crc kubenswrapper[4809]: I1206 06:53:35.447772 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" event={"ID":"cbf4cf62-024e-4703-a8b8-9aecda9cd26a","Type":"ContainerDied","Data":"030a0d8b374914264276796b77b60278c64c06d896d2041127b7a240ba65b24c"} Dec 06 06:53:35 crc kubenswrapper[4809]: I1206 06:53:35.447809 4809 scope.go:117] "RemoveContainer" containerID="083a8276457cb2db9f64bc581dbeeb57fcdb1feabe656331716e6ce96734d239" Dec 06 06:53:35 crc kubenswrapper[4809]: I1206 06:53:35.450867 4809 scope.go:117] "RemoveContainer" containerID="030a0d8b374914264276796b77b60278c64c06d896d2041127b7a240ba65b24c" Dec 06 06:53:35 crc kubenswrapper[4809]: E1206 06:53:35.451316 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:53:46 crc kubenswrapper[4809]: I1206 06:53:46.388852 4809 scope.go:117] "RemoveContainer" containerID="030a0d8b374914264276796b77b60278c64c06d896d2041127b7a240ba65b24c" Dec 06 06:53:46 crc kubenswrapper[4809]: E1206 06:53:46.389528 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:53:57 crc kubenswrapper[4809]: I1206 06:53:57.389754 4809 scope.go:117] "RemoveContainer" containerID="030a0d8b374914264276796b77b60278c64c06d896d2041127b7a240ba65b24c" Dec 06 06:53:57 crc kubenswrapper[4809]: E1206 06:53:57.390768 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:54:09 crc kubenswrapper[4809]: I1206 06:54:09.388920 4809 scope.go:117] "RemoveContainer" containerID="030a0d8b374914264276796b77b60278c64c06d896d2041127b7a240ba65b24c" Dec 06 06:54:09 crc kubenswrapper[4809]: E1206 06:54:09.390055 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:54:20 crc kubenswrapper[4809]: I1206 06:54:20.389249 4809 scope.go:117] "RemoveContainer" containerID="030a0d8b374914264276796b77b60278c64c06d896d2041127b7a240ba65b24c" Dec 06 06:54:20 crc kubenswrapper[4809]: E1206 06:54:20.389880 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:54:35 crc kubenswrapper[4809]: I1206 06:54:35.396127 4809 scope.go:117] "RemoveContainer" containerID="030a0d8b374914264276796b77b60278c64c06d896d2041127b7a240ba65b24c" Dec 06 06:54:35 crc kubenswrapper[4809]: E1206 06:54:35.396976 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:54:48 crc kubenswrapper[4809]: I1206 06:54:48.388438 4809 scope.go:117] "RemoveContainer" containerID="030a0d8b374914264276796b77b60278c64c06d896d2041127b7a240ba65b24c" Dec 06 06:54:48 crc kubenswrapper[4809]: E1206 06:54:48.389204 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:55:01 crc kubenswrapper[4809]: I1206 06:55:01.388626 4809 scope.go:117] "RemoveContainer" containerID="030a0d8b374914264276796b77b60278c64c06d896d2041127b7a240ba65b24c" Dec 06 06:55:01 crc kubenswrapper[4809]: E1206 06:55:01.389521 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:55:15 crc kubenswrapper[4809]: I1206 06:55:15.406410 4809 scope.go:117] "RemoveContainer" containerID="030a0d8b374914264276796b77b60278c64c06d896d2041127b7a240ba65b24c" Dec 06 06:55:15 crc kubenswrapper[4809]: E1206 06:55:15.407623 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:55:27 crc kubenswrapper[4809]: I1206 06:55:27.390321 4809 scope.go:117] "RemoveContainer" containerID="030a0d8b374914264276796b77b60278c64c06d896d2041127b7a240ba65b24c" Dec 06 06:55:27 crc kubenswrapper[4809]: E1206 06:55:27.391101 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:55:38 crc kubenswrapper[4809]: I1206 06:55:38.389398 4809 scope.go:117] "RemoveContainer" containerID="030a0d8b374914264276796b77b60278c64c06d896d2041127b7a240ba65b24c" Dec 06 06:55:38 crc kubenswrapper[4809]: E1206 06:55:38.390132 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:55:51 crc kubenswrapper[4809]: I1206 06:55:51.392074 4809 scope.go:117] "RemoveContainer" containerID="030a0d8b374914264276796b77b60278c64c06d896d2041127b7a240ba65b24c" Dec 06 06:55:51 crc kubenswrapper[4809]: E1206 06:55:51.393098 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:56:03 crc kubenswrapper[4809]: I1206 06:56:03.389161 4809 scope.go:117] "RemoveContainer" containerID="030a0d8b374914264276796b77b60278c64c06d896d2041127b7a240ba65b24c" Dec 06 06:56:03 crc kubenswrapper[4809]: E1206 06:56:03.390147 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:56:18 crc kubenswrapper[4809]: I1206 06:56:18.187052 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-c4c5c"] Dec 06 06:56:18 crc kubenswrapper[4809]: E1206 06:56:18.188443 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd7ad127-708d-4846-8b67-b14323c4de29" containerName="extract-content" Dec 06 06:56:18 crc kubenswrapper[4809]: I1206 06:56:18.188465 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd7ad127-708d-4846-8b67-b14323c4de29" containerName="extract-content" Dec 06 06:56:18 crc kubenswrapper[4809]: E1206 06:56:18.188545 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd7ad127-708d-4846-8b67-b14323c4de29" containerName="extract-utilities" Dec 06 06:56:18 crc kubenswrapper[4809]: I1206 06:56:18.188560 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd7ad127-708d-4846-8b67-b14323c4de29" containerName="extract-utilities" Dec 06 06:56:18 crc kubenswrapper[4809]: E1206 06:56:18.188586 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd7ad127-708d-4846-8b67-b14323c4de29" containerName="registry-server" Dec 06 06:56:18 crc kubenswrapper[4809]: I1206 06:56:18.188599 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd7ad127-708d-4846-8b67-b14323c4de29" containerName="registry-server" Dec 06 06:56:18 crc kubenswrapper[4809]: I1206 06:56:18.189062 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd7ad127-708d-4846-8b67-b14323c4de29" containerName="registry-server" Dec 06 06:56:18 crc kubenswrapper[4809]: I1206 06:56:18.191982 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c4c5c" Dec 06 06:56:18 crc kubenswrapper[4809]: I1206 06:56:18.218024 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-c4c5c"] Dec 06 06:56:18 crc kubenswrapper[4809]: I1206 06:56:18.370416 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35344077-f558-4635-87ce-70541624b0f6-catalog-content\") pod \"certified-operators-c4c5c\" (UID: \"35344077-f558-4635-87ce-70541624b0f6\") " pod="openshift-marketplace/certified-operators-c4c5c" Dec 06 06:56:18 crc kubenswrapper[4809]: I1206 06:56:18.370788 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-psrrv\" (UniqueName: \"kubernetes.io/projected/35344077-f558-4635-87ce-70541624b0f6-kube-api-access-psrrv\") pod \"certified-operators-c4c5c\" (UID: \"35344077-f558-4635-87ce-70541624b0f6\") " pod="openshift-marketplace/certified-operators-c4c5c" Dec 06 06:56:18 crc kubenswrapper[4809]: I1206 06:56:18.370821 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35344077-f558-4635-87ce-70541624b0f6-utilities\") pod \"certified-operators-c4c5c\" (UID: \"35344077-f558-4635-87ce-70541624b0f6\") " pod="openshift-marketplace/certified-operators-c4c5c" Dec 06 06:56:18 crc kubenswrapper[4809]: I1206 06:56:18.391176 4809 scope.go:117] "RemoveContainer" containerID="030a0d8b374914264276796b77b60278c64c06d896d2041127b7a240ba65b24c" Dec 06 06:56:18 crc kubenswrapper[4809]: E1206 06:56:18.391536 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:56:18 crc kubenswrapper[4809]: I1206 06:56:18.473671 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35344077-f558-4635-87ce-70541624b0f6-catalog-content\") pod \"certified-operators-c4c5c\" (UID: \"35344077-f558-4635-87ce-70541624b0f6\") " pod="openshift-marketplace/certified-operators-c4c5c" Dec 06 06:56:18 crc kubenswrapper[4809]: I1206 06:56:18.476079 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-psrrv\" (UniqueName: \"kubernetes.io/projected/35344077-f558-4635-87ce-70541624b0f6-kube-api-access-psrrv\") pod \"certified-operators-c4c5c\" (UID: \"35344077-f558-4635-87ce-70541624b0f6\") " pod="openshift-marketplace/certified-operators-c4c5c" Dec 06 06:56:18 crc kubenswrapper[4809]: I1206 06:56:18.476251 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35344077-f558-4635-87ce-70541624b0f6-utilities\") pod \"certified-operators-c4c5c\" (UID: \"35344077-f558-4635-87ce-70541624b0f6\") " pod="openshift-marketplace/certified-operators-c4c5c" Dec 06 06:56:18 crc kubenswrapper[4809]: I1206 06:56:18.474358 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35344077-f558-4635-87ce-70541624b0f6-catalog-content\") pod \"certified-operators-c4c5c\" (UID: \"35344077-f558-4635-87ce-70541624b0f6\") " pod="openshift-marketplace/certified-operators-c4c5c" Dec 06 06:56:18 crc kubenswrapper[4809]: I1206 06:56:18.477050 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35344077-f558-4635-87ce-70541624b0f6-utilities\") pod \"certified-operators-c4c5c\" (UID: \"35344077-f558-4635-87ce-70541624b0f6\") " pod="openshift-marketplace/certified-operators-c4c5c" Dec 06 06:56:18 crc kubenswrapper[4809]: I1206 06:56:18.500565 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-psrrv\" (UniqueName: \"kubernetes.io/projected/35344077-f558-4635-87ce-70541624b0f6-kube-api-access-psrrv\") pod \"certified-operators-c4c5c\" (UID: \"35344077-f558-4635-87ce-70541624b0f6\") " pod="openshift-marketplace/certified-operators-c4c5c" Dec 06 06:56:18 crc kubenswrapper[4809]: I1206 06:56:18.525737 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c4c5c" Dec 06 06:56:19 crc kubenswrapper[4809]: I1206 06:56:19.067542 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-c4c5c"] Dec 06 06:56:19 crc kubenswrapper[4809]: I1206 06:56:19.442005 4809 generic.go:334] "Generic (PLEG): container finished" podID="35344077-f558-4635-87ce-70541624b0f6" containerID="7ab8bf3437d0a10acb9f39ad1e8c39395b6303a530a5da4a9e09b410a5c6073a" exitCode=0 Dec 06 06:56:19 crc kubenswrapper[4809]: I1206 06:56:19.442045 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c4c5c" event={"ID":"35344077-f558-4635-87ce-70541624b0f6","Type":"ContainerDied","Data":"7ab8bf3437d0a10acb9f39ad1e8c39395b6303a530a5da4a9e09b410a5c6073a"} Dec 06 06:56:19 crc kubenswrapper[4809]: I1206 06:56:19.442080 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c4c5c" event={"ID":"35344077-f558-4635-87ce-70541624b0f6","Type":"ContainerStarted","Data":"a3d20546af3dd7e8d8efde42cf0c8f1ff6e5ba40ecf9127fd495fbbc21333194"} Dec 06 06:56:21 crc kubenswrapper[4809]: I1206 06:56:21.470149 4809 generic.go:334] "Generic (PLEG): container finished" podID="35344077-f558-4635-87ce-70541624b0f6" containerID="99ed2d4da41835424cc27949ab4677b5247895ca37a126a98641fe2ea301a12e" exitCode=0 Dec 06 06:56:21 crc kubenswrapper[4809]: I1206 06:56:21.470249 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c4c5c" event={"ID":"35344077-f558-4635-87ce-70541624b0f6","Type":"ContainerDied","Data":"99ed2d4da41835424cc27949ab4677b5247895ca37a126a98641fe2ea301a12e"} Dec 06 06:56:23 crc kubenswrapper[4809]: I1206 06:56:23.496813 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c4c5c" event={"ID":"35344077-f558-4635-87ce-70541624b0f6","Type":"ContainerStarted","Data":"10fcbba98615e0f55d9b6fb842d119f2f694bb4062456471c78ada5c14e44b72"} Dec 06 06:56:23 crc kubenswrapper[4809]: I1206 06:56:23.528437 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-c4c5c" podStartSLOduration=2.8338294939999997 podStartE2EDuration="5.528414898s" podCreationTimestamp="2025-12-06 06:56:18 +0000 UTC" firstStartedPulling="2025-12-06 06:56:19.443721097 +0000 UTC m=+3904.332704029" lastFinishedPulling="2025-12-06 06:56:22.138306461 +0000 UTC m=+3907.027289433" observedRunningTime="2025-12-06 06:56:23.514738218 +0000 UTC m=+3908.403721160" watchObservedRunningTime="2025-12-06 06:56:23.528414898 +0000 UTC m=+3908.417397850" Dec 06 06:56:28 crc kubenswrapper[4809]: I1206 06:56:28.526755 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-c4c5c" Dec 06 06:56:28 crc kubenswrapper[4809]: I1206 06:56:28.527241 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-c4c5c" Dec 06 06:56:28 crc kubenswrapper[4809]: I1206 06:56:28.579963 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-c4c5c" Dec 06 06:56:28 crc kubenswrapper[4809]: I1206 06:56:28.633735 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-c4c5c" Dec 06 06:56:28 crc kubenswrapper[4809]: I1206 06:56:28.825335 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-c4c5c"] Dec 06 06:56:30 crc kubenswrapper[4809]: I1206 06:56:30.582886 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-c4c5c" podUID="35344077-f558-4635-87ce-70541624b0f6" containerName="registry-server" containerID="cri-o://10fcbba98615e0f55d9b6fb842d119f2f694bb4062456471c78ada5c14e44b72" gracePeriod=2 Dec 06 06:56:31 crc kubenswrapper[4809]: I1206 06:56:31.601854 4809 generic.go:334] "Generic (PLEG): container finished" podID="35344077-f558-4635-87ce-70541624b0f6" containerID="10fcbba98615e0f55d9b6fb842d119f2f694bb4062456471c78ada5c14e44b72" exitCode=0 Dec 06 06:56:31 crc kubenswrapper[4809]: I1206 06:56:31.601924 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c4c5c" event={"ID":"35344077-f558-4635-87ce-70541624b0f6","Type":"ContainerDied","Data":"10fcbba98615e0f55d9b6fb842d119f2f694bb4062456471c78ada5c14e44b72"} Dec 06 06:56:31 crc kubenswrapper[4809]: I1206 06:56:31.602233 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c4c5c" event={"ID":"35344077-f558-4635-87ce-70541624b0f6","Type":"ContainerDied","Data":"a3d20546af3dd7e8d8efde42cf0c8f1ff6e5ba40ecf9127fd495fbbc21333194"} Dec 06 06:56:31 crc kubenswrapper[4809]: I1206 06:56:31.602250 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a3d20546af3dd7e8d8efde42cf0c8f1ff6e5ba40ecf9127fd495fbbc21333194" Dec 06 06:56:31 crc kubenswrapper[4809]: I1206 06:56:31.655579 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c4c5c" Dec 06 06:56:31 crc kubenswrapper[4809]: I1206 06:56:31.804977 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35344077-f558-4635-87ce-70541624b0f6-catalog-content\") pod \"35344077-f558-4635-87ce-70541624b0f6\" (UID: \"35344077-f558-4635-87ce-70541624b0f6\") " Dec 06 06:56:31 crc kubenswrapper[4809]: I1206 06:56:31.805054 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35344077-f558-4635-87ce-70541624b0f6-utilities\") pod \"35344077-f558-4635-87ce-70541624b0f6\" (UID: \"35344077-f558-4635-87ce-70541624b0f6\") " Dec 06 06:56:31 crc kubenswrapper[4809]: I1206 06:56:31.805133 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-psrrv\" (UniqueName: \"kubernetes.io/projected/35344077-f558-4635-87ce-70541624b0f6-kube-api-access-psrrv\") pod \"35344077-f558-4635-87ce-70541624b0f6\" (UID: \"35344077-f558-4635-87ce-70541624b0f6\") " Dec 06 06:56:31 crc kubenswrapper[4809]: I1206 06:56:31.805881 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35344077-f558-4635-87ce-70541624b0f6-utilities" (OuterVolumeSpecName: "utilities") pod "35344077-f558-4635-87ce-70541624b0f6" (UID: "35344077-f558-4635-87ce-70541624b0f6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:56:31 crc kubenswrapper[4809]: I1206 06:56:31.813418 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35344077-f558-4635-87ce-70541624b0f6-kube-api-access-psrrv" (OuterVolumeSpecName: "kube-api-access-psrrv") pod "35344077-f558-4635-87ce-70541624b0f6" (UID: "35344077-f558-4635-87ce-70541624b0f6"). InnerVolumeSpecName "kube-api-access-psrrv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:56:31 crc kubenswrapper[4809]: I1206 06:56:31.865437 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35344077-f558-4635-87ce-70541624b0f6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "35344077-f558-4635-87ce-70541624b0f6" (UID: "35344077-f558-4635-87ce-70541624b0f6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:56:31 crc kubenswrapper[4809]: I1206 06:56:31.908493 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35344077-f558-4635-87ce-70541624b0f6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:56:31 crc kubenswrapper[4809]: I1206 06:56:31.908553 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35344077-f558-4635-87ce-70541624b0f6-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:56:31 crc kubenswrapper[4809]: I1206 06:56:31.908567 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-psrrv\" (UniqueName: \"kubernetes.io/projected/35344077-f558-4635-87ce-70541624b0f6-kube-api-access-psrrv\") on node \"crc\" DevicePath \"\"" Dec 06 06:56:32 crc kubenswrapper[4809]: I1206 06:56:32.388975 4809 scope.go:117] "RemoveContainer" containerID="030a0d8b374914264276796b77b60278c64c06d896d2041127b7a240ba65b24c" Dec 06 06:56:32 crc kubenswrapper[4809]: E1206 06:56:32.389497 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:56:32 crc kubenswrapper[4809]: I1206 06:56:32.611840 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c4c5c" Dec 06 06:56:32 crc kubenswrapper[4809]: I1206 06:56:32.670418 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-c4c5c"] Dec 06 06:56:32 crc kubenswrapper[4809]: I1206 06:56:32.685752 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-c4c5c"] Dec 06 06:56:33 crc kubenswrapper[4809]: I1206 06:56:33.404192 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35344077-f558-4635-87ce-70541624b0f6" path="/var/lib/kubelet/pods/35344077-f558-4635-87ce-70541624b0f6/volumes" Dec 06 06:56:43 crc kubenswrapper[4809]: I1206 06:56:43.389541 4809 scope.go:117] "RemoveContainer" containerID="030a0d8b374914264276796b77b60278c64c06d896d2041127b7a240ba65b24c" Dec 06 06:56:43 crc kubenswrapper[4809]: E1206 06:56:43.390343 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:56:55 crc kubenswrapper[4809]: I1206 06:56:55.406526 4809 scope.go:117] "RemoveContainer" containerID="030a0d8b374914264276796b77b60278c64c06d896d2041127b7a240ba65b24c" Dec 06 06:56:55 crc kubenswrapper[4809]: E1206 06:56:55.407699 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:57:07 crc kubenswrapper[4809]: I1206 06:57:07.389339 4809 scope.go:117] "RemoveContainer" containerID="030a0d8b374914264276796b77b60278c64c06d896d2041127b7a240ba65b24c" Dec 06 06:57:07 crc kubenswrapper[4809]: E1206 06:57:07.390181 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:57:20 crc kubenswrapper[4809]: I1206 06:57:20.388925 4809 scope.go:117] "RemoveContainer" containerID="030a0d8b374914264276796b77b60278c64c06d896d2041127b7a240ba65b24c" Dec 06 06:57:20 crc kubenswrapper[4809]: E1206 06:57:20.389785 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:57:35 crc kubenswrapper[4809]: I1206 06:57:35.388954 4809 scope.go:117] "RemoveContainer" containerID="030a0d8b374914264276796b77b60278c64c06d896d2041127b7a240ba65b24c" Dec 06 06:57:35 crc kubenswrapper[4809]: E1206 06:57:35.389880 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:57:46 crc kubenswrapper[4809]: I1206 06:57:46.388539 4809 scope.go:117] "RemoveContainer" containerID="030a0d8b374914264276796b77b60278c64c06d896d2041127b7a240ba65b24c" Dec 06 06:57:46 crc kubenswrapper[4809]: E1206 06:57:46.389365 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:57:57 crc kubenswrapper[4809]: I1206 06:57:57.390006 4809 scope.go:117] "RemoveContainer" containerID="030a0d8b374914264276796b77b60278c64c06d896d2041127b7a240ba65b24c" Dec 06 06:57:57 crc kubenswrapper[4809]: E1206 06:57:57.391510 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:58:11 crc kubenswrapper[4809]: I1206 06:58:11.389046 4809 scope.go:117] "RemoveContainer" containerID="030a0d8b374914264276796b77b60278c64c06d896d2041127b7a240ba65b24c" Dec 06 06:58:11 crc kubenswrapper[4809]: E1206 06:58:11.390588 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:58:25 crc kubenswrapper[4809]: I1206 06:58:25.388770 4809 scope.go:117] "RemoveContainer" containerID="030a0d8b374914264276796b77b60278c64c06d896d2041127b7a240ba65b24c" Dec 06 06:58:25 crc kubenswrapper[4809]: E1206 06:58:25.389579 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 06:58:39 crc kubenswrapper[4809]: I1206 06:58:39.389390 4809 scope.go:117] "RemoveContainer" containerID="030a0d8b374914264276796b77b60278c64c06d896d2041127b7a240ba65b24c" Dec 06 06:58:40 crc kubenswrapper[4809]: I1206 06:58:40.166569 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" event={"ID":"cbf4cf62-024e-4703-a8b8-9aecda9cd26a","Type":"ContainerStarted","Data":"3c974c89cf5906fdd98dcc202d2af19f666c998fb5f2186f85d25c74486025db"} Dec 06 06:59:36 crc kubenswrapper[4809]: I1206 06:59:36.928800 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5hfq8"] Dec 06 06:59:36 crc kubenswrapper[4809]: E1206 06:59:36.930013 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35344077-f558-4635-87ce-70541624b0f6" containerName="registry-server" Dec 06 06:59:36 crc kubenswrapper[4809]: I1206 06:59:36.930033 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="35344077-f558-4635-87ce-70541624b0f6" containerName="registry-server" Dec 06 06:59:36 crc kubenswrapper[4809]: E1206 06:59:36.930075 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35344077-f558-4635-87ce-70541624b0f6" containerName="extract-content" Dec 06 06:59:36 crc kubenswrapper[4809]: I1206 06:59:36.930083 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="35344077-f558-4635-87ce-70541624b0f6" containerName="extract-content" Dec 06 06:59:36 crc kubenswrapper[4809]: E1206 06:59:36.930125 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35344077-f558-4635-87ce-70541624b0f6" containerName="extract-utilities" Dec 06 06:59:36 crc kubenswrapper[4809]: I1206 06:59:36.930134 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="35344077-f558-4635-87ce-70541624b0f6" containerName="extract-utilities" Dec 06 06:59:36 crc kubenswrapper[4809]: I1206 06:59:36.930447 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="35344077-f558-4635-87ce-70541624b0f6" containerName="registry-server" Dec 06 06:59:36 crc kubenswrapper[4809]: I1206 06:59:36.932677 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5hfq8" Dec 06 06:59:36 crc kubenswrapper[4809]: I1206 06:59:36.945719 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5hfq8"] Dec 06 06:59:36 crc kubenswrapper[4809]: I1206 06:59:36.973187 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxpxg\" (UniqueName: \"kubernetes.io/projected/1b568bb4-d459-4887-85f5-520e7197b5ac-kube-api-access-rxpxg\") pod \"community-operators-5hfq8\" (UID: \"1b568bb4-d459-4887-85f5-520e7197b5ac\") " pod="openshift-marketplace/community-operators-5hfq8" Dec 06 06:59:36 crc kubenswrapper[4809]: I1206 06:59:36.973565 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b568bb4-d459-4887-85f5-520e7197b5ac-catalog-content\") pod \"community-operators-5hfq8\" (UID: \"1b568bb4-d459-4887-85f5-520e7197b5ac\") " pod="openshift-marketplace/community-operators-5hfq8" Dec 06 06:59:36 crc kubenswrapper[4809]: I1206 06:59:36.973808 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b568bb4-d459-4887-85f5-520e7197b5ac-utilities\") pod \"community-operators-5hfq8\" (UID: \"1b568bb4-d459-4887-85f5-520e7197b5ac\") " pod="openshift-marketplace/community-operators-5hfq8" Dec 06 06:59:37 crc kubenswrapper[4809]: I1206 06:59:37.076336 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxpxg\" (UniqueName: \"kubernetes.io/projected/1b568bb4-d459-4887-85f5-520e7197b5ac-kube-api-access-rxpxg\") pod \"community-operators-5hfq8\" (UID: \"1b568bb4-d459-4887-85f5-520e7197b5ac\") " pod="openshift-marketplace/community-operators-5hfq8" Dec 06 06:59:37 crc kubenswrapper[4809]: I1206 06:59:37.076645 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b568bb4-d459-4887-85f5-520e7197b5ac-catalog-content\") pod \"community-operators-5hfq8\" (UID: \"1b568bb4-d459-4887-85f5-520e7197b5ac\") " pod="openshift-marketplace/community-operators-5hfq8" Dec 06 06:59:37 crc kubenswrapper[4809]: I1206 06:59:37.076840 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b568bb4-d459-4887-85f5-520e7197b5ac-utilities\") pod \"community-operators-5hfq8\" (UID: \"1b568bb4-d459-4887-85f5-520e7197b5ac\") " pod="openshift-marketplace/community-operators-5hfq8" Dec 06 06:59:37 crc kubenswrapper[4809]: I1206 06:59:37.077359 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b568bb4-d459-4887-85f5-520e7197b5ac-catalog-content\") pod \"community-operators-5hfq8\" (UID: \"1b568bb4-d459-4887-85f5-520e7197b5ac\") " pod="openshift-marketplace/community-operators-5hfq8" Dec 06 06:59:37 crc kubenswrapper[4809]: I1206 06:59:37.077399 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b568bb4-d459-4887-85f5-520e7197b5ac-utilities\") pod \"community-operators-5hfq8\" (UID: \"1b568bb4-d459-4887-85f5-520e7197b5ac\") " pod="openshift-marketplace/community-operators-5hfq8" Dec 06 06:59:37 crc kubenswrapper[4809]: I1206 06:59:37.104099 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxpxg\" (UniqueName: \"kubernetes.io/projected/1b568bb4-d459-4887-85f5-520e7197b5ac-kube-api-access-rxpxg\") pod \"community-operators-5hfq8\" (UID: \"1b568bb4-d459-4887-85f5-520e7197b5ac\") " pod="openshift-marketplace/community-operators-5hfq8" Dec 06 06:59:37 crc kubenswrapper[4809]: I1206 06:59:37.267419 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5hfq8" Dec 06 06:59:37 crc kubenswrapper[4809]: I1206 06:59:37.886836 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5hfq8"] Dec 06 06:59:38 crc kubenswrapper[4809]: I1206 06:59:38.916291 4809 generic.go:334] "Generic (PLEG): container finished" podID="1b568bb4-d459-4887-85f5-520e7197b5ac" containerID="145d9f184c0a6bdf7e135c7fe28528796f8670bc6e9ad2b656f882b834222c3c" exitCode=0 Dec 06 06:59:38 crc kubenswrapper[4809]: I1206 06:59:38.916446 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5hfq8" event={"ID":"1b568bb4-d459-4887-85f5-520e7197b5ac","Type":"ContainerDied","Data":"145d9f184c0a6bdf7e135c7fe28528796f8670bc6e9ad2b656f882b834222c3c"} Dec 06 06:59:38 crc kubenswrapper[4809]: I1206 06:59:38.916827 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5hfq8" event={"ID":"1b568bb4-d459-4887-85f5-520e7197b5ac","Type":"ContainerStarted","Data":"089cceebebd5c7835961a3a394058aba0ef4d69d907efb4df7732e805efc6fec"} Dec 06 06:59:38 crc kubenswrapper[4809]: I1206 06:59:38.919927 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 06:59:41 crc kubenswrapper[4809]: I1206 06:59:41.950024 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5hfq8" event={"ID":"1b568bb4-d459-4887-85f5-520e7197b5ac","Type":"ContainerStarted","Data":"b61cffbd2502c41b050eb1ae35809e585e80cb1f2921eb9e197d8d680ff49bce"} Dec 06 06:59:42 crc kubenswrapper[4809]: I1206 06:59:42.961803 4809 generic.go:334] "Generic (PLEG): container finished" podID="1b568bb4-d459-4887-85f5-520e7197b5ac" containerID="b61cffbd2502c41b050eb1ae35809e585e80cb1f2921eb9e197d8d680ff49bce" exitCode=0 Dec 06 06:59:42 crc kubenswrapper[4809]: I1206 06:59:42.962012 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5hfq8" event={"ID":"1b568bb4-d459-4887-85f5-520e7197b5ac","Type":"ContainerDied","Data":"b61cffbd2502c41b050eb1ae35809e585e80cb1f2921eb9e197d8d680ff49bce"} Dec 06 06:59:45 crc kubenswrapper[4809]: I1206 06:59:45.998526 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5hfq8" event={"ID":"1b568bb4-d459-4887-85f5-520e7197b5ac","Type":"ContainerStarted","Data":"25147faa2db47c45cddab143763b425215a8d02322e83ecd21cd40b0253a8ea6"} Dec 06 06:59:46 crc kubenswrapper[4809]: I1206 06:59:46.039084 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5hfq8" podStartSLOduration=3.952849408 podStartE2EDuration="10.03906082s" podCreationTimestamp="2025-12-06 06:59:36 +0000 UTC" firstStartedPulling="2025-12-06 06:59:38.919669484 +0000 UTC m=+4103.808652427" lastFinishedPulling="2025-12-06 06:59:45.005880897 +0000 UTC m=+4109.894863839" observedRunningTime="2025-12-06 06:59:46.026542961 +0000 UTC m=+4110.915525903" watchObservedRunningTime="2025-12-06 06:59:46.03906082 +0000 UTC m=+4110.928043762" Dec 06 06:59:47 crc kubenswrapper[4809]: I1206 06:59:47.268045 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5hfq8" Dec 06 06:59:47 crc kubenswrapper[4809]: I1206 06:59:47.268624 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5hfq8" Dec 06 06:59:48 crc kubenswrapper[4809]: I1206 06:59:48.322949 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-5hfq8" podUID="1b568bb4-d459-4887-85f5-520e7197b5ac" containerName="registry-server" probeResult="failure" output=< Dec 06 06:59:48 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Dec 06 06:59:48 crc kubenswrapper[4809]: > Dec 06 06:59:57 crc kubenswrapper[4809]: I1206 06:59:57.336195 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5hfq8" Dec 06 06:59:57 crc kubenswrapper[4809]: I1206 06:59:57.412753 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5hfq8" Dec 06 06:59:57 crc kubenswrapper[4809]: I1206 06:59:57.585029 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5hfq8"] Dec 06 06:59:59 crc kubenswrapper[4809]: I1206 06:59:59.169580 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5hfq8" podUID="1b568bb4-d459-4887-85f5-520e7197b5ac" containerName="registry-server" containerID="cri-o://25147faa2db47c45cddab143763b425215a8d02322e83ecd21cd40b0253a8ea6" gracePeriod=2 Dec 06 06:59:59 crc kubenswrapper[4809]: I1206 06:59:59.772683 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5hfq8" Dec 06 06:59:59 crc kubenswrapper[4809]: I1206 06:59:59.773730 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b568bb4-d459-4887-85f5-520e7197b5ac-utilities\") pod \"1b568bb4-d459-4887-85f5-520e7197b5ac\" (UID: \"1b568bb4-d459-4887-85f5-520e7197b5ac\") " Dec 06 06:59:59 crc kubenswrapper[4809]: I1206 06:59:59.774285 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b568bb4-d459-4887-85f5-520e7197b5ac-catalog-content\") pod \"1b568bb4-d459-4887-85f5-520e7197b5ac\" (UID: \"1b568bb4-d459-4887-85f5-520e7197b5ac\") " Dec 06 06:59:59 crc kubenswrapper[4809]: I1206 06:59:59.774321 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rxpxg\" (UniqueName: \"kubernetes.io/projected/1b568bb4-d459-4887-85f5-520e7197b5ac-kube-api-access-rxpxg\") pod \"1b568bb4-d459-4887-85f5-520e7197b5ac\" (UID: \"1b568bb4-d459-4887-85f5-520e7197b5ac\") " Dec 06 06:59:59 crc kubenswrapper[4809]: I1206 06:59:59.774595 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b568bb4-d459-4887-85f5-520e7197b5ac-utilities" (OuterVolumeSpecName: "utilities") pod "1b568bb4-d459-4887-85f5-520e7197b5ac" (UID: "1b568bb4-d459-4887-85f5-520e7197b5ac"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:59:59 crc kubenswrapper[4809]: I1206 06:59:59.775604 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b568bb4-d459-4887-85f5-520e7197b5ac-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:59:59 crc kubenswrapper[4809]: I1206 06:59:59.781270 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b568bb4-d459-4887-85f5-520e7197b5ac-kube-api-access-rxpxg" (OuterVolumeSpecName: "kube-api-access-rxpxg") pod "1b568bb4-d459-4887-85f5-520e7197b5ac" (UID: "1b568bb4-d459-4887-85f5-520e7197b5ac"). InnerVolumeSpecName "kube-api-access-rxpxg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:59:59 crc kubenswrapper[4809]: I1206 06:59:59.857067 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b568bb4-d459-4887-85f5-520e7197b5ac-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1b568bb4-d459-4887-85f5-520e7197b5ac" (UID: "1b568bb4-d459-4887-85f5-520e7197b5ac"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:59:59 crc kubenswrapper[4809]: I1206 06:59:59.877172 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b568bb4-d459-4887-85f5-520e7197b5ac-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:59:59 crc kubenswrapper[4809]: I1206 06:59:59.877219 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rxpxg\" (UniqueName: \"kubernetes.io/projected/1b568bb4-d459-4887-85f5-520e7197b5ac-kube-api-access-rxpxg\") on node \"crc\" DevicePath \"\"" Dec 06 07:00:00 crc kubenswrapper[4809]: I1206 07:00:00.181086 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416740-f52lt"] Dec 06 07:00:00 crc kubenswrapper[4809]: E1206 07:00:00.181742 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b568bb4-d459-4887-85f5-520e7197b5ac" containerName="extract-content" Dec 06 07:00:00 crc kubenswrapper[4809]: I1206 07:00:00.181759 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b568bb4-d459-4887-85f5-520e7197b5ac" containerName="extract-content" Dec 06 07:00:00 crc kubenswrapper[4809]: E1206 07:00:00.181834 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b568bb4-d459-4887-85f5-520e7197b5ac" containerName="extract-utilities" Dec 06 07:00:00 crc kubenswrapper[4809]: I1206 07:00:00.181842 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b568bb4-d459-4887-85f5-520e7197b5ac" containerName="extract-utilities" Dec 06 07:00:00 crc kubenswrapper[4809]: E1206 07:00:00.181859 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b568bb4-d459-4887-85f5-520e7197b5ac" containerName="registry-server" Dec 06 07:00:00 crc kubenswrapper[4809]: I1206 07:00:00.181865 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b568bb4-d459-4887-85f5-520e7197b5ac" containerName="registry-server" Dec 06 07:00:00 crc kubenswrapper[4809]: I1206 07:00:00.182212 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b568bb4-d459-4887-85f5-520e7197b5ac" containerName="registry-server" Dec 06 07:00:00 crc kubenswrapper[4809]: I1206 07:00:00.183248 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-f52lt" Dec 06 07:00:00 crc kubenswrapper[4809]: I1206 07:00:00.189551 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 07:00:00 crc kubenswrapper[4809]: I1206 07:00:00.190446 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 07:00:00 crc kubenswrapper[4809]: I1206 07:00:00.195505 4809 generic.go:334] "Generic (PLEG): container finished" podID="1b568bb4-d459-4887-85f5-520e7197b5ac" containerID="25147faa2db47c45cddab143763b425215a8d02322e83ecd21cd40b0253a8ea6" exitCode=0 Dec 06 07:00:00 crc kubenswrapper[4809]: I1206 07:00:00.195569 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5hfq8" event={"ID":"1b568bb4-d459-4887-85f5-520e7197b5ac","Type":"ContainerDied","Data":"25147faa2db47c45cddab143763b425215a8d02322e83ecd21cd40b0253a8ea6"} Dec 06 07:00:00 crc kubenswrapper[4809]: I1206 07:00:00.195628 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5hfq8" event={"ID":"1b568bb4-d459-4887-85f5-520e7197b5ac","Type":"ContainerDied","Data":"089cceebebd5c7835961a3a394058aba0ef4d69d907efb4df7732e805efc6fec"} Dec 06 07:00:00 crc kubenswrapper[4809]: I1206 07:00:00.195662 4809 scope.go:117] "RemoveContainer" containerID="25147faa2db47c45cddab143763b425215a8d02322e83ecd21cd40b0253a8ea6" Dec 06 07:00:00 crc kubenswrapper[4809]: I1206 07:00:00.195796 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5hfq8" Dec 06 07:00:00 crc kubenswrapper[4809]: I1206 07:00:00.216834 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c69eaeda-194c-4ace-ae88-a9f0f1fef8eb-config-volume\") pod \"collect-profiles-29416740-f52lt\" (UID: \"c69eaeda-194c-4ace-ae88-a9f0f1fef8eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-f52lt" Dec 06 07:00:00 crc kubenswrapper[4809]: I1206 07:00:00.219573 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xd6l8\" (UniqueName: \"kubernetes.io/projected/c69eaeda-194c-4ace-ae88-a9f0f1fef8eb-kube-api-access-xd6l8\") pod \"collect-profiles-29416740-f52lt\" (UID: \"c69eaeda-194c-4ace-ae88-a9f0f1fef8eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-f52lt" Dec 06 07:00:00 crc kubenswrapper[4809]: I1206 07:00:00.219781 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c69eaeda-194c-4ace-ae88-a9f0f1fef8eb-secret-volume\") pod \"collect-profiles-29416740-f52lt\" (UID: \"c69eaeda-194c-4ace-ae88-a9f0f1fef8eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-f52lt" Dec 06 07:00:00 crc kubenswrapper[4809]: I1206 07:00:00.236475 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416740-f52lt"] Dec 06 07:00:00 crc kubenswrapper[4809]: I1206 07:00:00.252110 4809 scope.go:117] "RemoveContainer" containerID="b61cffbd2502c41b050eb1ae35809e585e80cb1f2921eb9e197d8d680ff49bce" Dec 06 07:00:00 crc kubenswrapper[4809]: I1206 07:00:00.272120 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5hfq8"] Dec 06 07:00:00 crc kubenswrapper[4809]: I1206 07:00:00.280990 4809 scope.go:117] "RemoveContainer" containerID="145d9f184c0a6bdf7e135c7fe28528796f8670bc6e9ad2b656f882b834222c3c" Dec 06 07:00:00 crc kubenswrapper[4809]: I1206 07:00:00.283999 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5hfq8"] Dec 06 07:00:00 crc kubenswrapper[4809]: I1206 07:00:00.322623 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xd6l8\" (UniqueName: \"kubernetes.io/projected/c69eaeda-194c-4ace-ae88-a9f0f1fef8eb-kube-api-access-xd6l8\") pod \"collect-profiles-29416740-f52lt\" (UID: \"c69eaeda-194c-4ace-ae88-a9f0f1fef8eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-f52lt" Dec 06 07:00:00 crc kubenswrapper[4809]: I1206 07:00:00.322797 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c69eaeda-194c-4ace-ae88-a9f0f1fef8eb-secret-volume\") pod \"collect-profiles-29416740-f52lt\" (UID: \"c69eaeda-194c-4ace-ae88-a9f0f1fef8eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-f52lt" Dec 06 07:00:00 crc kubenswrapper[4809]: I1206 07:00:00.322911 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c69eaeda-194c-4ace-ae88-a9f0f1fef8eb-config-volume\") pod \"collect-profiles-29416740-f52lt\" (UID: \"c69eaeda-194c-4ace-ae88-a9f0f1fef8eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-f52lt" Dec 06 07:00:00 crc kubenswrapper[4809]: I1206 07:00:00.324168 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c69eaeda-194c-4ace-ae88-a9f0f1fef8eb-config-volume\") pod \"collect-profiles-29416740-f52lt\" (UID: \"c69eaeda-194c-4ace-ae88-a9f0f1fef8eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-f52lt" Dec 06 07:00:00 crc kubenswrapper[4809]: I1206 07:00:00.332086 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c69eaeda-194c-4ace-ae88-a9f0f1fef8eb-secret-volume\") pod \"collect-profiles-29416740-f52lt\" (UID: \"c69eaeda-194c-4ace-ae88-a9f0f1fef8eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-f52lt" Dec 06 07:00:00 crc kubenswrapper[4809]: I1206 07:00:00.340530 4809 scope.go:117] "RemoveContainer" containerID="25147faa2db47c45cddab143763b425215a8d02322e83ecd21cd40b0253a8ea6" Dec 06 07:00:00 crc kubenswrapper[4809]: I1206 07:00:00.340680 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xd6l8\" (UniqueName: \"kubernetes.io/projected/c69eaeda-194c-4ace-ae88-a9f0f1fef8eb-kube-api-access-xd6l8\") pod \"collect-profiles-29416740-f52lt\" (UID: \"c69eaeda-194c-4ace-ae88-a9f0f1fef8eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-f52lt" Dec 06 07:00:00 crc kubenswrapper[4809]: E1206 07:00:00.341677 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25147faa2db47c45cddab143763b425215a8d02322e83ecd21cd40b0253a8ea6\": container with ID starting with 25147faa2db47c45cddab143763b425215a8d02322e83ecd21cd40b0253a8ea6 not found: ID does not exist" containerID="25147faa2db47c45cddab143763b425215a8d02322e83ecd21cd40b0253a8ea6" Dec 06 07:00:00 crc kubenswrapper[4809]: I1206 07:00:00.341771 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25147faa2db47c45cddab143763b425215a8d02322e83ecd21cd40b0253a8ea6"} err="failed to get container status \"25147faa2db47c45cddab143763b425215a8d02322e83ecd21cd40b0253a8ea6\": rpc error: code = NotFound desc = could not find container \"25147faa2db47c45cddab143763b425215a8d02322e83ecd21cd40b0253a8ea6\": container with ID starting with 25147faa2db47c45cddab143763b425215a8d02322e83ecd21cd40b0253a8ea6 not found: ID does not exist" Dec 06 07:00:00 crc kubenswrapper[4809]: I1206 07:00:00.341802 4809 scope.go:117] "RemoveContainer" containerID="b61cffbd2502c41b050eb1ae35809e585e80cb1f2921eb9e197d8d680ff49bce" Dec 06 07:00:00 crc kubenswrapper[4809]: E1206 07:00:00.342177 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b61cffbd2502c41b050eb1ae35809e585e80cb1f2921eb9e197d8d680ff49bce\": container with ID starting with b61cffbd2502c41b050eb1ae35809e585e80cb1f2921eb9e197d8d680ff49bce not found: ID does not exist" containerID="b61cffbd2502c41b050eb1ae35809e585e80cb1f2921eb9e197d8d680ff49bce" Dec 06 07:00:00 crc kubenswrapper[4809]: I1206 07:00:00.342204 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b61cffbd2502c41b050eb1ae35809e585e80cb1f2921eb9e197d8d680ff49bce"} err="failed to get container status \"b61cffbd2502c41b050eb1ae35809e585e80cb1f2921eb9e197d8d680ff49bce\": rpc error: code = NotFound desc = could not find container \"b61cffbd2502c41b050eb1ae35809e585e80cb1f2921eb9e197d8d680ff49bce\": container with ID starting with b61cffbd2502c41b050eb1ae35809e585e80cb1f2921eb9e197d8d680ff49bce not found: ID does not exist" Dec 06 07:00:00 crc kubenswrapper[4809]: I1206 07:00:00.342220 4809 scope.go:117] "RemoveContainer" containerID="145d9f184c0a6bdf7e135c7fe28528796f8670bc6e9ad2b656f882b834222c3c" Dec 06 07:00:00 crc kubenswrapper[4809]: E1206 07:00:00.342448 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"145d9f184c0a6bdf7e135c7fe28528796f8670bc6e9ad2b656f882b834222c3c\": container with ID starting with 145d9f184c0a6bdf7e135c7fe28528796f8670bc6e9ad2b656f882b834222c3c not found: ID does not exist" containerID="145d9f184c0a6bdf7e135c7fe28528796f8670bc6e9ad2b656f882b834222c3c" Dec 06 07:00:00 crc kubenswrapper[4809]: I1206 07:00:00.342473 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"145d9f184c0a6bdf7e135c7fe28528796f8670bc6e9ad2b656f882b834222c3c"} err="failed to get container status \"145d9f184c0a6bdf7e135c7fe28528796f8670bc6e9ad2b656f882b834222c3c\": rpc error: code = NotFound desc = could not find container \"145d9f184c0a6bdf7e135c7fe28528796f8670bc6e9ad2b656f882b834222c3c\": container with ID starting with 145d9f184c0a6bdf7e135c7fe28528796f8670bc6e9ad2b656f882b834222c3c not found: ID does not exist" Dec 06 07:00:00 crc kubenswrapper[4809]: I1206 07:00:00.516385 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-f52lt" Dec 06 07:00:01 crc kubenswrapper[4809]: I1206 07:00:01.034780 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416740-f52lt"] Dec 06 07:00:01 crc kubenswrapper[4809]: I1206 07:00:01.214741 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-f52lt" event={"ID":"c69eaeda-194c-4ace-ae88-a9f0f1fef8eb","Type":"ContainerStarted","Data":"51bb0ea02d0646b0d1fc0c41ca7c530b96eab2cf8e086259bb6d654cbc1e6349"} Dec 06 07:00:01 crc kubenswrapper[4809]: I1206 07:00:01.408448 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b568bb4-d459-4887-85f5-520e7197b5ac" path="/var/lib/kubelet/pods/1b568bb4-d459-4887-85f5-520e7197b5ac/volumes" Dec 06 07:00:02 crc kubenswrapper[4809]: I1206 07:00:02.232183 4809 generic.go:334] "Generic (PLEG): container finished" podID="c69eaeda-194c-4ace-ae88-a9f0f1fef8eb" containerID="ba575d58f6157e118223fe08bc8768aa64bb0384cde48f1f6a0ac6f9c4517777" exitCode=0 Dec 06 07:00:02 crc kubenswrapper[4809]: I1206 07:00:02.232243 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-f52lt" event={"ID":"c69eaeda-194c-4ace-ae88-a9f0f1fef8eb","Type":"ContainerDied","Data":"ba575d58f6157e118223fe08bc8768aa64bb0384cde48f1f6a0ac6f9c4517777"} Dec 06 07:00:03 crc kubenswrapper[4809]: I1206 07:00:03.754430 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-f52lt" Dec 06 07:00:03 crc kubenswrapper[4809]: I1206 07:00:03.834254 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c69eaeda-194c-4ace-ae88-a9f0f1fef8eb-secret-volume\") pod \"c69eaeda-194c-4ace-ae88-a9f0f1fef8eb\" (UID: \"c69eaeda-194c-4ace-ae88-a9f0f1fef8eb\") " Dec 06 07:00:03 crc kubenswrapper[4809]: I1206 07:00:03.834481 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xd6l8\" (UniqueName: \"kubernetes.io/projected/c69eaeda-194c-4ace-ae88-a9f0f1fef8eb-kube-api-access-xd6l8\") pod \"c69eaeda-194c-4ace-ae88-a9f0f1fef8eb\" (UID: \"c69eaeda-194c-4ace-ae88-a9f0f1fef8eb\") " Dec 06 07:00:03 crc kubenswrapper[4809]: I1206 07:00:03.834724 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c69eaeda-194c-4ace-ae88-a9f0f1fef8eb-config-volume\") pod \"c69eaeda-194c-4ace-ae88-a9f0f1fef8eb\" (UID: \"c69eaeda-194c-4ace-ae88-a9f0f1fef8eb\") " Dec 06 07:00:03 crc kubenswrapper[4809]: I1206 07:00:03.835726 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c69eaeda-194c-4ace-ae88-a9f0f1fef8eb-config-volume" (OuterVolumeSpecName: "config-volume") pod "c69eaeda-194c-4ace-ae88-a9f0f1fef8eb" (UID: "c69eaeda-194c-4ace-ae88-a9f0f1fef8eb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:00:03 crc kubenswrapper[4809]: I1206 07:00:03.840509 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c69eaeda-194c-4ace-ae88-a9f0f1fef8eb-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c69eaeda-194c-4ace-ae88-a9f0f1fef8eb" (UID: "c69eaeda-194c-4ace-ae88-a9f0f1fef8eb"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:00:03 crc kubenswrapper[4809]: I1206 07:00:03.841032 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c69eaeda-194c-4ace-ae88-a9f0f1fef8eb-kube-api-access-xd6l8" (OuterVolumeSpecName: "kube-api-access-xd6l8") pod "c69eaeda-194c-4ace-ae88-a9f0f1fef8eb" (UID: "c69eaeda-194c-4ace-ae88-a9f0f1fef8eb"). InnerVolumeSpecName "kube-api-access-xd6l8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:00:03 crc kubenswrapper[4809]: I1206 07:00:03.937455 4809 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c69eaeda-194c-4ace-ae88-a9f0f1fef8eb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 07:00:03 crc kubenswrapper[4809]: I1206 07:00:03.937490 4809 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c69eaeda-194c-4ace-ae88-a9f0f1fef8eb-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 07:00:03 crc kubenswrapper[4809]: I1206 07:00:03.937500 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xd6l8\" (UniqueName: \"kubernetes.io/projected/c69eaeda-194c-4ace-ae88-a9f0f1fef8eb-kube-api-access-xd6l8\") on node \"crc\" DevicePath \"\"" Dec 06 07:00:04 crc kubenswrapper[4809]: I1206 07:00:04.256780 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-f52lt" event={"ID":"c69eaeda-194c-4ace-ae88-a9f0f1fef8eb","Type":"ContainerDied","Data":"51bb0ea02d0646b0d1fc0c41ca7c530b96eab2cf8e086259bb6d654cbc1e6349"} Dec 06 07:00:04 crc kubenswrapper[4809]: I1206 07:00:04.256835 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="51bb0ea02d0646b0d1fc0c41ca7c530b96eab2cf8e086259bb6d654cbc1e6349" Dec 06 07:00:04 crc kubenswrapper[4809]: I1206 07:00:04.257159 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-f52lt" Dec 06 07:00:04 crc kubenswrapper[4809]: I1206 07:00:04.838009 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416695-sc6ls"] Dec 06 07:00:04 crc kubenswrapper[4809]: I1206 07:00:04.849295 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416695-sc6ls"] Dec 06 07:00:05 crc kubenswrapper[4809]: I1206 07:00:05.413043 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d92fc9c-754a-403e-b49f-f0ee023f60d6" path="/var/lib/kubelet/pods/6d92fc9c-754a-403e-b49f-f0ee023f60d6/volumes" Dec 06 07:00:33 crc kubenswrapper[4809]: I1206 07:00:33.684140 4809 scope.go:117] "RemoveContainer" containerID="62d3284237ce1ce31a88a82b5a0128ef317c936516d5ccef3cbb0b1ba7ed5598" Dec 06 07:01:00 crc kubenswrapper[4809]: I1206 07:01:00.161285 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29416741-fwlb2"] Dec 06 07:01:00 crc kubenswrapper[4809]: E1206 07:01:00.162737 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c69eaeda-194c-4ace-ae88-a9f0f1fef8eb" containerName="collect-profiles" Dec 06 07:01:00 crc kubenswrapper[4809]: I1206 07:01:00.162762 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c69eaeda-194c-4ace-ae88-a9f0f1fef8eb" containerName="collect-profiles" Dec 06 07:01:00 crc kubenswrapper[4809]: I1206 07:01:00.163209 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="c69eaeda-194c-4ace-ae88-a9f0f1fef8eb" containerName="collect-profiles" Dec 06 07:01:00 crc kubenswrapper[4809]: I1206 07:01:00.164446 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29416741-fwlb2" Dec 06 07:01:00 crc kubenswrapper[4809]: I1206 07:01:00.172890 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29416741-fwlb2"] Dec 06 07:01:00 crc kubenswrapper[4809]: I1206 07:01:00.310916 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a360f0fd-86ee-4fab-bf60-aed62c57d106-combined-ca-bundle\") pod \"keystone-cron-29416741-fwlb2\" (UID: \"a360f0fd-86ee-4fab-bf60-aed62c57d106\") " pod="openstack/keystone-cron-29416741-fwlb2" Dec 06 07:01:00 crc kubenswrapper[4809]: I1206 07:01:00.310991 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hb2rs\" (UniqueName: \"kubernetes.io/projected/a360f0fd-86ee-4fab-bf60-aed62c57d106-kube-api-access-hb2rs\") pod \"keystone-cron-29416741-fwlb2\" (UID: \"a360f0fd-86ee-4fab-bf60-aed62c57d106\") " pod="openstack/keystone-cron-29416741-fwlb2" Dec 06 07:01:00 crc kubenswrapper[4809]: I1206 07:01:00.311245 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a360f0fd-86ee-4fab-bf60-aed62c57d106-config-data\") pod \"keystone-cron-29416741-fwlb2\" (UID: \"a360f0fd-86ee-4fab-bf60-aed62c57d106\") " pod="openstack/keystone-cron-29416741-fwlb2" Dec 06 07:01:00 crc kubenswrapper[4809]: I1206 07:01:00.311361 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a360f0fd-86ee-4fab-bf60-aed62c57d106-fernet-keys\") pod \"keystone-cron-29416741-fwlb2\" (UID: \"a360f0fd-86ee-4fab-bf60-aed62c57d106\") " pod="openstack/keystone-cron-29416741-fwlb2" Dec 06 07:01:00 crc kubenswrapper[4809]: I1206 07:01:00.413374 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a360f0fd-86ee-4fab-bf60-aed62c57d106-combined-ca-bundle\") pod \"keystone-cron-29416741-fwlb2\" (UID: \"a360f0fd-86ee-4fab-bf60-aed62c57d106\") " pod="openstack/keystone-cron-29416741-fwlb2" Dec 06 07:01:00 crc kubenswrapper[4809]: I1206 07:01:00.413428 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hb2rs\" (UniqueName: \"kubernetes.io/projected/a360f0fd-86ee-4fab-bf60-aed62c57d106-kube-api-access-hb2rs\") pod \"keystone-cron-29416741-fwlb2\" (UID: \"a360f0fd-86ee-4fab-bf60-aed62c57d106\") " pod="openstack/keystone-cron-29416741-fwlb2" Dec 06 07:01:00 crc kubenswrapper[4809]: I1206 07:01:00.413500 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a360f0fd-86ee-4fab-bf60-aed62c57d106-config-data\") pod \"keystone-cron-29416741-fwlb2\" (UID: \"a360f0fd-86ee-4fab-bf60-aed62c57d106\") " pod="openstack/keystone-cron-29416741-fwlb2" Dec 06 07:01:00 crc kubenswrapper[4809]: I1206 07:01:00.413542 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a360f0fd-86ee-4fab-bf60-aed62c57d106-fernet-keys\") pod \"keystone-cron-29416741-fwlb2\" (UID: \"a360f0fd-86ee-4fab-bf60-aed62c57d106\") " pod="openstack/keystone-cron-29416741-fwlb2" Dec 06 07:01:00 crc kubenswrapper[4809]: I1206 07:01:00.419485 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a360f0fd-86ee-4fab-bf60-aed62c57d106-fernet-keys\") pod \"keystone-cron-29416741-fwlb2\" (UID: \"a360f0fd-86ee-4fab-bf60-aed62c57d106\") " pod="openstack/keystone-cron-29416741-fwlb2" Dec 06 07:01:00 crc kubenswrapper[4809]: I1206 07:01:00.420123 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a360f0fd-86ee-4fab-bf60-aed62c57d106-combined-ca-bundle\") pod \"keystone-cron-29416741-fwlb2\" (UID: \"a360f0fd-86ee-4fab-bf60-aed62c57d106\") " pod="openstack/keystone-cron-29416741-fwlb2" Dec 06 07:01:00 crc kubenswrapper[4809]: I1206 07:01:00.426829 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a360f0fd-86ee-4fab-bf60-aed62c57d106-config-data\") pod \"keystone-cron-29416741-fwlb2\" (UID: \"a360f0fd-86ee-4fab-bf60-aed62c57d106\") " pod="openstack/keystone-cron-29416741-fwlb2" Dec 06 07:01:00 crc kubenswrapper[4809]: I1206 07:01:00.441165 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hb2rs\" (UniqueName: \"kubernetes.io/projected/a360f0fd-86ee-4fab-bf60-aed62c57d106-kube-api-access-hb2rs\") pod \"keystone-cron-29416741-fwlb2\" (UID: \"a360f0fd-86ee-4fab-bf60-aed62c57d106\") " pod="openstack/keystone-cron-29416741-fwlb2" Dec 06 07:01:00 crc kubenswrapper[4809]: I1206 07:01:00.492509 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29416741-fwlb2" Dec 06 07:01:01 crc kubenswrapper[4809]: I1206 07:01:01.032347 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29416741-fwlb2"] Dec 06 07:01:01 crc kubenswrapper[4809]: I1206 07:01:01.098515 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29416741-fwlb2" event={"ID":"a360f0fd-86ee-4fab-bf60-aed62c57d106","Type":"ContainerStarted","Data":"f570b603e293a661b3ab66bc37c96e11b61901cf6b1f5b9f338b48708ce606d2"} Dec 06 07:01:02 crc kubenswrapper[4809]: I1206 07:01:02.124163 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29416741-fwlb2" event={"ID":"a360f0fd-86ee-4fab-bf60-aed62c57d106","Type":"ContainerStarted","Data":"61036af82f502f045ee0bc5334607886673ff37b5493140db6d97e2042e5f600"} Dec 06 07:01:03 crc kubenswrapper[4809]: I1206 07:01:03.154873 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29416741-fwlb2" podStartSLOduration=3.154854118 podStartE2EDuration="3.154854118s" podCreationTimestamp="2025-12-06 07:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:01:03.154360464 +0000 UTC m=+4188.043343416" watchObservedRunningTime="2025-12-06 07:01:03.154854118 +0000 UTC m=+4188.043837060" Dec 06 07:01:04 crc kubenswrapper[4809]: I1206 07:01:04.496596 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:01:04 crc kubenswrapper[4809]: I1206 07:01:04.497712 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:01:05 crc kubenswrapper[4809]: I1206 07:01:05.158829 4809 generic.go:334] "Generic (PLEG): container finished" podID="a360f0fd-86ee-4fab-bf60-aed62c57d106" containerID="61036af82f502f045ee0bc5334607886673ff37b5493140db6d97e2042e5f600" exitCode=0 Dec 06 07:01:05 crc kubenswrapper[4809]: I1206 07:01:05.159218 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29416741-fwlb2" event={"ID":"a360f0fd-86ee-4fab-bf60-aed62c57d106","Type":"ContainerDied","Data":"61036af82f502f045ee0bc5334607886673ff37b5493140db6d97e2042e5f600"} Dec 06 07:01:07 crc kubenswrapper[4809]: I1206 07:01:07.319109 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29416741-fwlb2" Dec 06 07:01:07 crc kubenswrapper[4809]: I1206 07:01:07.480035 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hb2rs\" (UniqueName: \"kubernetes.io/projected/a360f0fd-86ee-4fab-bf60-aed62c57d106-kube-api-access-hb2rs\") pod \"a360f0fd-86ee-4fab-bf60-aed62c57d106\" (UID: \"a360f0fd-86ee-4fab-bf60-aed62c57d106\") " Dec 06 07:01:07 crc kubenswrapper[4809]: I1206 07:01:07.480400 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a360f0fd-86ee-4fab-bf60-aed62c57d106-config-data\") pod \"a360f0fd-86ee-4fab-bf60-aed62c57d106\" (UID: \"a360f0fd-86ee-4fab-bf60-aed62c57d106\") " Dec 06 07:01:07 crc kubenswrapper[4809]: I1206 07:01:07.480492 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a360f0fd-86ee-4fab-bf60-aed62c57d106-combined-ca-bundle\") pod \"a360f0fd-86ee-4fab-bf60-aed62c57d106\" (UID: \"a360f0fd-86ee-4fab-bf60-aed62c57d106\") " Dec 06 07:01:07 crc kubenswrapper[4809]: I1206 07:01:07.480673 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a360f0fd-86ee-4fab-bf60-aed62c57d106-fernet-keys\") pod \"a360f0fd-86ee-4fab-bf60-aed62c57d106\" (UID: \"a360f0fd-86ee-4fab-bf60-aed62c57d106\") " Dec 06 07:01:07 crc kubenswrapper[4809]: I1206 07:01:07.489444 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a360f0fd-86ee-4fab-bf60-aed62c57d106-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "a360f0fd-86ee-4fab-bf60-aed62c57d106" (UID: "a360f0fd-86ee-4fab-bf60-aed62c57d106"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:01:07 crc kubenswrapper[4809]: I1206 07:01:07.489690 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a360f0fd-86ee-4fab-bf60-aed62c57d106-kube-api-access-hb2rs" (OuterVolumeSpecName: "kube-api-access-hb2rs") pod "a360f0fd-86ee-4fab-bf60-aed62c57d106" (UID: "a360f0fd-86ee-4fab-bf60-aed62c57d106"). InnerVolumeSpecName "kube-api-access-hb2rs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:01:07 crc kubenswrapper[4809]: I1206 07:01:07.514759 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a360f0fd-86ee-4fab-bf60-aed62c57d106-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a360f0fd-86ee-4fab-bf60-aed62c57d106" (UID: "a360f0fd-86ee-4fab-bf60-aed62c57d106"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:01:07 crc kubenswrapper[4809]: I1206 07:01:07.543454 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a360f0fd-86ee-4fab-bf60-aed62c57d106-config-data" (OuterVolumeSpecName: "config-data") pod "a360f0fd-86ee-4fab-bf60-aed62c57d106" (UID: "a360f0fd-86ee-4fab-bf60-aed62c57d106"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:01:07 crc kubenswrapper[4809]: I1206 07:01:07.583582 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a360f0fd-86ee-4fab-bf60-aed62c57d106-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:01:07 crc kubenswrapper[4809]: I1206 07:01:07.583622 4809 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a360f0fd-86ee-4fab-bf60-aed62c57d106-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 06 07:01:07 crc kubenswrapper[4809]: I1206 07:01:07.583632 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hb2rs\" (UniqueName: \"kubernetes.io/projected/a360f0fd-86ee-4fab-bf60-aed62c57d106-kube-api-access-hb2rs\") on node \"crc\" DevicePath \"\"" Dec 06 07:01:07 crc kubenswrapper[4809]: I1206 07:01:07.583643 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a360f0fd-86ee-4fab-bf60-aed62c57d106-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:01:08 crc kubenswrapper[4809]: I1206 07:01:08.197349 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29416741-fwlb2" event={"ID":"a360f0fd-86ee-4fab-bf60-aed62c57d106","Type":"ContainerDied","Data":"f570b603e293a661b3ab66bc37c96e11b61901cf6b1f5b9f338b48708ce606d2"} Dec 06 07:01:08 crc kubenswrapper[4809]: I1206 07:01:08.197387 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f570b603e293a661b3ab66bc37c96e11b61901cf6b1f5b9f338b48708ce606d2" Dec 06 07:01:08 crc kubenswrapper[4809]: I1206 07:01:08.198016 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29416741-fwlb2" Dec 06 07:01:34 crc kubenswrapper[4809]: I1206 07:01:34.496094 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:01:34 crc kubenswrapper[4809]: I1206 07:01:34.497082 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:01:59 crc kubenswrapper[4809]: I1206 07:01:59.579191 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/metallb-operator-webhook-server-765b7d7f99-8dngt" podUID="60798ea0-75be-4a0c-a8c6-9fb431ba0e67" containerName="webhook-server" probeResult="failure" output="Get \"http://10.217.0.93:7472/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:02:04 crc kubenswrapper[4809]: I1206 07:02:04.496696 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:02:04 crc kubenswrapper[4809]: I1206 07:02:04.497272 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:02:04 crc kubenswrapper[4809]: I1206 07:02:04.497316 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-npms2" Dec 06 07:02:04 crc kubenswrapper[4809]: I1206 07:02:04.498197 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3c974c89cf5906fdd98dcc202d2af19f666c998fb5f2186f85d25c74486025db"} pod="openshift-machine-config-operator/machine-config-daemon-npms2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 07:02:04 crc kubenswrapper[4809]: I1206 07:02:04.498259 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" containerID="cri-o://3c974c89cf5906fdd98dcc202d2af19f666c998fb5f2186f85d25c74486025db" gracePeriod=600 Dec 06 07:02:04 crc kubenswrapper[4809]: I1206 07:02:04.874470 4809 generic.go:334] "Generic (PLEG): container finished" podID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerID="3c974c89cf5906fdd98dcc202d2af19f666c998fb5f2186f85d25c74486025db" exitCode=0 Dec 06 07:02:04 crc kubenswrapper[4809]: I1206 07:02:04.874551 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" event={"ID":"cbf4cf62-024e-4703-a8b8-9aecda9cd26a","Type":"ContainerDied","Data":"3c974c89cf5906fdd98dcc202d2af19f666c998fb5f2186f85d25c74486025db"} Dec 06 07:02:04 crc kubenswrapper[4809]: I1206 07:02:04.874777 4809 scope.go:117] "RemoveContainer" containerID="030a0d8b374914264276796b77b60278c64c06d896d2041127b7a240ba65b24c" Dec 06 07:02:05 crc kubenswrapper[4809]: I1206 07:02:05.919131 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" event={"ID":"cbf4cf62-024e-4703-a8b8-9aecda9cd26a","Type":"ContainerStarted","Data":"474210a91ed30ed36d67639fb33cd7c97c7a14000a24e601812704bd7b93c234"} Dec 06 07:02:37 crc kubenswrapper[4809]: I1206 07:02:37.297147 4809 scope.go:117] "RemoveContainer" containerID="99ed2d4da41835424cc27949ab4677b5247895ca37a126a98641fe2ea301a12e" Dec 06 07:02:37 crc kubenswrapper[4809]: I1206 07:02:37.332287 4809 scope.go:117] "RemoveContainer" containerID="10fcbba98615e0f55d9b6fb842d119f2f694bb4062456471c78ada5c14e44b72" Dec 06 07:02:37 crc kubenswrapper[4809]: I1206 07:02:37.391584 4809 scope.go:117] "RemoveContainer" containerID="7ab8bf3437d0a10acb9f39ad1e8c39395b6303a530a5da4a9e09b410a5c6073a" Dec 06 07:03:16 crc kubenswrapper[4809]: I1206 07:03:16.278696 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jbps4"] Dec 06 07:03:16 crc kubenswrapper[4809]: E1206 07:03:16.279682 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a360f0fd-86ee-4fab-bf60-aed62c57d106" containerName="keystone-cron" Dec 06 07:03:16 crc kubenswrapper[4809]: I1206 07:03:16.279696 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a360f0fd-86ee-4fab-bf60-aed62c57d106" containerName="keystone-cron" Dec 06 07:03:16 crc kubenswrapper[4809]: I1206 07:03:16.290859 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="a360f0fd-86ee-4fab-bf60-aed62c57d106" containerName="keystone-cron" Dec 06 07:03:16 crc kubenswrapper[4809]: I1206 07:03:16.292674 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jbps4"] Dec 06 07:03:16 crc kubenswrapper[4809]: I1206 07:03:16.292761 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jbps4" Dec 06 07:03:16 crc kubenswrapper[4809]: I1206 07:03:16.411183 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be479b27-8b8e-4618-b78f-156fc7089a73-utilities\") pod \"redhat-operators-jbps4\" (UID: \"be479b27-8b8e-4618-b78f-156fc7089a73\") " pod="openshift-marketplace/redhat-operators-jbps4" Dec 06 07:03:16 crc kubenswrapper[4809]: I1206 07:03:16.411276 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95mh7\" (UniqueName: \"kubernetes.io/projected/be479b27-8b8e-4618-b78f-156fc7089a73-kube-api-access-95mh7\") pod \"redhat-operators-jbps4\" (UID: \"be479b27-8b8e-4618-b78f-156fc7089a73\") " pod="openshift-marketplace/redhat-operators-jbps4" Dec 06 07:03:16 crc kubenswrapper[4809]: I1206 07:03:16.411398 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be479b27-8b8e-4618-b78f-156fc7089a73-catalog-content\") pod \"redhat-operators-jbps4\" (UID: \"be479b27-8b8e-4618-b78f-156fc7089a73\") " pod="openshift-marketplace/redhat-operators-jbps4" Dec 06 07:03:16 crc kubenswrapper[4809]: I1206 07:03:16.513584 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be479b27-8b8e-4618-b78f-156fc7089a73-utilities\") pod \"redhat-operators-jbps4\" (UID: \"be479b27-8b8e-4618-b78f-156fc7089a73\") " pod="openshift-marketplace/redhat-operators-jbps4" Dec 06 07:03:16 crc kubenswrapper[4809]: I1206 07:03:16.513648 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95mh7\" (UniqueName: \"kubernetes.io/projected/be479b27-8b8e-4618-b78f-156fc7089a73-kube-api-access-95mh7\") pod \"redhat-operators-jbps4\" (UID: \"be479b27-8b8e-4618-b78f-156fc7089a73\") " pod="openshift-marketplace/redhat-operators-jbps4" Dec 06 07:03:16 crc kubenswrapper[4809]: I1206 07:03:16.513696 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be479b27-8b8e-4618-b78f-156fc7089a73-catalog-content\") pod \"redhat-operators-jbps4\" (UID: \"be479b27-8b8e-4618-b78f-156fc7089a73\") " pod="openshift-marketplace/redhat-operators-jbps4" Dec 06 07:03:16 crc kubenswrapper[4809]: I1206 07:03:16.514858 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be479b27-8b8e-4618-b78f-156fc7089a73-utilities\") pod \"redhat-operators-jbps4\" (UID: \"be479b27-8b8e-4618-b78f-156fc7089a73\") " pod="openshift-marketplace/redhat-operators-jbps4" Dec 06 07:03:16 crc kubenswrapper[4809]: I1206 07:03:16.515434 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be479b27-8b8e-4618-b78f-156fc7089a73-catalog-content\") pod \"redhat-operators-jbps4\" (UID: \"be479b27-8b8e-4618-b78f-156fc7089a73\") " pod="openshift-marketplace/redhat-operators-jbps4" Dec 06 07:03:16 crc kubenswrapper[4809]: I1206 07:03:16.535092 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95mh7\" (UniqueName: \"kubernetes.io/projected/be479b27-8b8e-4618-b78f-156fc7089a73-kube-api-access-95mh7\") pod \"redhat-operators-jbps4\" (UID: \"be479b27-8b8e-4618-b78f-156fc7089a73\") " pod="openshift-marketplace/redhat-operators-jbps4" Dec 06 07:03:16 crc kubenswrapper[4809]: I1206 07:03:16.615313 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jbps4" Dec 06 07:03:17 crc kubenswrapper[4809]: I1206 07:03:17.270154 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jbps4"] Dec 06 07:03:17 crc kubenswrapper[4809]: I1206 07:03:17.915434 4809 generic.go:334] "Generic (PLEG): container finished" podID="be479b27-8b8e-4618-b78f-156fc7089a73" containerID="aad224bb10c4b1b7b921c73dd9f55dd0be089b30e02a943662b2a8becd798b6e" exitCode=0 Dec 06 07:03:17 crc kubenswrapper[4809]: I1206 07:03:17.915521 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jbps4" event={"ID":"be479b27-8b8e-4618-b78f-156fc7089a73","Type":"ContainerDied","Data":"aad224bb10c4b1b7b921c73dd9f55dd0be089b30e02a943662b2a8becd798b6e"} Dec 06 07:03:17 crc kubenswrapper[4809]: I1206 07:03:17.915716 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jbps4" event={"ID":"be479b27-8b8e-4618-b78f-156fc7089a73","Type":"ContainerStarted","Data":"2f256e28216bdfa91c01325e66383bdb78f0dd14524dc3b3afc643294c144467"} Dec 06 07:03:18 crc kubenswrapper[4809]: I1206 07:03:18.928871 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jbps4" event={"ID":"be479b27-8b8e-4618-b78f-156fc7089a73","Type":"ContainerStarted","Data":"1d047f59863f3beece2500aa1b3aa528db0bf28ec2f756a727f1b43ea090d207"} Dec 06 07:03:23 crc kubenswrapper[4809]: I1206 07:03:23.985968 4809 generic.go:334] "Generic (PLEG): container finished" podID="be479b27-8b8e-4618-b78f-156fc7089a73" containerID="1d047f59863f3beece2500aa1b3aa528db0bf28ec2f756a727f1b43ea090d207" exitCode=0 Dec 06 07:03:23 crc kubenswrapper[4809]: I1206 07:03:23.986089 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jbps4" event={"ID":"be479b27-8b8e-4618-b78f-156fc7089a73","Type":"ContainerDied","Data":"1d047f59863f3beece2500aa1b3aa528db0bf28ec2f756a727f1b43ea090d207"} Dec 06 07:03:25 crc kubenswrapper[4809]: I1206 07:03:25.000560 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jbps4" event={"ID":"be479b27-8b8e-4618-b78f-156fc7089a73","Type":"ContainerStarted","Data":"69d907505d3b0000ddf8b1e9cea98fbb9289fe2a53a4e7a1ec54ac1a0b10a09a"} Dec 06 07:03:26 crc kubenswrapper[4809]: I1206 07:03:26.617192 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jbps4" Dec 06 07:03:26 crc kubenswrapper[4809]: I1206 07:03:26.617555 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jbps4" Dec 06 07:03:27 crc kubenswrapper[4809]: I1206 07:03:27.668094 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jbps4" podUID="be479b27-8b8e-4618-b78f-156fc7089a73" containerName="registry-server" probeResult="failure" output=< Dec 06 07:03:27 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Dec 06 07:03:27 crc kubenswrapper[4809]: > Dec 06 07:03:36 crc kubenswrapper[4809]: I1206 07:03:36.668640 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jbps4" Dec 06 07:03:36 crc kubenswrapper[4809]: I1206 07:03:36.692447 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jbps4" podStartSLOduration=14.174769806 podStartE2EDuration="20.692422643s" podCreationTimestamp="2025-12-06 07:03:16 +0000 UTC" firstStartedPulling="2025-12-06 07:03:17.917611555 +0000 UTC m=+4322.806594497" lastFinishedPulling="2025-12-06 07:03:24.435264392 +0000 UTC m=+4329.324247334" observedRunningTime="2025-12-06 07:03:25.027680521 +0000 UTC m=+4329.916663463" watchObservedRunningTime="2025-12-06 07:03:36.692422643 +0000 UTC m=+4341.581405585" Dec 06 07:03:36 crc kubenswrapper[4809]: I1206 07:03:36.734310 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jbps4" Dec 06 07:03:36 crc kubenswrapper[4809]: I1206 07:03:36.911435 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jbps4"] Dec 06 07:03:38 crc kubenswrapper[4809]: I1206 07:03:38.182874 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jbps4" podUID="be479b27-8b8e-4618-b78f-156fc7089a73" containerName="registry-server" containerID="cri-o://69d907505d3b0000ddf8b1e9cea98fbb9289fe2a53a4e7a1ec54ac1a0b10a09a" gracePeriod=2 Dec 06 07:03:38 crc kubenswrapper[4809]: I1206 07:03:38.762702 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jbps4" Dec 06 07:03:38 crc kubenswrapper[4809]: I1206 07:03:38.871373 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be479b27-8b8e-4618-b78f-156fc7089a73-catalog-content\") pod \"be479b27-8b8e-4618-b78f-156fc7089a73\" (UID: \"be479b27-8b8e-4618-b78f-156fc7089a73\") " Dec 06 07:03:38 crc kubenswrapper[4809]: I1206 07:03:38.872410 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95mh7\" (UniqueName: \"kubernetes.io/projected/be479b27-8b8e-4618-b78f-156fc7089a73-kube-api-access-95mh7\") pod \"be479b27-8b8e-4618-b78f-156fc7089a73\" (UID: \"be479b27-8b8e-4618-b78f-156fc7089a73\") " Dec 06 07:03:38 crc kubenswrapper[4809]: I1206 07:03:38.872562 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be479b27-8b8e-4618-b78f-156fc7089a73-utilities\") pod \"be479b27-8b8e-4618-b78f-156fc7089a73\" (UID: \"be479b27-8b8e-4618-b78f-156fc7089a73\") " Dec 06 07:03:38 crc kubenswrapper[4809]: I1206 07:03:38.873319 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be479b27-8b8e-4618-b78f-156fc7089a73-utilities" (OuterVolumeSpecName: "utilities") pod "be479b27-8b8e-4618-b78f-156fc7089a73" (UID: "be479b27-8b8e-4618-b78f-156fc7089a73"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:03:38 crc kubenswrapper[4809]: I1206 07:03:38.873979 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be479b27-8b8e-4618-b78f-156fc7089a73-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 07:03:38 crc kubenswrapper[4809]: I1206 07:03:38.878480 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be479b27-8b8e-4618-b78f-156fc7089a73-kube-api-access-95mh7" (OuterVolumeSpecName: "kube-api-access-95mh7") pod "be479b27-8b8e-4618-b78f-156fc7089a73" (UID: "be479b27-8b8e-4618-b78f-156fc7089a73"). InnerVolumeSpecName "kube-api-access-95mh7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:03:38 crc kubenswrapper[4809]: I1206 07:03:38.976617 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95mh7\" (UniqueName: \"kubernetes.io/projected/be479b27-8b8e-4618-b78f-156fc7089a73-kube-api-access-95mh7\") on node \"crc\" DevicePath \"\"" Dec 06 07:03:38 crc kubenswrapper[4809]: I1206 07:03:38.990869 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be479b27-8b8e-4618-b78f-156fc7089a73-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "be479b27-8b8e-4618-b78f-156fc7089a73" (UID: "be479b27-8b8e-4618-b78f-156fc7089a73"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:03:39 crc kubenswrapper[4809]: I1206 07:03:39.080097 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be479b27-8b8e-4618-b78f-156fc7089a73-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 07:03:39 crc kubenswrapper[4809]: I1206 07:03:39.196762 4809 generic.go:334] "Generic (PLEG): container finished" podID="be479b27-8b8e-4618-b78f-156fc7089a73" containerID="69d907505d3b0000ddf8b1e9cea98fbb9289fe2a53a4e7a1ec54ac1a0b10a09a" exitCode=0 Dec 06 07:03:39 crc kubenswrapper[4809]: I1206 07:03:39.196802 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jbps4" event={"ID":"be479b27-8b8e-4618-b78f-156fc7089a73","Type":"ContainerDied","Data":"69d907505d3b0000ddf8b1e9cea98fbb9289fe2a53a4e7a1ec54ac1a0b10a09a"} Dec 06 07:03:39 crc kubenswrapper[4809]: I1206 07:03:39.196844 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jbps4" event={"ID":"be479b27-8b8e-4618-b78f-156fc7089a73","Type":"ContainerDied","Data":"2f256e28216bdfa91c01325e66383bdb78f0dd14524dc3b3afc643294c144467"} Dec 06 07:03:39 crc kubenswrapper[4809]: I1206 07:03:39.196861 4809 scope.go:117] "RemoveContainer" containerID="69d907505d3b0000ddf8b1e9cea98fbb9289fe2a53a4e7a1ec54ac1a0b10a09a" Dec 06 07:03:39 crc kubenswrapper[4809]: I1206 07:03:39.197065 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jbps4" Dec 06 07:03:39 crc kubenswrapper[4809]: I1206 07:03:39.237005 4809 scope.go:117] "RemoveContainer" containerID="1d047f59863f3beece2500aa1b3aa528db0bf28ec2f756a727f1b43ea090d207" Dec 06 07:03:39 crc kubenswrapper[4809]: I1206 07:03:39.241613 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jbps4"] Dec 06 07:03:39 crc kubenswrapper[4809]: I1206 07:03:39.252853 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jbps4"] Dec 06 07:03:39 crc kubenswrapper[4809]: I1206 07:03:39.263742 4809 scope.go:117] "RemoveContainer" containerID="aad224bb10c4b1b7b921c73dd9f55dd0be089b30e02a943662b2a8becd798b6e" Dec 06 07:03:39 crc kubenswrapper[4809]: I1206 07:03:39.343701 4809 scope.go:117] "RemoveContainer" containerID="69d907505d3b0000ddf8b1e9cea98fbb9289fe2a53a4e7a1ec54ac1a0b10a09a" Dec 06 07:03:39 crc kubenswrapper[4809]: E1206 07:03:39.346177 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69d907505d3b0000ddf8b1e9cea98fbb9289fe2a53a4e7a1ec54ac1a0b10a09a\": container with ID starting with 69d907505d3b0000ddf8b1e9cea98fbb9289fe2a53a4e7a1ec54ac1a0b10a09a not found: ID does not exist" containerID="69d907505d3b0000ddf8b1e9cea98fbb9289fe2a53a4e7a1ec54ac1a0b10a09a" Dec 06 07:03:39 crc kubenswrapper[4809]: I1206 07:03:39.346226 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69d907505d3b0000ddf8b1e9cea98fbb9289fe2a53a4e7a1ec54ac1a0b10a09a"} err="failed to get container status \"69d907505d3b0000ddf8b1e9cea98fbb9289fe2a53a4e7a1ec54ac1a0b10a09a\": rpc error: code = NotFound desc = could not find container \"69d907505d3b0000ddf8b1e9cea98fbb9289fe2a53a4e7a1ec54ac1a0b10a09a\": container with ID starting with 69d907505d3b0000ddf8b1e9cea98fbb9289fe2a53a4e7a1ec54ac1a0b10a09a not found: ID does not exist" Dec 06 07:03:39 crc kubenswrapper[4809]: I1206 07:03:39.346285 4809 scope.go:117] "RemoveContainer" containerID="1d047f59863f3beece2500aa1b3aa528db0bf28ec2f756a727f1b43ea090d207" Dec 06 07:03:39 crc kubenswrapper[4809]: E1206 07:03:39.346555 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d047f59863f3beece2500aa1b3aa528db0bf28ec2f756a727f1b43ea090d207\": container with ID starting with 1d047f59863f3beece2500aa1b3aa528db0bf28ec2f756a727f1b43ea090d207 not found: ID does not exist" containerID="1d047f59863f3beece2500aa1b3aa528db0bf28ec2f756a727f1b43ea090d207" Dec 06 07:03:39 crc kubenswrapper[4809]: I1206 07:03:39.346586 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d047f59863f3beece2500aa1b3aa528db0bf28ec2f756a727f1b43ea090d207"} err="failed to get container status \"1d047f59863f3beece2500aa1b3aa528db0bf28ec2f756a727f1b43ea090d207\": rpc error: code = NotFound desc = could not find container \"1d047f59863f3beece2500aa1b3aa528db0bf28ec2f756a727f1b43ea090d207\": container with ID starting with 1d047f59863f3beece2500aa1b3aa528db0bf28ec2f756a727f1b43ea090d207 not found: ID does not exist" Dec 06 07:03:39 crc kubenswrapper[4809]: I1206 07:03:39.346605 4809 scope.go:117] "RemoveContainer" containerID="aad224bb10c4b1b7b921c73dd9f55dd0be089b30e02a943662b2a8becd798b6e" Dec 06 07:03:39 crc kubenswrapper[4809]: E1206 07:03:39.346925 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aad224bb10c4b1b7b921c73dd9f55dd0be089b30e02a943662b2a8becd798b6e\": container with ID starting with aad224bb10c4b1b7b921c73dd9f55dd0be089b30e02a943662b2a8becd798b6e not found: ID does not exist" containerID="aad224bb10c4b1b7b921c73dd9f55dd0be089b30e02a943662b2a8becd798b6e" Dec 06 07:03:39 crc kubenswrapper[4809]: I1206 07:03:39.347063 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aad224bb10c4b1b7b921c73dd9f55dd0be089b30e02a943662b2a8becd798b6e"} err="failed to get container status \"aad224bb10c4b1b7b921c73dd9f55dd0be089b30e02a943662b2a8becd798b6e\": rpc error: code = NotFound desc = could not find container \"aad224bb10c4b1b7b921c73dd9f55dd0be089b30e02a943662b2a8becd798b6e\": container with ID starting with aad224bb10c4b1b7b921c73dd9f55dd0be089b30e02a943662b2a8becd798b6e not found: ID does not exist" Dec 06 07:03:39 crc kubenswrapper[4809]: I1206 07:03:39.404392 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be479b27-8b8e-4618-b78f-156fc7089a73" path="/var/lib/kubelet/pods/be479b27-8b8e-4618-b78f-156fc7089a73/volumes" Dec 06 07:04:34 crc kubenswrapper[4809]: I1206 07:04:34.497127 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:04:34 crc kubenswrapper[4809]: I1206 07:04:34.498228 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:05:04 crc kubenswrapper[4809]: I1206 07:05:04.496349 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:05:04 crc kubenswrapper[4809]: I1206 07:05:04.497033 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:05:34 crc kubenswrapper[4809]: I1206 07:05:34.496692 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:05:34 crc kubenswrapper[4809]: I1206 07:05:34.498523 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:05:34 crc kubenswrapper[4809]: I1206 07:05:34.498894 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-npms2" Dec 06 07:05:34 crc kubenswrapper[4809]: I1206 07:05:34.500331 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"474210a91ed30ed36d67639fb33cd7c97c7a14000a24e601812704bd7b93c234"} pod="openshift-machine-config-operator/machine-config-daemon-npms2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 07:05:34 crc kubenswrapper[4809]: I1206 07:05:34.500595 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" containerID="cri-o://474210a91ed30ed36d67639fb33cd7c97c7a14000a24e601812704bd7b93c234" gracePeriod=600 Dec 06 07:05:34 crc kubenswrapper[4809]: E1206 07:05:34.710768 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:05:35 crc kubenswrapper[4809]: I1206 07:05:35.460643 4809 generic.go:334] "Generic (PLEG): container finished" podID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerID="474210a91ed30ed36d67639fb33cd7c97c7a14000a24e601812704bd7b93c234" exitCode=0 Dec 06 07:05:35 crc kubenswrapper[4809]: I1206 07:05:35.461189 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" event={"ID":"cbf4cf62-024e-4703-a8b8-9aecda9cd26a","Type":"ContainerDied","Data":"474210a91ed30ed36d67639fb33cd7c97c7a14000a24e601812704bd7b93c234"} Dec 06 07:05:35 crc kubenswrapper[4809]: I1206 07:05:35.461297 4809 scope.go:117] "RemoveContainer" containerID="3c974c89cf5906fdd98dcc202d2af19f666c998fb5f2186f85d25c74486025db" Dec 06 07:05:35 crc kubenswrapper[4809]: I1206 07:05:35.462427 4809 scope.go:117] "RemoveContainer" containerID="474210a91ed30ed36d67639fb33cd7c97c7a14000a24e601812704bd7b93c234" Dec 06 07:05:35 crc kubenswrapper[4809]: E1206 07:05:35.462909 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:05:46 crc kubenswrapper[4809]: I1206 07:05:46.389325 4809 scope.go:117] "RemoveContainer" containerID="474210a91ed30ed36d67639fb33cd7c97c7a14000a24e601812704bd7b93c234" Dec 06 07:05:46 crc kubenswrapper[4809]: E1206 07:05:46.390342 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:05:58 crc kubenswrapper[4809]: I1206 07:05:58.389093 4809 scope.go:117] "RemoveContainer" containerID="474210a91ed30ed36d67639fb33cd7c97c7a14000a24e601812704bd7b93c234" Dec 06 07:05:58 crc kubenswrapper[4809]: E1206 07:05:58.389920 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:06:10 crc kubenswrapper[4809]: I1206 07:06:10.389322 4809 scope.go:117] "RemoveContainer" containerID="474210a91ed30ed36d67639fb33cd7c97c7a14000a24e601812704bd7b93c234" Dec 06 07:06:10 crc kubenswrapper[4809]: E1206 07:06:10.391230 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:06:13 crc kubenswrapper[4809]: I1206 07:06:13.894835 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jwmsn"] Dec 06 07:06:13 crc kubenswrapper[4809]: E1206 07:06:13.895911 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be479b27-8b8e-4618-b78f-156fc7089a73" containerName="extract-utilities" Dec 06 07:06:13 crc kubenswrapper[4809]: I1206 07:06:13.895926 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="be479b27-8b8e-4618-b78f-156fc7089a73" containerName="extract-utilities" Dec 06 07:06:13 crc kubenswrapper[4809]: E1206 07:06:13.895994 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be479b27-8b8e-4618-b78f-156fc7089a73" containerName="extract-content" Dec 06 07:06:13 crc kubenswrapper[4809]: I1206 07:06:13.896004 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="be479b27-8b8e-4618-b78f-156fc7089a73" containerName="extract-content" Dec 06 07:06:13 crc kubenswrapper[4809]: E1206 07:06:13.896017 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be479b27-8b8e-4618-b78f-156fc7089a73" containerName="registry-server" Dec 06 07:06:13 crc kubenswrapper[4809]: I1206 07:06:13.896025 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="be479b27-8b8e-4618-b78f-156fc7089a73" containerName="registry-server" Dec 06 07:06:13 crc kubenswrapper[4809]: I1206 07:06:13.896342 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="be479b27-8b8e-4618-b78f-156fc7089a73" containerName="registry-server" Dec 06 07:06:13 crc kubenswrapper[4809]: I1206 07:06:13.898426 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jwmsn" Dec 06 07:06:13 crc kubenswrapper[4809]: I1206 07:06:13.926320 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jwmsn"] Dec 06 07:06:13 crc kubenswrapper[4809]: I1206 07:06:13.981700 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea34817a-9f07-44a7-b57f-7c9940de89b0-catalog-content\") pod \"redhat-marketplace-jwmsn\" (UID: \"ea34817a-9f07-44a7-b57f-7c9940de89b0\") " pod="openshift-marketplace/redhat-marketplace-jwmsn" Dec 06 07:06:13 crc kubenswrapper[4809]: I1206 07:06:13.981850 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea34817a-9f07-44a7-b57f-7c9940de89b0-utilities\") pod \"redhat-marketplace-jwmsn\" (UID: \"ea34817a-9f07-44a7-b57f-7c9940de89b0\") " pod="openshift-marketplace/redhat-marketplace-jwmsn" Dec 06 07:06:13 crc kubenswrapper[4809]: I1206 07:06:13.981971 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hptq2\" (UniqueName: \"kubernetes.io/projected/ea34817a-9f07-44a7-b57f-7c9940de89b0-kube-api-access-hptq2\") pod \"redhat-marketplace-jwmsn\" (UID: \"ea34817a-9f07-44a7-b57f-7c9940de89b0\") " pod="openshift-marketplace/redhat-marketplace-jwmsn" Dec 06 07:06:14 crc kubenswrapper[4809]: I1206 07:06:14.083924 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hptq2\" (UniqueName: \"kubernetes.io/projected/ea34817a-9f07-44a7-b57f-7c9940de89b0-kube-api-access-hptq2\") pod \"redhat-marketplace-jwmsn\" (UID: \"ea34817a-9f07-44a7-b57f-7c9940de89b0\") " pod="openshift-marketplace/redhat-marketplace-jwmsn" Dec 06 07:06:14 crc kubenswrapper[4809]: I1206 07:06:14.084096 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea34817a-9f07-44a7-b57f-7c9940de89b0-catalog-content\") pod \"redhat-marketplace-jwmsn\" (UID: \"ea34817a-9f07-44a7-b57f-7c9940de89b0\") " pod="openshift-marketplace/redhat-marketplace-jwmsn" Dec 06 07:06:14 crc kubenswrapper[4809]: I1206 07:06:14.084191 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea34817a-9f07-44a7-b57f-7c9940de89b0-utilities\") pod \"redhat-marketplace-jwmsn\" (UID: \"ea34817a-9f07-44a7-b57f-7c9940de89b0\") " pod="openshift-marketplace/redhat-marketplace-jwmsn" Dec 06 07:06:14 crc kubenswrapper[4809]: I1206 07:06:14.084700 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea34817a-9f07-44a7-b57f-7c9940de89b0-catalog-content\") pod \"redhat-marketplace-jwmsn\" (UID: \"ea34817a-9f07-44a7-b57f-7c9940de89b0\") " pod="openshift-marketplace/redhat-marketplace-jwmsn" Dec 06 07:06:14 crc kubenswrapper[4809]: I1206 07:06:14.084722 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea34817a-9f07-44a7-b57f-7c9940de89b0-utilities\") pod \"redhat-marketplace-jwmsn\" (UID: \"ea34817a-9f07-44a7-b57f-7c9940de89b0\") " pod="openshift-marketplace/redhat-marketplace-jwmsn" Dec 06 07:06:14 crc kubenswrapper[4809]: I1206 07:06:14.105040 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hptq2\" (UniqueName: \"kubernetes.io/projected/ea34817a-9f07-44a7-b57f-7c9940de89b0-kube-api-access-hptq2\") pod \"redhat-marketplace-jwmsn\" (UID: \"ea34817a-9f07-44a7-b57f-7c9940de89b0\") " pod="openshift-marketplace/redhat-marketplace-jwmsn" Dec 06 07:06:14 crc kubenswrapper[4809]: I1206 07:06:14.226895 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jwmsn" Dec 06 07:06:14 crc kubenswrapper[4809]: I1206 07:06:14.750877 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jwmsn"] Dec 06 07:06:14 crc kubenswrapper[4809]: I1206 07:06:14.968294 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jwmsn" event={"ID":"ea34817a-9f07-44a7-b57f-7c9940de89b0","Type":"ContainerStarted","Data":"bd34a895b1300e1ca2016d20a8d56836d105b0ccb26a37eaccd699d92df547df"} Dec 06 07:06:16 crc kubenswrapper[4809]: I1206 07:06:16.003841 4809 generic.go:334] "Generic (PLEG): container finished" podID="ea34817a-9f07-44a7-b57f-7c9940de89b0" containerID="dea5281f86a76991a16957a5054ded528256384a9998079f7ff42dcfb8a6d4cc" exitCode=0 Dec 06 07:06:16 crc kubenswrapper[4809]: I1206 07:06:16.003984 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jwmsn" event={"ID":"ea34817a-9f07-44a7-b57f-7c9940de89b0","Type":"ContainerDied","Data":"dea5281f86a76991a16957a5054ded528256384a9998079f7ff42dcfb8a6d4cc"} Dec 06 07:06:16 crc kubenswrapper[4809]: I1206 07:06:16.008787 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 07:06:18 crc kubenswrapper[4809]: I1206 07:06:18.033544 4809 generic.go:334] "Generic (PLEG): container finished" podID="ea34817a-9f07-44a7-b57f-7c9940de89b0" containerID="dee5480ad59364efab3a57cb85fa132511ac362f70d0c41854f573f23e3b2f20" exitCode=0 Dec 06 07:06:18 crc kubenswrapper[4809]: I1206 07:06:18.033602 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jwmsn" event={"ID":"ea34817a-9f07-44a7-b57f-7c9940de89b0","Type":"ContainerDied","Data":"dee5480ad59364efab3a57cb85fa132511ac362f70d0c41854f573f23e3b2f20"} Dec 06 07:06:19 crc kubenswrapper[4809]: I1206 07:06:19.049053 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jwmsn" event={"ID":"ea34817a-9f07-44a7-b57f-7c9940de89b0","Type":"ContainerStarted","Data":"0cd351db33eb94912c7d67298ccd4390dca2c4b2f1d237e74ced92fe7a04ffcf"} Dec 06 07:06:19 crc kubenswrapper[4809]: I1206 07:06:19.070260 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jwmsn" podStartSLOduration=3.369877818 podStartE2EDuration="6.070240691s" podCreationTimestamp="2025-12-06 07:06:13 +0000 UTC" firstStartedPulling="2025-12-06 07:06:16.008555952 +0000 UTC m=+4500.897538894" lastFinishedPulling="2025-12-06 07:06:18.708918815 +0000 UTC m=+4503.597901767" observedRunningTime="2025-12-06 07:06:19.069714357 +0000 UTC m=+4503.958697299" watchObservedRunningTime="2025-12-06 07:06:19.070240691 +0000 UTC m=+4503.959223633" Dec 06 07:06:24 crc kubenswrapper[4809]: I1206 07:06:24.227548 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jwmsn" Dec 06 07:06:24 crc kubenswrapper[4809]: I1206 07:06:24.228623 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jwmsn" Dec 06 07:06:24 crc kubenswrapper[4809]: I1206 07:06:24.282623 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jwmsn" Dec 06 07:06:25 crc kubenswrapper[4809]: I1206 07:06:25.397785 4809 scope.go:117] "RemoveContainer" containerID="474210a91ed30ed36d67639fb33cd7c97c7a14000a24e601812704bd7b93c234" Dec 06 07:06:25 crc kubenswrapper[4809]: E1206 07:06:25.398295 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:06:25 crc kubenswrapper[4809]: I1206 07:06:25.625277 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jwmsn" Dec 06 07:06:27 crc kubenswrapper[4809]: I1206 07:06:27.883627 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jwmsn"] Dec 06 07:06:27 crc kubenswrapper[4809]: I1206 07:06:27.884577 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jwmsn" podUID="ea34817a-9f07-44a7-b57f-7c9940de89b0" containerName="registry-server" containerID="cri-o://0cd351db33eb94912c7d67298ccd4390dca2c4b2f1d237e74ced92fe7a04ffcf" gracePeriod=2 Dec 06 07:06:28 crc kubenswrapper[4809]: I1206 07:06:28.175070 4809 generic.go:334] "Generic (PLEG): container finished" podID="ea34817a-9f07-44a7-b57f-7c9940de89b0" containerID="0cd351db33eb94912c7d67298ccd4390dca2c4b2f1d237e74ced92fe7a04ffcf" exitCode=0 Dec 06 07:06:28 crc kubenswrapper[4809]: I1206 07:06:28.175112 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jwmsn" event={"ID":"ea34817a-9f07-44a7-b57f-7c9940de89b0","Type":"ContainerDied","Data":"0cd351db33eb94912c7d67298ccd4390dca2c4b2f1d237e74ced92fe7a04ffcf"} Dec 06 07:06:29 crc kubenswrapper[4809]: I1206 07:06:29.016064 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jwmsn" Dec 06 07:06:29 crc kubenswrapper[4809]: I1206 07:06:29.076666 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hptq2\" (UniqueName: \"kubernetes.io/projected/ea34817a-9f07-44a7-b57f-7c9940de89b0-kube-api-access-hptq2\") pod \"ea34817a-9f07-44a7-b57f-7c9940de89b0\" (UID: \"ea34817a-9f07-44a7-b57f-7c9940de89b0\") " Dec 06 07:06:29 crc kubenswrapper[4809]: I1206 07:06:29.076967 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea34817a-9f07-44a7-b57f-7c9940de89b0-utilities\") pod \"ea34817a-9f07-44a7-b57f-7c9940de89b0\" (UID: \"ea34817a-9f07-44a7-b57f-7c9940de89b0\") " Dec 06 07:06:29 crc kubenswrapper[4809]: I1206 07:06:29.077173 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea34817a-9f07-44a7-b57f-7c9940de89b0-catalog-content\") pod \"ea34817a-9f07-44a7-b57f-7c9940de89b0\" (UID: \"ea34817a-9f07-44a7-b57f-7c9940de89b0\") " Dec 06 07:06:29 crc kubenswrapper[4809]: I1206 07:06:29.077693 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea34817a-9f07-44a7-b57f-7c9940de89b0-utilities" (OuterVolumeSpecName: "utilities") pod "ea34817a-9f07-44a7-b57f-7c9940de89b0" (UID: "ea34817a-9f07-44a7-b57f-7c9940de89b0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:06:29 crc kubenswrapper[4809]: I1206 07:06:29.094822 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea34817a-9f07-44a7-b57f-7c9940de89b0-kube-api-access-hptq2" (OuterVolumeSpecName: "kube-api-access-hptq2") pod "ea34817a-9f07-44a7-b57f-7c9940de89b0" (UID: "ea34817a-9f07-44a7-b57f-7c9940de89b0"). InnerVolumeSpecName "kube-api-access-hptq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:06:29 crc kubenswrapper[4809]: I1206 07:06:29.100847 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea34817a-9f07-44a7-b57f-7c9940de89b0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ea34817a-9f07-44a7-b57f-7c9940de89b0" (UID: "ea34817a-9f07-44a7-b57f-7c9940de89b0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:06:29 crc kubenswrapper[4809]: I1206 07:06:29.179216 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hptq2\" (UniqueName: \"kubernetes.io/projected/ea34817a-9f07-44a7-b57f-7c9940de89b0-kube-api-access-hptq2\") on node \"crc\" DevicePath \"\"" Dec 06 07:06:29 crc kubenswrapper[4809]: I1206 07:06:29.179244 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea34817a-9f07-44a7-b57f-7c9940de89b0-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 07:06:29 crc kubenswrapper[4809]: I1206 07:06:29.179254 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea34817a-9f07-44a7-b57f-7c9940de89b0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 07:06:29 crc kubenswrapper[4809]: I1206 07:06:29.190075 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jwmsn" event={"ID":"ea34817a-9f07-44a7-b57f-7c9940de89b0","Type":"ContainerDied","Data":"bd34a895b1300e1ca2016d20a8d56836d105b0ccb26a37eaccd699d92df547df"} Dec 06 07:06:29 crc kubenswrapper[4809]: I1206 07:06:29.190338 4809 scope.go:117] "RemoveContainer" containerID="0cd351db33eb94912c7d67298ccd4390dca2c4b2f1d237e74ced92fe7a04ffcf" Dec 06 07:06:29 crc kubenswrapper[4809]: I1206 07:06:29.190178 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jwmsn" Dec 06 07:06:29 crc kubenswrapper[4809]: I1206 07:06:29.210102 4809 scope.go:117] "RemoveContainer" containerID="dee5480ad59364efab3a57cb85fa132511ac362f70d0c41854f573f23e3b2f20" Dec 06 07:06:29 crc kubenswrapper[4809]: I1206 07:06:29.239459 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jwmsn"] Dec 06 07:06:29 crc kubenswrapper[4809]: I1206 07:06:29.253655 4809 scope.go:117] "RemoveContainer" containerID="dea5281f86a76991a16957a5054ded528256384a9998079f7ff42dcfb8a6d4cc" Dec 06 07:06:29 crc kubenswrapper[4809]: I1206 07:06:29.256198 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jwmsn"] Dec 06 07:06:29 crc kubenswrapper[4809]: I1206 07:06:29.402325 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea34817a-9f07-44a7-b57f-7c9940de89b0" path="/var/lib/kubelet/pods/ea34817a-9f07-44a7-b57f-7c9940de89b0/volumes" Dec 06 07:06:40 crc kubenswrapper[4809]: I1206 07:06:40.389064 4809 scope.go:117] "RemoveContainer" containerID="474210a91ed30ed36d67639fb33cd7c97c7a14000a24e601812704bd7b93c234" Dec 06 07:06:40 crc kubenswrapper[4809]: E1206 07:06:40.391711 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:06:53 crc kubenswrapper[4809]: I1206 07:06:53.390805 4809 scope.go:117] "RemoveContainer" containerID="474210a91ed30ed36d67639fb33cd7c97c7a14000a24e601812704bd7b93c234" Dec 06 07:06:53 crc kubenswrapper[4809]: E1206 07:06:53.391697 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:06:58 crc kubenswrapper[4809]: I1206 07:06:58.629466 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-v8k7r"] Dec 06 07:06:58 crc kubenswrapper[4809]: E1206 07:06:58.630634 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea34817a-9f07-44a7-b57f-7c9940de89b0" containerName="registry-server" Dec 06 07:06:58 crc kubenswrapper[4809]: I1206 07:06:58.630653 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea34817a-9f07-44a7-b57f-7c9940de89b0" containerName="registry-server" Dec 06 07:06:58 crc kubenswrapper[4809]: E1206 07:06:58.630710 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea34817a-9f07-44a7-b57f-7c9940de89b0" containerName="extract-utilities" Dec 06 07:06:58 crc kubenswrapper[4809]: I1206 07:06:58.630719 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea34817a-9f07-44a7-b57f-7c9940de89b0" containerName="extract-utilities" Dec 06 07:06:58 crc kubenswrapper[4809]: E1206 07:06:58.630729 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea34817a-9f07-44a7-b57f-7c9940de89b0" containerName="extract-content" Dec 06 07:06:58 crc kubenswrapper[4809]: I1206 07:06:58.630736 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea34817a-9f07-44a7-b57f-7c9940de89b0" containerName="extract-content" Dec 06 07:06:58 crc kubenswrapper[4809]: I1206 07:06:58.631088 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea34817a-9f07-44a7-b57f-7c9940de89b0" containerName="registry-server" Dec 06 07:06:58 crc kubenswrapper[4809]: I1206 07:06:58.633651 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v8k7r" Dec 06 07:06:58 crc kubenswrapper[4809]: I1206 07:06:58.641907 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v8k7r"] Dec 06 07:06:58 crc kubenswrapper[4809]: I1206 07:06:58.785133 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6m85\" (UniqueName: \"kubernetes.io/projected/3bee09dc-6321-4073-8c73-5fd3dcaa7e3a-kube-api-access-n6m85\") pod \"certified-operators-v8k7r\" (UID: \"3bee09dc-6321-4073-8c73-5fd3dcaa7e3a\") " pod="openshift-marketplace/certified-operators-v8k7r" Dec 06 07:06:58 crc kubenswrapper[4809]: I1206 07:06:58.785686 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bee09dc-6321-4073-8c73-5fd3dcaa7e3a-utilities\") pod \"certified-operators-v8k7r\" (UID: \"3bee09dc-6321-4073-8c73-5fd3dcaa7e3a\") " pod="openshift-marketplace/certified-operators-v8k7r" Dec 06 07:06:58 crc kubenswrapper[4809]: I1206 07:06:58.785790 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bee09dc-6321-4073-8c73-5fd3dcaa7e3a-catalog-content\") pod \"certified-operators-v8k7r\" (UID: \"3bee09dc-6321-4073-8c73-5fd3dcaa7e3a\") " pod="openshift-marketplace/certified-operators-v8k7r" Dec 06 07:06:58 crc kubenswrapper[4809]: I1206 07:06:58.889485 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bee09dc-6321-4073-8c73-5fd3dcaa7e3a-utilities\") pod \"certified-operators-v8k7r\" (UID: \"3bee09dc-6321-4073-8c73-5fd3dcaa7e3a\") " pod="openshift-marketplace/certified-operators-v8k7r" Dec 06 07:06:58 crc kubenswrapper[4809]: I1206 07:06:58.889617 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bee09dc-6321-4073-8c73-5fd3dcaa7e3a-catalog-content\") pod \"certified-operators-v8k7r\" (UID: \"3bee09dc-6321-4073-8c73-5fd3dcaa7e3a\") " pod="openshift-marketplace/certified-operators-v8k7r" Dec 06 07:06:58 crc kubenswrapper[4809]: I1206 07:06:58.889794 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6m85\" (UniqueName: \"kubernetes.io/projected/3bee09dc-6321-4073-8c73-5fd3dcaa7e3a-kube-api-access-n6m85\") pod \"certified-operators-v8k7r\" (UID: \"3bee09dc-6321-4073-8c73-5fd3dcaa7e3a\") " pod="openshift-marketplace/certified-operators-v8k7r" Dec 06 07:06:58 crc kubenswrapper[4809]: I1206 07:06:58.890196 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bee09dc-6321-4073-8c73-5fd3dcaa7e3a-catalog-content\") pod \"certified-operators-v8k7r\" (UID: \"3bee09dc-6321-4073-8c73-5fd3dcaa7e3a\") " pod="openshift-marketplace/certified-operators-v8k7r" Dec 06 07:06:58 crc kubenswrapper[4809]: I1206 07:06:58.890400 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bee09dc-6321-4073-8c73-5fd3dcaa7e3a-utilities\") pod \"certified-operators-v8k7r\" (UID: \"3bee09dc-6321-4073-8c73-5fd3dcaa7e3a\") " pod="openshift-marketplace/certified-operators-v8k7r" Dec 06 07:06:58 crc kubenswrapper[4809]: I1206 07:06:58.924514 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6m85\" (UniqueName: \"kubernetes.io/projected/3bee09dc-6321-4073-8c73-5fd3dcaa7e3a-kube-api-access-n6m85\") pod \"certified-operators-v8k7r\" (UID: \"3bee09dc-6321-4073-8c73-5fd3dcaa7e3a\") " pod="openshift-marketplace/certified-operators-v8k7r" Dec 06 07:06:58 crc kubenswrapper[4809]: I1206 07:06:58.972745 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v8k7r" Dec 06 07:06:59 crc kubenswrapper[4809]: I1206 07:06:59.284508 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v8k7r"] Dec 06 07:06:59 crc kubenswrapper[4809]: I1206 07:06:59.519344 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v8k7r" event={"ID":"3bee09dc-6321-4073-8c73-5fd3dcaa7e3a","Type":"ContainerStarted","Data":"d273a64de1f7160aefea7a3c3722ac5d5010a50558189951c730b92f5d95121a"} Dec 06 07:07:00 crc kubenswrapper[4809]: I1206 07:07:00.534400 4809 generic.go:334] "Generic (PLEG): container finished" podID="3bee09dc-6321-4073-8c73-5fd3dcaa7e3a" containerID="e916b7863f23eacee4c554aaa3a1a4fc90063d87d4e35acf82421c84a0e03d9e" exitCode=0 Dec 06 07:07:00 crc kubenswrapper[4809]: I1206 07:07:00.534442 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v8k7r" event={"ID":"3bee09dc-6321-4073-8c73-5fd3dcaa7e3a","Type":"ContainerDied","Data":"e916b7863f23eacee4c554aaa3a1a4fc90063d87d4e35acf82421c84a0e03d9e"} Dec 06 07:07:01 crc kubenswrapper[4809]: I1206 07:07:01.549175 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v8k7r" event={"ID":"3bee09dc-6321-4073-8c73-5fd3dcaa7e3a","Type":"ContainerStarted","Data":"9b5b2ff50e8b6159f14825e5801c94f144bca7c2cb6670208db2980563852fcf"} Dec 06 07:07:02 crc kubenswrapper[4809]: I1206 07:07:02.572504 4809 generic.go:334] "Generic (PLEG): container finished" podID="3bee09dc-6321-4073-8c73-5fd3dcaa7e3a" containerID="9b5b2ff50e8b6159f14825e5801c94f144bca7c2cb6670208db2980563852fcf" exitCode=0 Dec 06 07:07:02 crc kubenswrapper[4809]: I1206 07:07:02.572879 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v8k7r" event={"ID":"3bee09dc-6321-4073-8c73-5fd3dcaa7e3a","Type":"ContainerDied","Data":"9b5b2ff50e8b6159f14825e5801c94f144bca7c2cb6670208db2980563852fcf"} Dec 06 07:07:04 crc kubenswrapper[4809]: I1206 07:07:04.389144 4809 scope.go:117] "RemoveContainer" containerID="474210a91ed30ed36d67639fb33cd7c97c7a14000a24e601812704bd7b93c234" Dec 06 07:07:04 crc kubenswrapper[4809]: E1206 07:07:04.389984 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:07:04 crc kubenswrapper[4809]: I1206 07:07:04.596391 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v8k7r" event={"ID":"3bee09dc-6321-4073-8c73-5fd3dcaa7e3a","Type":"ContainerStarted","Data":"a8b6e3d53a8b5d348094849f2735aa7bc1b384fe5b5b2941c2ec4884e80da639"} Dec 06 07:07:04 crc kubenswrapper[4809]: I1206 07:07:04.623544 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-v8k7r" podStartSLOduration=4.123457687 podStartE2EDuration="6.623526581s" podCreationTimestamp="2025-12-06 07:06:58 +0000 UTC" firstStartedPulling="2025-12-06 07:07:00.537565848 +0000 UTC m=+4545.426548830" lastFinishedPulling="2025-12-06 07:07:03.037634782 +0000 UTC m=+4547.926617724" observedRunningTime="2025-12-06 07:07:04.615164025 +0000 UTC m=+4549.504146967" watchObservedRunningTime="2025-12-06 07:07:04.623526581 +0000 UTC m=+4549.512509523" Dec 06 07:07:08 crc kubenswrapper[4809]: I1206 07:07:08.972973 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-v8k7r" Dec 06 07:07:08 crc kubenswrapper[4809]: I1206 07:07:08.973617 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-v8k7r" Dec 06 07:07:09 crc kubenswrapper[4809]: I1206 07:07:09.121589 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-v8k7r" Dec 06 07:07:09 crc kubenswrapper[4809]: I1206 07:07:09.707683 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-v8k7r" Dec 06 07:07:13 crc kubenswrapper[4809]: I1206 07:07:13.024401 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v8k7r"] Dec 06 07:07:13 crc kubenswrapper[4809]: I1206 07:07:13.025444 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-v8k7r" podUID="3bee09dc-6321-4073-8c73-5fd3dcaa7e3a" containerName="registry-server" containerID="cri-o://a8b6e3d53a8b5d348094849f2735aa7bc1b384fe5b5b2941c2ec4884e80da639" gracePeriod=2 Dec 06 07:07:13 crc kubenswrapper[4809]: I1206 07:07:13.541359 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v8k7r" Dec 06 07:07:13 crc kubenswrapper[4809]: I1206 07:07:13.677990 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6m85\" (UniqueName: \"kubernetes.io/projected/3bee09dc-6321-4073-8c73-5fd3dcaa7e3a-kube-api-access-n6m85\") pod \"3bee09dc-6321-4073-8c73-5fd3dcaa7e3a\" (UID: \"3bee09dc-6321-4073-8c73-5fd3dcaa7e3a\") " Dec 06 07:07:13 crc kubenswrapper[4809]: I1206 07:07:13.678062 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bee09dc-6321-4073-8c73-5fd3dcaa7e3a-catalog-content\") pod \"3bee09dc-6321-4073-8c73-5fd3dcaa7e3a\" (UID: \"3bee09dc-6321-4073-8c73-5fd3dcaa7e3a\") " Dec 06 07:07:13 crc kubenswrapper[4809]: I1206 07:07:13.678351 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bee09dc-6321-4073-8c73-5fd3dcaa7e3a-utilities\") pod \"3bee09dc-6321-4073-8c73-5fd3dcaa7e3a\" (UID: \"3bee09dc-6321-4073-8c73-5fd3dcaa7e3a\") " Dec 06 07:07:13 crc kubenswrapper[4809]: I1206 07:07:13.679784 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3bee09dc-6321-4073-8c73-5fd3dcaa7e3a-utilities" (OuterVolumeSpecName: "utilities") pod "3bee09dc-6321-4073-8c73-5fd3dcaa7e3a" (UID: "3bee09dc-6321-4073-8c73-5fd3dcaa7e3a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:07:13 crc kubenswrapper[4809]: I1206 07:07:13.684640 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bee09dc-6321-4073-8c73-5fd3dcaa7e3a-kube-api-access-n6m85" (OuterVolumeSpecName: "kube-api-access-n6m85") pod "3bee09dc-6321-4073-8c73-5fd3dcaa7e3a" (UID: "3bee09dc-6321-4073-8c73-5fd3dcaa7e3a"). InnerVolumeSpecName "kube-api-access-n6m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:07:13 crc kubenswrapper[4809]: I1206 07:07:13.691904 4809 generic.go:334] "Generic (PLEG): container finished" podID="3bee09dc-6321-4073-8c73-5fd3dcaa7e3a" containerID="a8b6e3d53a8b5d348094849f2735aa7bc1b384fe5b5b2941c2ec4884e80da639" exitCode=0 Dec 06 07:07:13 crc kubenswrapper[4809]: I1206 07:07:13.691958 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v8k7r" event={"ID":"3bee09dc-6321-4073-8c73-5fd3dcaa7e3a","Type":"ContainerDied","Data":"a8b6e3d53a8b5d348094849f2735aa7bc1b384fe5b5b2941c2ec4884e80da639"} Dec 06 07:07:13 crc kubenswrapper[4809]: I1206 07:07:13.691985 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v8k7r" event={"ID":"3bee09dc-6321-4073-8c73-5fd3dcaa7e3a","Type":"ContainerDied","Data":"d273a64de1f7160aefea7a3c3722ac5d5010a50558189951c730b92f5d95121a"} Dec 06 07:07:13 crc kubenswrapper[4809]: I1206 07:07:13.692004 4809 scope.go:117] "RemoveContainer" containerID="a8b6e3d53a8b5d348094849f2735aa7bc1b384fe5b5b2941c2ec4884e80da639" Dec 06 07:07:13 crc kubenswrapper[4809]: I1206 07:07:13.692150 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v8k7r" Dec 06 07:07:13 crc kubenswrapper[4809]: I1206 07:07:13.724371 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3bee09dc-6321-4073-8c73-5fd3dcaa7e3a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3bee09dc-6321-4073-8c73-5fd3dcaa7e3a" (UID: "3bee09dc-6321-4073-8c73-5fd3dcaa7e3a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:07:13 crc kubenswrapper[4809]: I1206 07:07:13.756508 4809 scope.go:117] "RemoveContainer" containerID="9b5b2ff50e8b6159f14825e5801c94f144bca7c2cb6670208db2980563852fcf" Dec 06 07:07:13 crc kubenswrapper[4809]: I1206 07:07:13.781046 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bee09dc-6321-4073-8c73-5fd3dcaa7e3a-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 07:07:13 crc kubenswrapper[4809]: I1206 07:07:13.781083 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6m85\" (UniqueName: \"kubernetes.io/projected/3bee09dc-6321-4073-8c73-5fd3dcaa7e3a-kube-api-access-n6m85\") on node \"crc\" DevicePath \"\"" Dec 06 07:07:13 crc kubenswrapper[4809]: I1206 07:07:13.781093 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bee09dc-6321-4073-8c73-5fd3dcaa7e3a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 07:07:13 crc kubenswrapper[4809]: I1206 07:07:13.797795 4809 scope.go:117] "RemoveContainer" containerID="e916b7863f23eacee4c554aaa3a1a4fc90063d87d4e35acf82421c84a0e03d9e" Dec 06 07:07:13 crc kubenswrapper[4809]: I1206 07:07:13.836621 4809 scope.go:117] "RemoveContainer" containerID="a8b6e3d53a8b5d348094849f2735aa7bc1b384fe5b5b2941c2ec4884e80da639" Dec 06 07:07:13 crc kubenswrapper[4809]: E1206 07:07:13.837227 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8b6e3d53a8b5d348094849f2735aa7bc1b384fe5b5b2941c2ec4884e80da639\": container with ID starting with a8b6e3d53a8b5d348094849f2735aa7bc1b384fe5b5b2941c2ec4884e80da639 not found: ID does not exist" containerID="a8b6e3d53a8b5d348094849f2735aa7bc1b384fe5b5b2941c2ec4884e80da639" Dec 06 07:07:13 crc kubenswrapper[4809]: I1206 07:07:13.837278 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8b6e3d53a8b5d348094849f2735aa7bc1b384fe5b5b2941c2ec4884e80da639"} err="failed to get container status \"a8b6e3d53a8b5d348094849f2735aa7bc1b384fe5b5b2941c2ec4884e80da639\": rpc error: code = NotFound desc = could not find container \"a8b6e3d53a8b5d348094849f2735aa7bc1b384fe5b5b2941c2ec4884e80da639\": container with ID starting with a8b6e3d53a8b5d348094849f2735aa7bc1b384fe5b5b2941c2ec4884e80da639 not found: ID does not exist" Dec 06 07:07:13 crc kubenswrapper[4809]: I1206 07:07:13.837305 4809 scope.go:117] "RemoveContainer" containerID="9b5b2ff50e8b6159f14825e5801c94f144bca7c2cb6670208db2980563852fcf" Dec 06 07:07:13 crc kubenswrapper[4809]: E1206 07:07:13.837680 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b5b2ff50e8b6159f14825e5801c94f144bca7c2cb6670208db2980563852fcf\": container with ID starting with 9b5b2ff50e8b6159f14825e5801c94f144bca7c2cb6670208db2980563852fcf not found: ID does not exist" containerID="9b5b2ff50e8b6159f14825e5801c94f144bca7c2cb6670208db2980563852fcf" Dec 06 07:07:13 crc kubenswrapper[4809]: I1206 07:07:13.837711 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b5b2ff50e8b6159f14825e5801c94f144bca7c2cb6670208db2980563852fcf"} err="failed to get container status \"9b5b2ff50e8b6159f14825e5801c94f144bca7c2cb6670208db2980563852fcf\": rpc error: code = NotFound desc = could not find container \"9b5b2ff50e8b6159f14825e5801c94f144bca7c2cb6670208db2980563852fcf\": container with ID starting with 9b5b2ff50e8b6159f14825e5801c94f144bca7c2cb6670208db2980563852fcf not found: ID does not exist" Dec 06 07:07:13 crc kubenswrapper[4809]: I1206 07:07:13.837733 4809 scope.go:117] "RemoveContainer" containerID="e916b7863f23eacee4c554aaa3a1a4fc90063d87d4e35acf82421c84a0e03d9e" Dec 06 07:07:13 crc kubenswrapper[4809]: E1206 07:07:13.837925 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e916b7863f23eacee4c554aaa3a1a4fc90063d87d4e35acf82421c84a0e03d9e\": container with ID starting with e916b7863f23eacee4c554aaa3a1a4fc90063d87d4e35acf82421c84a0e03d9e not found: ID does not exist" containerID="e916b7863f23eacee4c554aaa3a1a4fc90063d87d4e35acf82421c84a0e03d9e" Dec 06 07:07:13 crc kubenswrapper[4809]: I1206 07:07:13.837959 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e916b7863f23eacee4c554aaa3a1a4fc90063d87d4e35acf82421c84a0e03d9e"} err="failed to get container status \"e916b7863f23eacee4c554aaa3a1a4fc90063d87d4e35acf82421c84a0e03d9e\": rpc error: code = NotFound desc = could not find container \"e916b7863f23eacee4c554aaa3a1a4fc90063d87d4e35acf82421c84a0e03d9e\": container with ID starting with e916b7863f23eacee4c554aaa3a1a4fc90063d87d4e35acf82421c84a0e03d9e not found: ID does not exist" Dec 06 07:07:14 crc kubenswrapper[4809]: I1206 07:07:14.032358 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v8k7r"] Dec 06 07:07:14 crc kubenswrapper[4809]: I1206 07:07:14.044183 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-v8k7r"] Dec 06 07:07:15 crc kubenswrapper[4809]: I1206 07:07:15.410370 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3bee09dc-6321-4073-8c73-5fd3dcaa7e3a" path="/var/lib/kubelet/pods/3bee09dc-6321-4073-8c73-5fd3dcaa7e3a/volumes" Dec 06 07:07:19 crc kubenswrapper[4809]: I1206 07:07:19.388885 4809 scope.go:117] "RemoveContainer" containerID="474210a91ed30ed36d67639fb33cd7c97c7a14000a24e601812704bd7b93c234" Dec 06 07:07:19 crc kubenswrapper[4809]: E1206 07:07:19.389818 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:07:34 crc kubenswrapper[4809]: I1206 07:07:34.389451 4809 scope.go:117] "RemoveContainer" containerID="474210a91ed30ed36d67639fb33cd7c97c7a14000a24e601812704bd7b93c234" Dec 06 07:07:34 crc kubenswrapper[4809]: E1206 07:07:34.390564 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:07:47 crc kubenswrapper[4809]: I1206 07:07:47.394364 4809 scope.go:117] "RemoveContainer" containerID="474210a91ed30ed36d67639fb33cd7c97c7a14000a24e601812704bd7b93c234" Dec 06 07:07:47 crc kubenswrapper[4809]: E1206 07:07:47.395556 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:07:59 crc kubenswrapper[4809]: I1206 07:07:59.388244 4809 scope.go:117] "RemoveContainer" containerID="474210a91ed30ed36d67639fb33cd7c97c7a14000a24e601812704bd7b93c234" Dec 06 07:07:59 crc kubenswrapper[4809]: E1206 07:07:59.388971 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:08:11 crc kubenswrapper[4809]: I1206 07:08:11.389625 4809 scope.go:117] "RemoveContainer" containerID="474210a91ed30ed36d67639fb33cd7c97c7a14000a24e601812704bd7b93c234" Dec 06 07:08:11 crc kubenswrapper[4809]: E1206 07:08:11.391464 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:08:24 crc kubenswrapper[4809]: I1206 07:08:24.388466 4809 scope.go:117] "RemoveContainer" containerID="474210a91ed30ed36d67639fb33cd7c97c7a14000a24e601812704bd7b93c234" Dec 06 07:08:24 crc kubenswrapper[4809]: E1206 07:08:24.389470 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:08:37 crc kubenswrapper[4809]: I1206 07:08:37.389254 4809 scope.go:117] "RemoveContainer" containerID="474210a91ed30ed36d67639fb33cd7c97c7a14000a24e601812704bd7b93c234" Dec 06 07:08:37 crc kubenswrapper[4809]: E1206 07:08:37.390141 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:08:48 crc kubenswrapper[4809]: I1206 07:08:48.389324 4809 scope.go:117] "RemoveContainer" containerID="474210a91ed30ed36d67639fb33cd7c97c7a14000a24e601812704bd7b93c234" Dec 06 07:08:48 crc kubenswrapper[4809]: E1206 07:08:48.390240 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:09:01 crc kubenswrapper[4809]: I1206 07:09:01.389280 4809 scope.go:117] "RemoveContainer" containerID="474210a91ed30ed36d67639fb33cd7c97c7a14000a24e601812704bd7b93c234" Dec 06 07:09:01 crc kubenswrapper[4809]: E1206 07:09:01.390192 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:09:13 crc kubenswrapper[4809]: I1206 07:09:13.394058 4809 scope.go:117] "RemoveContainer" containerID="474210a91ed30ed36d67639fb33cd7c97c7a14000a24e601812704bd7b93c234" Dec 06 07:09:13 crc kubenswrapper[4809]: E1206 07:09:13.397089 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:09:25 crc kubenswrapper[4809]: I1206 07:09:25.399456 4809 scope.go:117] "RemoveContainer" containerID="474210a91ed30ed36d67639fb33cd7c97c7a14000a24e601812704bd7b93c234" Dec 06 07:09:25 crc kubenswrapper[4809]: E1206 07:09:25.400209 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:09:37 crc kubenswrapper[4809]: I1206 07:09:37.389621 4809 scope.go:117] "RemoveContainer" containerID="474210a91ed30ed36d67639fb33cd7c97c7a14000a24e601812704bd7b93c234" Dec 06 07:09:37 crc kubenswrapper[4809]: E1206 07:09:37.390483 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:09:49 crc kubenswrapper[4809]: I1206 07:09:49.393736 4809 scope.go:117] "RemoveContainer" containerID="474210a91ed30ed36d67639fb33cd7c97c7a14000a24e601812704bd7b93c234" Dec 06 07:09:49 crc kubenswrapper[4809]: E1206 07:09:49.394710 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:10:03 crc kubenswrapper[4809]: I1206 07:10:03.389191 4809 scope.go:117] "RemoveContainer" containerID="474210a91ed30ed36d67639fb33cd7c97c7a14000a24e601812704bd7b93c234" Dec 06 07:10:03 crc kubenswrapper[4809]: E1206 07:10:03.389721 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:10:15 crc kubenswrapper[4809]: I1206 07:10:15.405037 4809 scope.go:117] "RemoveContainer" containerID="474210a91ed30ed36d67639fb33cd7c97c7a14000a24e601812704bd7b93c234" Dec 06 07:10:15 crc kubenswrapper[4809]: E1206 07:10:15.405835 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:10:30 crc kubenswrapper[4809]: I1206 07:10:30.388234 4809 scope.go:117] "RemoveContainer" containerID="474210a91ed30ed36d67639fb33cd7c97c7a14000a24e601812704bd7b93c234" Dec 06 07:10:30 crc kubenswrapper[4809]: E1206 07:10:30.389186 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:10:44 crc kubenswrapper[4809]: I1206 07:10:44.389592 4809 scope.go:117] "RemoveContainer" containerID="474210a91ed30ed36d67639fb33cd7c97c7a14000a24e601812704bd7b93c234" Dec 06 07:10:46 crc kubenswrapper[4809]: I1206 07:10:46.139066 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" event={"ID":"cbf4cf62-024e-4703-a8b8-9aecda9cd26a","Type":"ContainerStarted","Data":"9ddac882b8b7232d7d5f4ba25a08bfffc6501c262ddc8c7ab993417cc0299ce0"} Dec 06 07:12:16 crc kubenswrapper[4809]: I1206 07:12:16.167206 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/telemetry-operator-controller-manager-b4644b845-zxwvr" podUID="24aed9fd-d792-4bce-82a8-3ea4d70094b4" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.117:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:12:23 crc kubenswrapper[4809]: I1206 07:12:23.782729 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Dec 06 07:12:24 crc kubenswrapper[4809]: I1206 07:12:24.031376 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerName="ceilometer-central-agent" probeResult="failure" output=< Dec 06 07:12:24 crc kubenswrapper[4809]: Unkown error: Expecting value: line 1 column 1 (char 0) Dec 06 07:12:24 crc kubenswrapper[4809]: > Dec 06 07:12:28 crc kubenswrapper[4809]: I1206 07:12:28.094916 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerName="ceilometer-central-agent" probeResult="failure" output=< Dec 06 07:12:28 crc kubenswrapper[4809]: Unkown error: Expecting value: line 1 column 1 (char 0) Dec 06 07:12:28 crc kubenswrapper[4809]: > Dec 06 07:12:28 crc kubenswrapper[4809]: I1206 07:12:28.095481 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/ceilometer-0" Dec 06 07:12:28 crc kubenswrapper[4809]: I1206 07:12:28.096538 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="ceilometer-central-agent" containerStatusID={"Type":"cri-o","ID":"92f800a6ae2ed5e64c1afe14ad830a380ec334e44b69d5e189323d25b3c22531"} pod="openstack/ceilometer-0" containerMessage="Container ceilometer-central-agent failed liveness probe, will be restarted" Dec 06 07:12:28 crc kubenswrapper[4809]: I1206 07:12:28.096605 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerName="ceilometer-central-agent" containerID="cri-o://92f800a6ae2ed5e64c1afe14ad830a380ec334e44b69d5e189323d25b3c22531" gracePeriod=30 Dec 06 07:12:29 crc kubenswrapper[4809]: I1206 07:12:29.857363 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 07:12:30 crc kubenswrapper[4809]: I1206 07:12:30.406405 4809 generic.go:334] "Generic (PLEG): container finished" podID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerID="92f800a6ae2ed5e64c1afe14ad830a380ec334e44b69d5e189323d25b3c22531" exitCode=0 Dec 06 07:12:30 crc kubenswrapper[4809]: I1206 07:12:30.406498 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3d4b42d3-04a4-40e3-803b-11c77024f612","Type":"ContainerDied","Data":"92f800a6ae2ed5e64c1afe14ad830a380ec334e44b69d5e189323d25b3c22531"} Dec 06 07:12:30 crc kubenswrapper[4809]: I1206 07:12:30.406727 4809 scope.go:117] "RemoveContainer" containerID="644e6da2ea964da6dc3103aea27cc874c7220fb372f243ee56009f3395b2d887" Dec 06 07:12:31 crc kubenswrapper[4809]: I1206 07:12:31.419043 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3d4b42d3-04a4-40e3-803b-11c77024f612","Type":"ContainerStarted","Data":"1ed2ac2be19d734f28ddb67018abd7ca0033f58827c7e4ab6c9c7cfe52be430c"} Dec 06 07:13:04 crc kubenswrapper[4809]: I1206 07:13:04.496117 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:13:04 crc kubenswrapper[4809]: I1206 07:13:04.496789 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:13:34 crc kubenswrapper[4809]: I1206 07:13:34.496130 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:13:34 crc kubenswrapper[4809]: I1206 07:13:34.496744 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:14:04 crc kubenswrapper[4809]: I1206 07:14:04.497069 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:14:04 crc kubenswrapper[4809]: I1206 07:14:04.498001 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:14:04 crc kubenswrapper[4809]: I1206 07:14:04.498062 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-npms2" Dec 06 07:14:04 crc kubenswrapper[4809]: I1206 07:14:04.499120 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9ddac882b8b7232d7d5f4ba25a08bfffc6501c262ddc8c7ab993417cc0299ce0"} pod="openshift-machine-config-operator/machine-config-daemon-npms2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 07:14:04 crc kubenswrapper[4809]: I1206 07:14:04.499186 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" containerID="cri-o://9ddac882b8b7232d7d5f4ba25a08bfffc6501c262ddc8c7ab993417cc0299ce0" gracePeriod=600 Dec 06 07:14:04 crc kubenswrapper[4809]: I1206 07:14:04.978489 4809 generic.go:334] "Generic (PLEG): container finished" podID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerID="9ddac882b8b7232d7d5f4ba25a08bfffc6501c262ddc8c7ab993417cc0299ce0" exitCode=0 Dec 06 07:14:04 crc kubenswrapper[4809]: I1206 07:14:04.978539 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" event={"ID":"cbf4cf62-024e-4703-a8b8-9aecda9cd26a","Type":"ContainerDied","Data":"9ddac882b8b7232d7d5f4ba25a08bfffc6501c262ddc8c7ab993417cc0299ce0"} Dec 06 07:14:04 crc kubenswrapper[4809]: I1206 07:14:04.978812 4809 scope.go:117] "RemoveContainer" containerID="474210a91ed30ed36d67639fb33cd7c97c7a14000a24e601812704bd7b93c234" Dec 06 07:14:10 crc kubenswrapper[4809]: I1206 07:14:10.511224 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="cert-manager/cert-manager-webhook-5655c58dd6-7vvw8" podUID="886a998b-a62e-412c-ae50-fd1ba3a4a098" containerName="cert-manager-webhook" probeResult="failure" output="Get \"http://10.217.0.44:6080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:14:10 crc kubenswrapper[4809]: I1206 07:14:10.511320 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/controller-f8648f98b-d4fgg" podUID="073938c5-bc80-47d3-91e8-1d6d7da0f000" containerName="controller" probeResult="failure" output="Get \"http://10.217.0.95:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:14:13 crc kubenswrapper[4809]: I1206 07:14:13.781919 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerName="ceilometer-notification-agent" probeResult="failure" output="command timed out" Dec 06 07:14:15 crc kubenswrapper[4809]: I1206 07:14:15.659229 4809 patch_prober.go:28] interesting pod/router-default-5444994796-n6z57 container/router namespace/openshift-ingress: Liveness probe status=failure output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:14:15 crc kubenswrapper[4809]: I1206 07:14:15.659641 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-ingress/router-default-5444994796-n6z57" podUID="e5356f1d-5afd-47bd-89af-72abfa8f901f" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:14:17 crc kubenswrapper[4809]: I1206 07:14:17.146693 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" event={"ID":"cbf4cf62-024e-4703-a8b8-9aecda9cd26a","Type":"ContainerStarted","Data":"af73883c3be3c80377e0f4806f2ecfa3ce57913ac4edf4c8383f744be4172b66"} Dec 06 07:15:00 crc kubenswrapper[4809]: I1206 07:15:00.156141 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416755-jq7qg"] Dec 06 07:15:00 crc kubenswrapper[4809]: E1206 07:15:00.157313 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bee09dc-6321-4073-8c73-5fd3dcaa7e3a" containerName="extract-content" Dec 06 07:15:00 crc kubenswrapper[4809]: I1206 07:15:00.157333 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bee09dc-6321-4073-8c73-5fd3dcaa7e3a" containerName="extract-content" Dec 06 07:15:00 crc kubenswrapper[4809]: E1206 07:15:00.157353 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bee09dc-6321-4073-8c73-5fd3dcaa7e3a" containerName="extract-utilities" Dec 06 07:15:00 crc kubenswrapper[4809]: I1206 07:15:00.157361 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bee09dc-6321-4073-8c73-5fd3dcaa7e3a" containerName="extract-utilities" Dec 06 07:15:00 crc kubenswrapper[4809]: E1206 07:15:00.157420 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bee09dc-6321-4073-8c73-5fd3dcaa7e3a" containerName="registry-server" Dec 06 07:15:00 crc kubenswrapper[4809]: I1206 07:15:00.157429 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bee09dc-6321-4073-8c73-5fd3dcaa7e3a" containerName="registry-server" Dec 06 07:15:00 crc kubenswrapper[4809]: I1206 07:15:00.157679 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bee09dc-6321-4073-8c73-5fd3dcaa7e3a" containerName="registry-server" Dec 06 07:15:00 crc kubenswrapper[4809]: I1206 07:15:00.164025 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416755-jq7qg" Dec 06 07:15:00 crc kubenswrapper[4809]: I1206 07:15:00.168067 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 07:15:00 crc kubenswrapper[4809]: I1206 07:15:00.168527 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 07:15:00 crc kubenswrapper[4809]: I1206 07:15:00.191468 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416755-jq7qg"] Dec 06 07:15:00 crc kubenswrapper[4809]: I1206 07:15:00.298214 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ef6236c2-1bac-49bb-9ad7-642921a9dc35-config-volume\") pod \"collect-profiles-29416755-jq7qg\" (UID: \"ef6236c2-1bac-49bb-9ad7-642921a9dc35\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416755-jq7qg" Dec 06 07:15:00 crc kubenswrapper[4809]: I1206 07:15:00.298556 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmrh5\" (UniqueName: \"kubernetes.io/projected/ef6236c2-1bac-49bb-9ad7-642921a9dc35-kube-api-access-zmrh5\") pod \"collect-profiles-29416755-jq7qg\" (UID: \"ef6236c2-1bac-49bb-9ad7-642921a9dc35\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416755-jq7qg" Dec 06 07:15:00 crc kubenswrapper[4809]: I1206 07:15:00.298988 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ef6236c2-1bac-49bb-9ad7-642921a9dc35-secret-volume\") pod \"collect-profiles-29416755-jq7qg\" (UID: \"ef6236c2-1bac-49bb-9ad7-642921a9dc35\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416755-jq7qg" Dec 06 07:15:00 crc kubenswrapper[4809]: I1206 07:15:00.401633 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ef6236c2-1bac-49bb-9ad7-642921a9dc35-secret-volume\") pod \"collect-profiles-29416755-jq7qg\" (UID: \"ef6236c2-1bac-49bb-9ad7-642921a9dc35\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416755-jq7qg" Dec 06 07:15:00 crc kubenswrapper[4809]: I1206 07:15:00.401694 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ef6236c2-1bac-49bb-9ad7-642921a9dc35-config-volume\") pod \"collect-profiles-29416755-jq7qg\" (UID: \"ef6236c2-1bac-49bb-9ad7-642921a9dc35\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416755-jq7qg" Dec 06 07:15:00 crc kubenswrapper[4809]: I1206 07:15:00.401753 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmrh5\" (UniqueName: \"kubernetes.io/projected/ef6236c2-1bac-49bb-9ad7-642921a9dc35-kube-api-access-zmrh5\") pod \"collect-profiles-29416755-jq7qg\" (UID: \"ef6236c2-1bac-49bb-9ad7-642921a9dc35\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416755-jq7qg" Dec 06 07:15:00 crc kubenswrapper[4809]: I1206 07:15:00.403227 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ef6236c2-1bac-49bb-9ad7-642921a9dc35-config-volume\") pod \"collect-profiles-29416755-jq7qg\" (UID: \"ef6236c2-1bac-49bb-9ad7-642921a9dc35\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416755-jq7qg" Dec 06 07:15:00 crc kubenswrapper[4809]: I1206 07:15:00.408809 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ef6236c2-1bac-49bb-9ad7-642921a9dc35-secret-volume\") pod \"collect-profiles-29416755-jq7qg\" (UID: \"ef6236c2-1bac-49bb-9ad7-642921a9dc35\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416755-jq7qg" Dec 06 07:15:00 crc kubenswrapper[4809]: I1206 07:15:00.419481 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmrh5\" (UniqueName: \"kubernetes.io/projected/ef6236c2-1bac-49bb-9ad7-642921a9dc35-kube-api-access-zmrh5\") pod \"collect-profiles-29416755-jq7qg\" (UID: \"ef6236c2-1bac-49bb-9ad7-642921a9dc35\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416755-jq7qg" Dec 06 07:15:00 crc kubenswrapper[4809]: I1206 07:15:00.495415 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416755-jq7qg" Dec 06 07:15:01 crc kubenswrapper[4809]: I1206 07:15:01.040142 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416755-jq7qg"] Dec 06 07:15:01 crc kubenswrapper[4809]: I1206 07:15:01.650918 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416755-jq7qg" event={"ID":"ef6236c2-1bac-49bb-9ad7-642921a9dc35","Type":"ContainerStarted","Data":"83232476a9438e2c6c8a017430a796a52b8a241342e4ca8e0b860933c7fb678e"} Dec 06 07:15:02 crc kubenswrapper[4809]: I1206 07:15:02.662754 4809 generic.go:334] "Generic (PLEG): container finished" podID="ef6236c2-1bac-49bb-9ad7-642921a9dc35" containerID="4245cbd1539e6b09fd533abf15544b6fc3d70a5b5fb73d22d71224f4f56e3873" exitCode=0 Dec 06 07:15:02 crc kubenswrapper[4809]: I1206 07:15:02.662908 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416755-jq7qg" event={"ID":"ef6236c2-1bac-49bb-9ad7-642921a9dc35","Type":"ContainerDied","Data":"4245cbd1539e6b09fd533abf15544b6fc3d70a5b5fb73d22d71224f4f56e3873"} Dec 06 07:15:04 crc kubenswrapper[4809]: I1206 07:15:04.089823 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416755-jq7qg" Dec 06 07:15:04 crc kubenswrapper[4809]: I1206 07:15:04.193478 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmrh5\" (UniqueName: \"kubernetes.io/projected/ef6236c2-1bac-49bb-9ad7-642921a9dc35-kube-api-access-zmrh5\") pod \"ef6236c2-1bac-49bb-9ad7-642921a9dc35\" (UID: \"ef6236c2-1bac-49bb-9ad7-642921a9dc35\") " Dec 06 07:15:04 crc kubenswrapper[4809]: I1206 07:15:04.193713 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ef6236c2-1bac-49bb-9ad7-642921a9dc35-config-volume\") pod \"ef6236c2-1bac-49bb-9ad7-642921a9dc35\" (UID: \"ef6236c2-1bac-49bb-9ad7-642921a9dc35\") " Dec 06 07:15:04 crc kubenswrapper[4809]: I1206 07:15:04.193794 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ef6236c2-1bac-49bb-9ad7-642921a9dc35-secret-volume\") pod \"ef6236c2-1bac-49bb-9ad7-642921a9dc35\" (UID: \"ef6236c2-1bac-49bb-9ad7-642921a9dc35\") " Dec 06 07:15:04 crc kubenswrapper[4809]: I1206 07:15:04.194493 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef6236c2-1bac-49bb-9ad7-642921a9dc35-config-volume" (OuterVolumeSpecName: "config-volume") pod "ef6236c2-1bac-49bb-9ad7-642921a9dc35" (UID: "ef6236c2-1bac-49bb-9ad7-642921a9dc35"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:15:04 crc kubenswrapper[4809]: I1206 07:15:04.208194 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef6236c2-1bac-49bb-9ad7-642921a9dc35-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ef6236c2-1bac-49bb-9ad7-642921a9dc35" (UID: "ef6236c2-1bac-49bb-9ad7-642921a9dc35"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:15:04 crc kubenswrapper[4809]: I1206 07:15:04.208378 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef6236c2-1bac-49bb-9ad7-642921a9dc35-kube-api-access-zmrh5" (OuterVolumeSpecName: "kube-api-access-zmrh5") pod "ef6236c2-1bac-49bb-9ad7-642921a9dc35" (UID: "ef6236c2-1bac-49bb-9ad7-642921a9dc35"). InnerVolumeSpecName "kube-api-access-zmrh5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:15:04 crc kubenswrapper[4809]: I1206 07:15:04.296369 4809 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ef6236c2-1bac-49bb-9ad7-642921a9dc35-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:04 crc kubenswrapper[4809]: I1206 07:15:04.296409 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmrh5\" (UniqueName: \"kubernetes.io/projected/ef6236c2-1bac-49bb-9ad7-642921a9dc35-kube-api-access-zmrh5\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:04 crc kubenswrapper[4809]: I1206 07:15:04.296422 4809 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ef6236c2-1bac-49bb-9ad7-642921a9dc35-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 07:15:04 crc kubenswrapper[4809]: I1206 07:15:04.687139 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416755-jq7qg" event={"ID":"ef6236c2-1bac-49bb-9ad7-642921a9dc35","Type":"ContainerDied","Data":"83232476a9438e2c6c8a017430a796a52b8a241342e4ca8e0b860933c7fb678e"} Dec 06 07:15:04 crc kubenswrapper[4809]: I1206 07:15:04.687399 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="83232476a9438e2c6c8a017430a796a52b8a241342e4ca8e0b860933c7fb678e" Dec 06 07:15:04 crc kubenswrapper[4809]: I1206 07:15:04.687206 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416755-jq7qg" Dec 06 07:15:05 crc kubenswrapper[4809]: I1206 07:15:05.202731 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416710-sjkdm"] Dec 06 07:15:05 crc kubenswrapper[4809]: I1206 07:15:05.217217 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416710-sjkdm"] Dec 06 07:15:05 crc kubenswrapper[4809]: I1206 07:15:05.406264 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e97b08e-c54b-42c8-9e6f-893215182a8e" path="/var/lib/kubelet/pods/3e97b08e-c54b-42c8-9e6f-893215182a8e/volumes" Dec 06 07:15:37 crc kubenswrapper[4809]: I1206 07:15:37.820148 4809 scope.go:117] "RemoveContainer" containerID="40c2ed333d9348bfc04097dc0c2374eeac937cb4c29d26f625c8626cba492c73" Dec 06 07:16:34 crc kubenswrapper[4809]: I1206 07:16:34.496709 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:16:34 crc kubenswrapper[4809]: I1206 07:16:34.497865 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:17:04 crc kubenswrapper[4809]: I1206 07:17:04.496588 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:17:04 crc kubenswrapper[4809]: I1206 07:17:04.497442 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:17:10 crc kubenswrapper[4809]: I1206 07:17:10.326812 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dt8kb"] Dec 06 07:17:10 crc kubenswrapper[4809]: E1206 07:17:10.328319 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef6236c2-1bac-49bb-9ad7-642921a9dc35" containerName="collect-profiles" Dec 06 07:17:10 crc kubenswrapper[4809]: I1206 07:17:10.328338 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef6236c2-1bac-49bb-9ad7-642921a9dc35" containerName="collect-profiles" Dec 06 07:17:10 crc kubenswrapper[4809]: I1206 07:17:10.328658 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef6236c2-1bac-49bb-9ad7-642921a9dc35" containerName="collect-profiles" Dec 06 07:17:10 crc kubenswrapper[4809]: I1206 07:17:10.331260 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dt8kb" Dec 06 07:17:10 crc kubenswrapper[4809]: I1206 07:17:10.360275 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dt8kb"] Dec 06 07:17:10 crc kubenswrapper[4809]: I1206 07:17:10.508163 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/869d8981-4bea-4606-8a08-5d668025d1cb-catalog-content\") pod \"certified-operators-dt8kb\" (UID: \"869d8981-4bea-4606-8a08-5d668025d1cb\") " pod="openshift-marketplace/certified-operators-dt8kb" Dec 06 07:17:10 crc kubenswrapper[4809]: I1206 07:17:10.508356 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/869d8981-4bea-4606-8a08-5d668025d1cb-utilities\") pod \"certified-operators-dt8kb\" (UID: \"869d8981-4bea-4606-8a08-5d668025d1cb\") " pod="openshift-marketplace/certified-operators-dt8kb" Dec 06 07:17:10 crc kubenswrapper[4809]: I1206 07:17:10.508391 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csp8m\" (UniqueName: \"kubernetes.io/projected/869d8981-4bea-4606-8a08-5d668025d1cb-kube-api-access-csp8m\") pod \"certified-operators-dt8kb\" (UID: \"869d8981-4bea-4606-8a08-5d668025d1cb\") " pod="openshift-marketplace/certified-operators-dt8kb" Dec 06 07:17:10 crc kubenswrapper[4809]: I1206 07:17:10.611238 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/869d8981-4bea-4606-8a08-5d668025d1cb-catalog-content\") pod \"certified-operators-dt8kb\" (UID: \"869d8981-4bea-4606-8a08-5d668025d1cb\") " pod="openshift-marketplace/certified-operators-dt8kb" Dec 06 07:17:10 crc kubenswrapper[4809]: I1206 07:17:10.611352 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/869d8981-4bea-4606-8a08-5d668025d1cb-utilities\") pod \"certified-operators-dt8kb\" (UID: \"869d8981-4bea-4606-8a08-5d668025d1cb\") " pod="openshift-marketplace/certified-operators-dt8kb" Dec 06 07:17:10 crc kubenswrapper[4809]: I1206 07:17:10.611376 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csp8m\" (UniqueName: \"kubernetes.io/projected/869d8981-4bea-4606-8a08-5d668025d1cb-kube-api-access-csp8m\") pod \"certified-operators-dt8kb\" (UID: \"869d8981-4bea-4606-8a08-5d668025d1cb\") " pod="openshift-marketplace/certified-operators-dt8kb" Dec 06 07:17:10 crc kubenswrapper[4809]: I1206 07:17:10.611864 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/869d8981-4bea-4606-8a08-5d668025d1cb-catalog-content\") pod \"certified-operators-dt8kb\" (UID: \"869d8981-4bea-4606-8a08-5d668025d1cb\") " pod="openshift-marketplace/certified-operators-dt8kb" Dec 06 07:17:10 crc kubenswrapper[4809]: I1206 07:17:10.611917 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/869d8981-4bea-4606-8a08-5d668025d1cb-utilities\") pod \"certified-operators-dt8kb\" (UID: \"869d8981-4bea-4606-8a08-5d668025d1cb\") " pod="openshift-marketplace/certified-operators-dt8kb" Dec 06 07:17:10 crc kubenswrapper[4809]: I1206 07:17:10.632678 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csp8m\" (UniqueName: \"kubernetes.io/projected/869d8981-4bea-4606-8a08-5d668025d1cb-kube-api-access-csp8m\") pod \"certified-operators-dt8kb\" (UID: \"869d8981-4bea-4606-8a08-5d668025d1cb\") " pod="openshift-marketplace/certified-operators-dt8kb" Dec 06 07:17:10 crc kubenswrapper[4809]: I1206 07:17:10.657138 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dt8kb" Dec 06 07:17:13 crc kubenswrapper[4809]: I1206 07:17:13.841803 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dt8kb"] Dec 06 07:17:14 crc kubenswrapper[4809]: I1206 07:17:14.092507 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dt8kb" event={"ID":"869d8981-4bea-4606-8a08-5d668025d1cb","Type":"ContainerStarted","Data":"b013e18b38ec523c32ecfde8644ad5764edae82d853fdec18d98f6e635027ef6"} Dec 06 07:17:14 crc kubenswrapper[4809]: I1206 07:17:14.092862 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dt8kb" event={"ID":"869d8981-4bea-4606-8a08-5d668025d1cb","Type":"ContainerStarted","Data":"5cf4bcb9b28df432d7f69c73439a469cd1766abe8e284a1c8ab3ac0c74260b2b"} Dec 06 07:17:15 crc kubenswrapper[4809]: I1206 07:17:15.105136 4809 generic.go:334] "Generic (PLEG): container finished" podID="869d8981-4bea-4606-8a08-5d668025d1cb" containerID="b013e18b38ec523c32ecfde8644ad5764edae82d853fdec18d98f6e635027ef6" exitCode=0 Dec 06 07:17:15 crc kubenswrapper[4809]: I1206 07:17:15.105230 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dt8kb" event={"ID":"869d8981-4bea-4606-8a08-5d668025d1cb","Type":"ContainerDied","Data":"b013e18b38ec523c32ecfde8644ad5764edae82d853fdec18d98f6e635027ef6"} Dec 06 07:17:17 crc kubenswrapper[4809]: I1206 07:17:17.128843 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dt8kb" event={"ID":"869d8981-4bea-4606-8a08-5d668025d1cb","Type":"ContainerStarted","Data":"9ff00d424bef2b276ac790a01b83efad16b1ce292f6edaf6eff5ea5ccaa48861"} Dec 06 07:17:20 crc kubenswrapper[4809]: I1206 07:17:20.307990 4809 generic.go:334] "Generic (PLEG): container finished" podID="869d8981-4bea-4606-8a08-5d668025d1cb" containerID="9ff00d424bef2b276ac790a01b83efad16b1ce292f6edaf6eff5ea5ccaa48861" exitCode=0 Dec 06 07:17:20 crc kubenswrapper[4809]: I1206 07:17:20.308485 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dt8kb" event={"ID":"869d8981-4bea-4606-8a08-5d668025d1cb","Type":"ContainerDied","Data":"9ff00d424bef2b276ac790a01b83efad16b1ce292f6edaf6eff5ea5ccaa48861"} Dec 06 07:17:22 crc kubenswrapper[4809]: I1206 07:17:22.334897 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dt8kb" event={"ID":"869d8981-4bea-4606-8a08-5d668025d1cb","Type":"ContainerStarted","Data":"1fc1962f2576f1809c92491ecd86b05a6e515b6c818db357fc28c4950ae6c39f"} Dec 06 07:17:22 crc kubenswrapper[4809]: I1206 07:17:22.356637 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dt8kb" podStartSLOduration=6.304409612 podStartE2EDuration="12.356617014s" podCreationTimestamp="2025-12-06 07:17:10 +0000 UTC" firstStartedPulling="2025-12-06 07:17:15.107435185 +0000 UTC m=+5159.996418127" lastFinishedPulling="2025-12-06 07:17:21.159642587 +0000 UTC m=+5166.048625529" observedRunningTime="2025-12-06 07:17:22.350073687 +0000 UTC m=+5167.239056639" watchObservedRunningTime="2025-12-06 07:17:22.356617014 +0000 UTC m=+5167.245599956" Dec 06 07:17:30 crc kubenswrapper[4809]: I1206 07:17:30.657232 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dt8kb" Dec 06 07:17:30 crc kubenswrapper[4809]: I1206 07:17:30.657830 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dt8kb" Dec 06 07:17:30 crc kubenswrapper[4809]: I1206 07:17:30.720918 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dt8kb" Dec 06 07:17:32 crc kubenswrapper[4809]: I1206 07:17:32.025366 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dt8kb" Dec 06 07:17:32 crc kubenswrapper[4809]: I1206 07:17:32.080802 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dt8kb"] Dec 06 07:17:33 crc kubenswrapper[4809]: I1206 07:17:33.457375 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dt8kb" podUID="869d8981-4bea-4606-8a08-5d668025d1cb" containerName="registry-server" containerID="cri-o://1fc1962f2576f1809c92491ecd86b05a6e515b6c818db357fc28c4950ae6c39f" gracePeriod=2 Dec 06 07:17:34 crc kubenswrapper[4809]: I1206 07:17:34.475005 4809 generic.go:334] "Generic (PLEG): container finished" podID="869d8981-4bea-4606-8a08-5d668025d1cb" containerID="1fc1962f2576f1809c92491ecd86b05a6e515b6c818db357fc28c4950ae6c39f" exitCode=0 Dec 06 07:17:34 crc kubenswrapper[4809]: I1206 07:17:34.475079 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dt8kb" event={"ID":"869d8981-4bea-4606-8a08-5d668025d1cb","Type":"ContainerDied","Data":"1fc1962f2576f1809c92491ecd86b05a6e515b6c818db357fc28c4950ae6c39f"} Dec 06 07:17:34 crc kubenswrapper[4809]: I1206 07:17:34.496037 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:17:34 crc kubenswrapper[4809]: I1206 07:17:34.496085 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:17:34 crc kubenswrapper[4809]: I1206 07:17:34.496122 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-npms2" Dec 06 07:17:34 crc kubenswrapper[4809]: I1206 07:17:34.497075 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"af73883c3be3c80377e0f4806f2ecfa3ce57913ac4edf4c8383f744be4172b66"} pod="openshift-machine-config-operator/machine-config-daemon-npms2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 07:17:34 crc kubenswrapper[4809]: I1206 07:17:34.497125 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" containerID="cri-o://af73883c3be3c80377e0f4806f2ecfa3ce57913ac4edf4c8383f744be4172b66" gracePeriod=600 Dec 06 07:17:34 crc kubenswrapper[4809]: E1206 07:17:34.624540 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:17:34 crc kubenswrapper[4809]: I1206 07:17:34.802461 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dt8kb" Dec 06 07:17:34 crc kubenswrapper[4809]: I1206 07:17:34.957025 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/869d8981-4bea-4606-8a08-5d668025d1cb-catalog-content\") pod \"869d8981-4bea-4606-8a08-5d668025d1cb\" (UID: \"869d8981-4bea-4606-8a08-5d668025d1cb\") " Dec 06 07:17:34 crc kubenswrapper[4809]: I1206 07:17:34.957159 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-csp8m\" (UniqueName: \"kubernetes.io/projected/869d8981-4bea-4606-8a08-5d668025d1cb-kube-api-access-csp8m\") pod \"869d8981-4bea-4606-8a08-5d668025d1cb\" (UID: \"869d8981-4bea-4606-8a08-5d668025d1cb\") " Dec 06 07:17:34 crc kubenswrapper[4809]: I1206 07:17:34.957234 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/869d8981-4bea-4606-8a08-5d668025d1cb-utilities\") pod \"869d8981-4bea-4606-8a08-5d668025d1cb\" (UID: \"869d8981-4bea-4606-8a08-5d668025d1cb\") " Dec 06 07:17:34 crc kubenswrapper[4809]: I1206 07:17:34.958153 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/869d8981-4bea-4606-8a08-5d668025d1cb-utilities" (OuterVolumeSpecName: "utilities") pod "869d8981-4bea-4606-8a08-5d668025d1cb" (UID: "869d8981-4bea-4606-8a08-5d668025d1cb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:17:34 crc kubenswrapper[4809]: I1206 07:17:34.965702 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/869d8981-4bea-4606-8a08-5d668025d1cb-kube-api-access-csp8m" (OuterVolumeSpecName: "kube-api-access-csp8m") pod "869d8981-4bea-4606-8a08-5d668025d1cb" (UID: "869d8981-4bea-4606-8a08-5d668025d1cb"). InnerVolumeSpecName "kube-api-access-csp8m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:17:35 crc kubenswrapper[4809]: I1206 07:17:35.015709 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/869d8981-4bea-4606-8a08-5d668025d1cb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "869d8981-4bea-4606-8a08-5d668025d1cb" (UID: "869d8981-4bea-4606-8a08-5d668025d1cb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:17:35 crc kubenswrapper[4809]: I1206 07:17:35.060106 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/869d8981-4bea-4606-8a08-5d668025d1cb-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:35 crc kubenswrapper[4809]: I1206 07:17:35.060141 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-csp8m\" (UniqueName: \"kubernetes.io/projected/869d8981-4bea-4606-8a08-5d668025d1cb-kube-api-access-csp8m\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:35 crc kubenswrapper[4809]: I1206 07:17:35.060151 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/869d8981-4bea-4606-8a08-5d668025d1cb-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 07:17:35 crc kubenswrapper[4809]: I1206 07:17:35.491116 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dt8kb" event={"ID":"869d8981-4bea-4606-8a08-5d668025d1cb","Type":"ContainerDied","Data":"5cf4bcb9b28df432d7f69c73439a469cd1766abe8e284a1c8ab3ac0c74260b2b"} Dec 06 07:17:35 crc kubenswrapper[4809]: I1206 07:17:35.492460 4809 scope.go:117] "RemoveContainer" containerID="1fc1962f2576f1809c92491ecd86b05a6e515b6c818db357fc28c4950ae6c39f" Dec 06 07:17:35 crc kubenswrapper[4809]: I1206 07:17:35.491162 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dt8kb" Dec 06 07:17:35 crc kubenswrapper[4809]: I1206 07:17:35.497898 4809 generic.go:334] "Generic (PLEG): container finished" podID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerID="af73883c3be3c80377e0f4806f2ecfa3ce57913ac4edf4c8383f744be4172b66" exitCode=0 Dec 06 07:17:35 crc kubenswrapper[4809]: I1206 07:17:35.497914 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" event={"ID":"cbf4cf62-024e-4703-a8b8-9aecda9cd26a","Type":"ContainerDied","Data":"af73883c3be3c80377e0f4806f2ecfa3ce57913ac4edf4c8383f744be4172b66"} Dec 06 07:17:35 crc kubenswrapper[4809]: I1206 07:17:35.498996 4809 scope.go:117] "RemoveContainer" containerID="af73883c3be3c80377e0f4806f2ecfa3ce57913ac4edf4c8383f744be4172b66" Dec 06 07:17:35 crc kubenswrapper[4809]: E1206 07:17:35.499493 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:17:35 crc kubenswrapper[4809]: I1206 07:17:35.543221 4809 scope.go:117] "RemoveContainer" containerID="9ff00d424bef2b276ac790a01b83efad16b1ce292f6edaf6eff5ea5ccaa48861" Dec 06 07:17:35 crc kubenswrapper[4809]: I1206 07:17:35.543966 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dt8kb"] Dec 06 07:17:35 crc kubenswrapper[4809]: I1206 07:17:35.561049 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dt8kb"] Dec 06 07:17:35 crc kubenswrapper[4809]: I1206 07:17:35.579792 4809 scope.go:117] "RemoveContainer" containerID="b013e18b38ec523c32ecfde8644ad5764edae82d853fdec18d98f6e635027ef6" Dec 06 07:17:35 crc kubenswrapper[4809]: I1206 07:17:35.635074 4809 scope.go:117] "RemoveContainer" containerID="9ddac882b8b7232d7d5f4ba25a08bfffc6501c262ddc8c7ab993417cc0299ce0" Dec 06 07:17:37 crc kubenswrapper[4809]: I1206 07:17:37.407096 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="869d8981-4bea-4606-8a08-5d668025d1cb" path="/var/lib/kubelet/pods/869d8981-4bea-4606-8a08-5d668025d1cb/volumes" Dec 06 07:17:48 crc kubenswrapper[4809]: I1206 07:17:48.389315 4809 scope.go:117] "RemoveContainer" containerID="af73883c3be3c80377e0f4806f2ecfa3ce57913ac4edf4c8383f744be4172b66" Dec 06 07:17:48 crc kubenswrapper[4809]: E1206 07:17:48.393852 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:17:51 crc kubenswrapper[4809]: I1206 07:17:51.727361 4809 patch_prober.go:28] interesting pod/console-5cccfc6d89-9j7zs container/console namespace/openshift-console: Readiness probe status=failure output="Get \"https://10.217.0.133:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:17:51 crc kubenswrapper[4809]: I1206 07:17:51.728899 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/console-5cccfc6d89-9j7zs" podUID="ac4de3b3-b463-4e5b-b935-ee6561519ddf" containerName="console" probeResult="failure" output="Get \"https://10.217.0.133:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:18:03 crc kubenswrapper[4809]: I1206 07:18:03.388668 4809 scope.go:117] "RemoveContainer" containerID="af73883c3be3c80377e0f4806f2ecfa3ce57913ac4edf4c8383f744be4172b66" Dec 06 07:18:03 crc kubenswrapper[4809]: E1206 07:18:03.389556 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:18:14 crc kubenswrapper[4809]: I1206 07:18:14.389028 4809 scope.go:117] "RemoveContainer" containerID="af73883c3be3c80377e0f4806f2ecfa3ce57913ac4edf4c8383f744be4172b66" Dec 06 07:18:14 crc kubenswrapper[4809]: E1206 07:18:14.389924 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:18:25 crc kubenswrapper[4809]: I1206 07:18:25.399484 4809 scope.go:117] "RemoveContainer" containerID="af73883c3be3c80377e0f4806f2ecfa3ce57913ac4edf4c8383f744be4172b66" Dec 06 07:18:25 crc kubenswrapper[4809]: E1206 07:18:25.401023 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:18:39 crc kubenswrapper[4809]: I1206 07:18:39.388610 4809 scope.go:117] "RemoveContainer" containerID="af73883c3be3c80377e0f4806f2ecfa3ce57913ac4edf4c8383f744be4172b66" Dec 06 07:18:39 crc kubenswrapper[4809]: E1206 07:18:39.389520 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:18:46 crc kubenswrapper[4809]: I1206 07:18:46.718323 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Dec 06 07:18:46 crc kubenswrapper[4809]: E1206 07:18:46.719634 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="869d8981-4bea-4606-8a08-5d668025d1cb" containerName="registry-server" Dec 06 07:18:46 crc kubenswrapper[4809]: I1206 07:18:46.719657 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="869d8981-4bea-4606-8a08-5d668025d1cb" containerName="registry-server" Dec 06 07:18:46 crc kubenswrapper[4809]: E1206 07:18:46.719696 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="869d8981-4bea-4606-8a08-5d668025d1cb" containerName="extract-utilities" Dec 06 07:18:46 crc kubenswrapper[4809]: I1206 07:18:46.719706 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="869d8981-4bea-4606-8a08-5d668025d1cb" containerName="extract-utilities" Dec 06 07:18:46 crc kubenswrapper[4809]: E1206 07:18:46.719753 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="869d8981-4bea-4606-8a08-5d668025d1cb" containerName="extract-content" Dec 06 07:18:46 crc kubenswrapper[4809]: I1206 07:18:46.719762 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="869d8981-4bea-4606-8a08-5d668025d1cb" containerName="extract-content" Dec 06 07:18:46 crc kubenswrapper[4809]: I1206 07:18:46.720076 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="869d8981-4bea-4606-8a08-5d668025d1cb" containerName="registry-server" Dec 06 07:18:46 crc kubenswrapper[4809]: I1206 07:18:46.721195 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 06 07:18:46 crc kubenswrapper[4809]: I1206 07:18:46.723517 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 06 07:18:46 crc kubenswrapper[4809]: I1206 07:18:46.723782 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-cfdpf" Dec 06 07:18:46 crc kubenswrapper[4809]: I1206 07:18:46.724007 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 06 07:18:46 crc kubenswrapper[4809]: I1206 07:18:46.724595 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 06 07:18:46 crc kubenswrapper[4809]: I1206 07:18:46.731695 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 06 07:18:46 crc kubenswrapper[4809]: I1206 07:18:46.813770 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/973f75f0-3e24-4648-b57b-864065deb9b1-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"973f75f0-3e24-4648-b57b-864065deb9b1\") " pod="openstack/tempest-tests-tempest" Dec 06 07:18:46 crc kubenswrapper[4809]: I1206 07:18:46.814000 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/973f75f0-3e24-4648-b57b-864065deb9b1-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"973f75f0-3e24-4648-b57b-864065deb9b1\") " pod="openstack/tempest-tests-tempest" Dec 06 07:18:46 crc kubenswrapper[4809]: I1206 07:18:46.814057 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/973f75f0-3e24-4648-b57b-864065deb9b1-config-data\") pod \"tempest-tests-tempest\" (UID: \"973f75f0-3e24-4648-b57b-864065deb9b1\") " pod="openstack/tempest-tests-tempest" Dec 06 07:18:46 crc kubenswrapper[4809]: I1206 07:18:46.916232 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/973f75f0-3e24-4648-b57b-864065deb9b1-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"973f75f0-3e24-4648-b57b-864065deb9b1\") " pod="openstack/tempest-tests-tempest" Dec 06 07:18:46 crc kubenswrapper[4809]: I1206 07:18:46.916312 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/973f75f0-3e24-4648-b57b-864065deb9b1-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"973f75f0-3e24-4648-b57b-864065deb9b1\") " pod="openstack/tempest-tests-tempest" Dec 06 07:18:46 crc kubenswrapper[4809]: I1206 07:18:46.916366 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkqlh\" (UniqueName: \"kubernetes.io/projected/973f75f0-3e24-4648-b57b-864065deb9b1-kube-api-access-bkqlh\") pod \"tempest-tests-tempest\" (UID: \"973f75f0-3e24-4648-b57b-864065deb9b1\") " pod="openstack/tempest-tests-tempest" Dec 06 07:18:46 crc kubenswrapper[4809]: I1206 07:18:46.916427 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tempest-tests-tempest\" (UID: \"973f75f0-3e24-4648-b57b-864065deb9b1\") " pod="openstack/tempest-tests-tempest" Dec 06 07:18:46 crc kubenswrapper[4809]: I1206 07:18:46.916511 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/973f75f0-3e24-4648-b57b-864065deb9b1-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"973f75f0-3e24-4648-b57b-864065deb9b1\") " pod="openstack/tempest-tests-tempest" Dec 06 07:18:46 crc kubenswrapper[4809]: I1206 07:18:46.916578 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/973f75f0-3e24-4648-b57b-864065deb9b1-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"973f75f0-3e24-4648-b57b-864065deb9b1\") " pod="openstack/tempest-tests-tempest" Dec 06 07:18:46 crc kubenswrapper[4809]: I1206 07:18:46.916602 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/973f75f0-3e24-4648-b57b-864065deb9b1-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"973f75f0-3e24-4648-b57b-864065deb9b1\") " pod="openstack/tempest-tests-tempest" Dec 06 07:18:46 crc kubenswrapper[4809]: I1206 07:18:46.916640 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/973f75f0-3e24-4648-b57b-864065deb9b1-config-data\") pod \"tempest-tests-tempest\" (UID: \"973f75f0-3e24-4648-b57b-864065deb9b1\") " pod="openstack/tempest-tests-tempest" Dec 06 07:18:46 crc kubenswrapper[4809]: I1206 07:18:46.916662 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/973f75f0-3e24-4648-b57b-864065deb9b1-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"973f75f0-3e24-4648-b57b-864065deb9b1\") " pod="openstack/tempest-tests-tempest" Dec 06 07:18:46 crc kubenswrapper[4809]: I1206 07:18:46.919738 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/973f75f0-3e24-4648-b57b-864065deb9b1-config-data\") pod \"tempest-tests-tempest\" (UID: \"973f75f0-3e24-4648-b57b-864065deb9b1\") " pod="openstack/tempest-tests-tempest" Dec 06 07:18:46 crc kubenswrapper[4809]: I1206 07:18:46.920120 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/973f75f0-3e24-4648-b57b-864065deb9b1-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"973f75f0-3e24-4648-b57b-864065deb9b1\") " pod="openstack/tempest-tests-tempest" Dec 06 07:18:46 crc kubenswrapper[4809]: I1206 07:18:46.925562 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/973f75f0-3e24-4648-b57b-864065deb9b1-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"973f75f0-3e24-4648-b57b-864065deb9b1\") " pod="openstack/tempest-tests-tempest" Dec 06 07:18:47 crc kubenswrapper[4809]: I1206 07:18:47.018785 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/973f75f0-3e24-4648-b57b-864065deb9b1-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"973f75f0-3e24-4648-b57b-864065deb9b1\") " pod="openstack/tempest-tests-tempest" Dec 06 07:18:47 crc kubenswrapper[4809]: I1206 07:18:47.019281 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/973f75f0-3e24-4648-b57b-864065deb9b1-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"973f75f0-3e24-4648-b57b-864065deb9b1\") " pod="openstack/tempest-tests-tempest" Dec 06 07:18:47 crc kubenswrapper[4809]: I1206 07:18:47.019412 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/973f75f0-3e24-4648-b57b-864065deb9b1-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"973f75f0-3e24-4648-b57b-864065deb9b1\") " pod="openstack/tempest-tests-tempest" Dec 06 07:18:47 crc kubenswrapper[4809]: I1206 07:18:47.019653 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/973f75f0-3e24-4648-b57b-864065deb9b1-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"973f75f0-3e24-4648-b57b-864065deb9b1\") " pod="openstack/tempest-tests-tempest" Dec 06 07:18:47 crc kubenswrapper[4809]: I1206 07:18:47.019752 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/973f75f0-3e24-4648-b57b-864065deb9b1-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"973f75f0-3e24-4648-b57b-864065deb9b1\") " pod="openstack/tempest-tests-tempest" Dec 06 07:18:47 crc kubenswrapper[4809]: I1206 07:18:47.019881 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkqlh\" (UniqueName: \"kubernetes.io/projected/973f75f0-3e24-4648-b57b-864065deb9b1-kube-api-access-bkqlh\") pod \"tempest-tests-tempest\" (UID: \"973f75f0-3e24-4648-b57b-864065deb9b1\") " pod="openstack/tempest-tests-tempest" Dec 06 07:18:47 crc kubenswrapper[4809]: I1206 07:18:47.020026 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tempest-tests-tempest\" (UID: \"973f75f0-3e24-4648-b57b-864065deb9b1\") " pod="openstack/tempest-tests-tempest" Dec 06 07:18:47 crc kubenswrapper[4809]: I1206 07:18:47.020978 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/973f75f0-3e24-4648-b57b-864065deb9b1-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"973f75f0-3e24-4648-b57b-864065deb9b1\") " pod="openstack/tempest-tests-tempest" Dec 06 07:18:47 crc kubenswrapper[4809]: I1206 07:18:47.022228 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tempest-tests-tempest\" (UID: \"973f75f0-3e24-4648-b57b-864065deb9b1\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/tempest-tests-tempest" Dec 06 07:18:47 crc kubenswrapper[4809]: I1206 07:18:47.023010 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/973f75f0-3e24-4648-b57b-864065deb9b1-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"973f75f0-3e24-4648-b57b-864065deb9b1\") " pod="openstack/tempest-tests-tempest" Dec 06 07:18:47 crc kubenswrapper[4809]: I1206 07:18:47.023319 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/973f75f0-3e24-4648-b57b-864065deb9b1-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"973f75f0-3e24-4648-b57b-864065deb9b1\") " pod="openstack/tempest-tests-tempest" Dec 06 07:18:47 crc kubenswrapper[4809]: I1206 07:18:47.041070 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkqlh\" (UniqueName: \"kubernetes.io/projected/973f75f0-3e24-4648-b57b-864065deb9b1-kube-api-access-bkqlh\") pod \"tempest-tests-tempest\" (UID: \"973f75f0-3e24-4648-b57b-864065deb9b1\") " pod="openstack/tempest-tests-tempest" Dec 06 07:18:47 crc kubenswrapper[4809]: I1206 07:18:47.061304 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tempest-tests-tempest\" (UID: \"973f75f0-3e24-4648-b57b-864065deb9b1\") " pod="openstack/tempest-tests-tempest" Dec 06 07:18:47 crc kubenswrapper[4809]: I1206 07:18:47.342571 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 06 07:18:47 crc kubenswrapper[4809]: I1206 07:18:47.856018 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 06 07:18:47 crc kubenswrapper[4809]: W1206 07:18:47.889403 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod973f75f0_3e24_4648_b57b_864065deb9b1.slice/crio-1ce794f012a2205d331438d97153b6f6aeee06635e5a44bc945d8ad86fc0fbb2 WatchSource:0}: Error finding container 1ce794f012a2205d331438d97153b6f6aeee06635e5a44bc945d8ad86fc0fbb2: Status 404 returned error can't find the container with id 1ce794f012a2205d331438d97153b6f6aeee06635e5a44bc945d8ad86fc0fbb2 Dec 06 07:18:47 crc kubenswrapper[4809]: I1206 07:18:47.891876 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 07:18:48 crc kubenswrapper[4809]: I1206 07:18:48.321482 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"973f75f0-3e24-4648-b57b-864065deb9b1","Type":"ContainerStarted","Data":"1ce794f012a2205d331438d97153b6f6aeee06635e5a44bc945d8ad86fc0fbb2"} Dec 06 07:18:52 crc kubenswrapper[4809]: I1206 07:18:52.391063 4809 scope.go:117] "RemoveContainer" containerID="af73883c3be3c80377e0f4806f2ecfa3ce57913ac4edf4c8383f744be4172b66" Dec 06 07:18:52 crc kubenswrapper[4809]: E1206 07:18:52.393384 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:19:04 crc kubenswrapper[4809]: I1206 07:19:04.389300 4809 scope.go:117] "RemoveContainer" containerID="af73883c3be3c80377e0f4806f2ecfa3ce57913ac4edf4c8383f744be4172b66" Dec 06 07:19:04 crc kubenswrapper[4809]: E1206 07:19:04.390364 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:19:13 crc kubenswrapper[4809]: I1206 07:19:13.762203 4809 trace.go:236] Trace[1966278646]: "Calculate volume metrics of config-data-generated for pod openstack/openstack-cell1-galera-0" (06-Dec-2025 07:19:12.536) (total time: 1224ms): Dec 06 07:19:13 crc kubenswrapper[4809]: Trace[1966278646]: [1.22438342s] [1.22438342s] END Dec 06 07:19:18 crc kubenswrapper[4809]: I1206 07:19:18.388666 4809 scope.go:117] "RemoveContainer" containerID="af73883c3be3c80377e0f4806f2ecfa3ce57913ac4edf4c8383f744be4172b66" Dec 06 07:19:18 crc kubenswrapper[4809]: E1206 07:19:18.389571 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:19:23 crc kubenswrapper[4809]: I1206 07:19:23.217763 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.682686751s: [/var/lib/containers/storage/overlay/16afad53e653d3e637e5df40959b6b5710d9258ef689b69ce6d8c2bd72f56f97/diff /var/log/pods/openstack_openstackclient_6c6578f3-bf7b-4892-abc9-5f4f145c8d36/openstackclient/0.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:19:32 crc kubenswrapper[4809]: I1206 07:19:32.388974 4809 scope.go:117] "RemoveContainer" containerID="af73883c3be3c80377e0f4806f2ecfa3ce57913ac4edf4c8383f744be4172b66" Dec 06 07:19:32 crc kubenswrapper[4809]: E1206 07:19:32.391208 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:19:44 crc kubenswrapper[4809]: E1206 07:19:44.273804 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Dec 06 07:19:44 crc kubenswrapper[4809]: E1206 07:19:44.276394 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bkqlh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(973f75f0-3e24-4648-b57b-864065deb9b1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 07:19:44 crc kubenswrapper[4809]: E1206 07:19:44.277686 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="973f75f0-3e24-4648-b57b-864065deb9b1" Dec 06 07:19:45 crc kubenswrapper[4809]: E1206 07:19:45.013591 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="973f75f0-3e24-4648-b57b-864065deb9b1" Dec 06 07:19:46 crc kubenswrapper[4809]: I1206 07:19:46.388425 4809 scope.go:117] "RemoveContainer" containerID="af73883c3be3c80377e0f4806f2ecfa3ce57913ac4edf4c8383f744be4172b66" Dec 06 07:19:46 crc kubenswrapper[4809]: E1206 07:19:46.389116 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:19:56 crc kubenswrapper[4809]: I1206 07:19:56.583088 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-sh8jm"] Dec 06 07:19:56 crc kubenswrapper[4809]: I1206 07:19:56.586838 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sh8jm" Dec 06 07:19:56 crc kubenswrapper[4809]: I1206 07:19:56.603272 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sh8jm"] Dec 06 07:19:56 crc kubenswrapper[4809]: I1206 07:19:56.697099 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f5f3c13-919e-490d-9796-80d82cd12a53-utilities\") pod \"redhat-marketplace-sh8jm\" (UID: \"8f5f3c13-919e-490d-9796-80d82cd12a53\") " pod="openshift-marketplace/redhat-marketplace-sh8jm" Dec 06 07:19:56 crc kubenswrapper[4809]: I1206 07:19:56.697601 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f5f3c13-919e-490d-9796-80d82cd12a53-catalog-content\") pod \"redhat-marketplace-sh8jm\" (UID: \"8f5f3c13-919e-490d-9796-80d82cd12a53\") " pod="openshift-marketplace/redhat-marketplace-sh8jm" Dec 06 07:19:56 crc kubenswrapper[4809]: I1206 07:19:56.697981 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzr2h\" (UniqueName: \"kubernetes.io/projected/8f5f3c13-919e-490d-9796-80d82cd12a53-kube-api-access-fzr2h\") pod \"redhat-marketplace-sh8jm\" (UID: \"8f5f3c13-919e-490d-9796-80d82cd12a53\") " pod="openshift-marketplace/redhat-marketplace-sh8jm" Dec 06 07:19:56 crc kubenswrapper[4809]: I1206 07:19:56.800175 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzr2h\" (UniqueName: \"kubernetes.io/projected/8f5f3c13-919e-490d-9796-80d82cd12a53-kube-api-access-fzr2h\") pod \"redhat-marketplace-sh8jm\" (UID: \"8f5f3c13-919e-490d-9796-80d82cd12a53\") " pod="openshift-marketplace/redhat-marketplace-sh8jm" Dec 06 07:19:56 crc kubenswrapper[4809]: I1206 07:19:56.800352 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f5f3c13-919e-490d-9796-80d82cd12a53-utilities\") pod \"redhat-marketplace-sh8jm\" (UID: \"8f5f3c13-919e-490d-9796-80d82cd12a53\") " pod="openshift-marketplace/redhat-marketplace-sh8jm" Dec 06 07:19:56 crc kubenswrapper[4809]: I1206 07:19:56.800462 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f5f3c13-919e-490d-9796-80d82cd12a53-catalog-content\") pod \"redhat-marketplace-sh8jm\" (UID: \"8f5f3c13-919e-490d-9796-80d82cd12a53\") " pod="openshift-marketplace/redhat-marketplace-sh8jm" Dec 06 07:19:56 crc kubenswrapper[4809]: I1206 07:19:56.801072 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f5f3c13-919e-490d-9796-80d82cd12a53-utilities\") pod \"redhat-marketplace-sh8jm\" (UID: \"8f5f3c13-919e-490d-9796-80d82cd12a53\") " pod="openshift-marketplace/redhat-marketplace-sh8jm" Dec 06 07:19:56 crc kubenswrapper[4809]: I1206 07:19:56.801120 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f5f3c13-919e-490d-9796-80d82cd12a53-catalog-content\") pod \"redhat-marketplace-sh8jm\" (UID: \"8f5f3c13-919e-490d-9796-80d82cd12a53\") " pod="openshift-marketplace/redhat-marketplace-sh8jm" Dec 06 07:19:56 crc kubenswrapper[4809]: I1206 07:19:56.984530 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzr2h\" (UniqueName: \"kubernetes.io/projected/8f5f3c13-919e-490d-9796-80d82cd12a53-kube-api-access-fzr2h\") pod \"redhat-marketplace-sh8jm\" (UID: \"8f5f3c13-919e-490d-9796-80d82cd12a53\") " pod="openshift-marketplace/redhat-marketplace-sh8jm" Dec 06 07:19:57 crc kubenswrapper[4809]: I1206 07:19:57.267402 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sh8jm" Dec 06 07:19:57 crc kubenswrapper[4809]: I1206 07:19:57.389341 4809 scope.go:117] "RemoveContainer" containerID="af73883c3be3c80377e0f4806f2ecfa3ce57913ac4edf4c8383f744be4172b66" Dec 06 07:19:57 crc kubenswrapper[4809]: E1206 07:19:57.389983 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:19:58 crc kubenswrapper[4809]: W1206 07:19:58.063122 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f5f3c13_919e_490d_9796_80d82cd12a53.slice/crio-0d06c24ef1bfd7fc7b45aa59961e5c70ca0f7cd0b93352ddd22197101138ac17 WatchSource:0}: Error finding container 0d06c24ef1bfd7fc7b45aa59961e5c70ca0f7cd0b93352ddd22197101138ac17: Status 404 returned error can't find the container with id 0d06c24ef1bfd7fc7b45aa59961e5c70ca0f7cd0b93352ddd22197101138ac17 Dec 06 07:19:58 crc kubenswrapper[4809]: I1206 07:19:58.073459 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sh8jm"] Dec 06 07:19:58 crc kubenswrapper[4809]: I1206 07:19:58.151473 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sh8jm" event={"ID":"8f5f3c13-919e-490d-9796-80d82cd12a53","Type":"ContainerStarted","Data":"0d06c24ef1bfd7fc7b45aa59961e5c70ca0f7cd0b93352ddd22197101138ac17"} Dec 06 07:19:58 crc kubenswrapper[4809]: I1206 07:19:58.985714 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6p627"] Dec 06 07:19:58 crc kubenswrapper[4809]: I1206 07:19:58.991353 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6p627" Dec 06 07:19:59 crc kubenswrapper[4809]: I1206 07:19:59.006540 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6p627"] Dec 06 07:19:59 crc kubenswrapper[4809]: I1206 07:19:59.058351 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1435a636-4c7d-44fc-ae60-c1ebc049d0c3-utilities\") pod \"community-operators-6p627\" (UID: \"1435a636-4c7d-44fc-ae60-c1ebc049d0c3\") " pod="openshift-marketplace/community-operators-6p627" Dec 06 07:19:59 crc kubenswrapper[4809]: I1206 07:19:59.058606 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4c2c\" (UniqueName: \"kubernetes.io/projected/1435a636-4c7d-44fc-ae60-c1ebc049d0c3-kube-api-access-f4c2c\") pod \"community-operators-6p627\" (UID: \"1435a636-4c7d-44fc-ae60-c1ebc049d0c3\") " pod="openshift-marketplace/community-operators-6p627" Dec 06 07:19:59 crc kubenswrapper[4809]: I1206 07:19:59.058683 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1435a636-4c7d-44fc-ae60-c1ebc049d0c3-catalog-content\") pod \"community-operators-6p627\" (UID: \"1435a636-4c7d-44fc-ae60-c1ebc049d0c3\") " pod="openshift-marketplace/community-operators-6p627" Dec 06 07:19:59 crc kubenswrapper[4809]: I1206 07:19:59.161707 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1435a636-4c7d-44fc-ae60-c1ebc049d0c3-utilities\") pod \"community-operators-6p627\" (UID: \"1435a636-4c7d-44fc-ae60-c1ebc049d0c3\") " pod="openshift-marketplace/community-operators-6p627" Dec 06 07:19:59 crc kubenswrapper[4809]: I1206 07:19:59.161902 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4c2c\" (UniqueName: \"kubernetes.io/projected/1435a636-4c7d-44fc-ae60-c1ebc049d0c3-kube-api-access-f4c2c\") pod \"community-operators-6p627\" (UID: \"1435a636-4c7d-44fc-ae60-c1ebc049d0c3\") " pod="openshift-marketplace/community-operators-6p627" Dec 06 07:19:59 crc kubenswrapper[4809]: I1206 07:19:59.161978 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1435a636-4c7d-44fc-ae60-c1ebc049d0c3-catalog-content\") pod \"community-operators-6p627\" (UID: \"1435a636-4c7d-44fc-ae60-c1ebc049d0c3\") " pod="openshift-marketplace/community-operators-6p627" Dec 06 07:19:59 crc kubenswrapper[4809]: I1206 07:19:59.162675 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1435a636-4c7d-44fc-ae60-c1ebc049d0c3-utilities\") pod \"community-operators-6p627\" (UID: \"1435a636-4c7d-44fc-ae60-c1ebc049d0c3\") " pod="openshift-marketplace/community-operators-6p627" Dec 06 07:19:59 crc kubenswrapper[4809]: I1206 07:19:59.162691 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1435a636-4c7d-44fc-ae60-c1ebc049d0c3-catalog-content\") pod \"community-operators-6p627\" (UID: \"1435a636-4c7d-44fc-ae60-c1ebc049d0c3\") " pod="openshift-marketplace/community-operators-6p627" Dec 06 07:19:59 crc kubenswrapper[4809]: I1206 07:19:59.176091 4809 generic.go:334] "Generic (PLEG): container finished" podID="8f5f3c13-919e-490d-9796-80d82cd12a53" containerID="4fca7c336c7eb96d2b9d5a117a50b6807503243c5239d55959e73cd8d339fb96" exitCode=0 Dec 06 07:19:59 crc kubenswrapper[4809]: I1206 07:19:59.176161 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sh8jm" event={"ID":"8f5f3c13-919e-490d-9796-80d82cd12a53","Type":"ContainerDied","Data":"4fca7c336c7eb96d2b9d5a117a50b6807503243c5239d55959e73cd8d339fb96"} Dec 06 07:19:59 crc kubenswrapper[4809]: I1206 07:19:59.194062 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4c2c\" (UniqueName: \"kubernetes.io/projected/1435a636-4c7d-44fc-ae60-c1ebc049d0c3-kube-api-access-f4c2c\") pod \"community-operators-6p627\" (UID: \"1435a636-4c7d-44fc-ae60-c1ebc049d0c3\") " pod="openshift-marketplace/community-operators-6p627" Dec 06 07:19:59 crc kubenswrapper[4809]: I1206 07:19:59.196070 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-58pxg"] Dec 06 07:19:59 crc kubenswrapper[4809]: I1206 07:19:59.206043 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-58pxg" Dec 06 07:19:59 crc kubenswrapper[4809]: I1206 07:19:59.217830 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-58pxg"] Dec 06 07:19:59 crc kubenswrapper[4809]: I1206 07:19:59.267753 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fde5c8b4-ad5d-48ae-bc7f-3464d809dc92-utilities\") pod \"redhat-operators-58pxg\" (UID: \"fde5c8b4-ad5d-48ae-bc7f-3464d809dc92\") " pod="openshift-marketplace/redhat-operators-58pxg" Dec 06 07:19:59 crc kubenswrapper[4809]: I1206 07:19:59.268317 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fde5c8b4-ad5d-48ae-bc7f-3464d809dc92-catalog-content\") pod \"redhat-operators-58pxg\" (UID: \"fde5c8b4-ad5d-48ae-bc7f-3464d809dc92\") " pod="openshift-marketplace/redhat-operators-58pxg" Dec 06 07:19:59 crc kubenswrapper[4809]: I1206 07:19:59.268412 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbttt\" (UniqueName: \"kubernetes.io/projected/fde5c8b4-ad5d-48ae-bc7f-3464d809dc92-kube-api-access-rbttt\") pod \"redhat-operators-58pxg\" (UID: \"fde5c8b4-ad5d-48ae-bc7f-3464d809dc92\") " pod="openshift-marketplace/redhat-operators-58pxg" Dec 06 07:19:59 crc kubenswrapper[4809]: I1206 07:19:59.331923 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6p627" Dec 06 07:19:59 crc kubenswrapper[4809]: I1206 07:19:59.371449 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fde5c8b4-ad5d-48ae-bc7f-3464d809dc92-utilities\") pod \"redhat-operators-58pxg\" (UID: \"fde5c8b4-ad5d-48ae-bc7f-3464d809dc92\") " pod="openshift-marketplace/redhat-operators-58pxg" Dec 06 07:19:59 crc kubenswrapper[4809]: I1206 07:19:59.371564 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fde5c8b4-ad5d-48ae-bc7f-3464d809dc92-catalog-content\") pod \"redhat-operators-58pxg\" (UID: \"fde5c8b4-ad5d-48ae-bc7f-3464d809dc92\") " pod="openshift-marketplace/redhat-operators-58pxg" Dec 06 07:19:59 crc kubenswrapper[4809]: I1206 07:19:59.371619 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbttt\" (UniqueName: \"kubernetes.io/projected/fde5c8b4-ad5d-48ae-bc7f-3464d809dc92-kube-api-access-rbttt\") pod \"redhat-operators-58pxg\" (UID: \"fde5c8b4-ad5d-48ae-bc7f-3464d809dc92\") " pod="openshift-marketplace/redhat-operators-58pxg" Dec 06 07:19:59 crc kubenswrapper[4809]: I1206 07:19:59.372566 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fde5c8b4-ad5d-48ae-bc7f-3464d809dc92-utilities\") pod \"redhat-operators-58pxg\" (UID: \"fde5c8b4-ad5d-48ae-bc7f-3464d809dc92\") " pod="openshift-marketplace/redhat-operators-58pxg" Dec 06 07:19:59 crc kubenswrapper[4809]: I1206 07:19:59.372855 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fde5c8b4-ad5d-48ae-bc7f-3464d809dc92-catalog-content\") pod \"redhat-operators-58pxg\" (UID: \"fde5c8b4-ad5d-48ae-bc7f-3464d809dc92\") " pod="openshift-marketplace/redhat-operators-58pxg" Dec 06 07:19:59 crc kubenswrapper[4809]: I1206 07:19:59.445628 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbttt\" (UniqueName: \"kubernetes.io/projected/fde5c8b4-ad5d-48ae-bc7f-3464d809dc92-kube-api-access-rbttt\") pod \"redhat-operators-58pxg\" (UID: \"fde5c8b4-ad5d-48ae-bc7f-3464d809dc92\") " pod="openshift-marketplace/redhat-operators-58pxg" Dec 06 07:19:59 crc kubenswrapper[4809]: I1206 07:19:59.626804 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-58pxg" Dec 06 07:20:00 crc kubenswrapper[4809]: I1206 07:20:00.034186 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 06 07:20:00 crc kubenswrapper[4809]: I1206 07:20:00.102752 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6p627"] Dec 06 07:20:00 crc kubenswrapper[4809]: W1206 07:20:00.113542 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1435a636_4c7d_44fc_ae60_c1ebc049d0c3.slice/crio-eb567013cbbf13d58109b66398303c9f00a949d29eb136d73b8889392f06da5a WatchSource:0}: Error finding container eb567013cbbf13d58109b66398303c9f00a949d29eb136d73b8889392f06da5a: Status 404 returned error can't find the container with id eb567013cbbf13d58109b66398303c9f00a949d29eb136d73b8889392f06da5a Dec 06 07:20:00 crc kubenswrapper[4809]: I1206 07:20:00.188664 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6p627" event={"ID":"1435a636-4c7d-44fc-ae60-c1ebc049d0c3","Type":"ContainerStarted","Data":"eb567013cbbf13d58109b66398303c9f00a949d29eb136d73b8889392f06da5a"} Dec 06 07:20:00 crc kubenswrapper[4809]: W1206 07:20:00.244686 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfde5c8b4_ad5d_48ae_bc7f_3464d809dc92.slice/crio-bcd1d79370b2423ba7423b7caa3c83d9d38464f0456480322985038fc6d884eb WatchSource:0}: Error finding container bcd1d79370b2423ba7423b7caa3c83d9d38464f0456480322985038fc6d884eb: Status 404 returned error can't find the container with id bcd1d79370b2423ba7423b7caa3c83d9d38464f0456480322985038fc6d884eb Dec 06 07:20:00 crc kubenswrapper[4809]: I1206 07:20:00.249525 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-58pxg"] Dec 06 07:20:00 crc kubenswrapper[4809]: E1206 07:20:00.956984 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f5f3c13_919e_490d_9796_80d82cd12a53.slice/crio-conmon-dad9f45d255826350ce9093c7c8f08c4c68726d32eb0e8bee06101654fb0d288.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f5f3c13_919e_490d_9796_80d82cd12a53.slice/crio-dad9f45d255826350ce9093c7c8f08c4c68726d32eb0e8bee06101654fb0d288.scope\": RecentStats: unable to find data in memory cache]" Dec 06 07:20:01 crc kubenswrapper[4809]: I1206 07:20:01.203226 4809 generic.go:334] "Generic (PLEG): container finished" podID="fde5c8b4-ad5d-48ae-bc7f-3464d809dc92" containerID="e60d977c0f934e0236adce381ec03c98f2ff6f04a49e656ed3235d17a7d41ce9" exitCode=0 Dec 06 07:20:01 crc kubenswrapper[4809]: I1206 07:20:01.203891 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-58pxg" event={"ID":"fde5c8b4-ad5d-48ae-bc7f-3464d809dc92","Type":"ContainerDied","Data":"e60d977c0f934e0236adce381ec03c98f2ff6f04a49e656ed3235d17a7d41ce9"} Dec 06 07:20:01 crc kubenswrapper[4809]: I1206 07:20:01.203976 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-58pxg" event={"ID":"fde5c8b4-ad5d-48ae-bc7f-3464d809dc92","Type":"ContainerStarted","Data":"bcd1d79370b2423ba7423b7caa3c83d9d38464f0456480322985038fc6d884eb"} Dec 06 07:20:01 crc kubenswrapper[4809]: I1206 07:20:01.206243 4809 generic.go:334] "Generic (PLEG): container finished" podID="1435a636-4c7d-44fc-ae60-c1ebc049d0c3" containerID="11ad3f86328e9313e44ea51baf678ee472a2e294b818bd8c58fa0a6763a057e4" exitCode=0 Dec 06 07:20:01 crc kubenswrapper[4809]: I1206 07:20:01.206368 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6p627" event={"ID":"1435a636-4c7d-44fc-ae60-c1ebc049d0c3","Type":"ContainerDied","Data":"11ad3f86328e9313e44ea51baf678ee472a2e294b818bd8c58fa0a6763a057e4"} Dec 06 07:20:01 crc kubenswrapper[4809]: I1206 07:20:01.212926 4809 generic.go:334] "Generic (PLEG): container finished" podID="8f5f3c13-919e-490d-9796-80d82cd12a53" containerID="dad9f45d255826350ce9093c7c8f08c4c68726d32eb0e8bee06101654fb0d288" exitCode=0 Dec 06 07:20:01 crc kubenswrapper[4809]: I1206 07:20:01.212988 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sh8jm" event={"ID":"8f5f3c13-919e-490d-9796-80d82cd12a53","Type":"ContainerDied","Data":"dad9f45d255826350ce9093c7c8f08c4c68726d32eb0e8bee06101654fb0d288"} Dec 06 07:20:03 crc kubenswrapper[4809]: I1206 07:20:03.248636 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"973f75f0-3e24-4648-b57b-864065deb9b1","Type":"ContainerStarted","Data":"54afc4c460a902cc89e61598f063fb9f68c0ca80c46e96e0e702615a45bab812"} Dec 06 07:20:03 crc kubenswrapper[4809]: I1206 07:20:03.282039 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=6.145545522 podStartE2EDuration="1m18.282017138s" podCreationTimestamp="2025-12-06 07:18:45 +0000 UTC" firstStartedPulling="2025-12-06 07:18:47.891690959 +0000 UTC m=+5252.780673901" lastFinishedPulling="2025-12-06 07:20:00.028162575 +0000 UTC m=+5324.917145517" observedRunningTime="2025-12-06 07:20:03.273774705 +0000 UTC m=+5328.162757657" watchObservedRunningTime="2025-12-06 07:20:03.282017138 +0000 UTC m=+5328.171000080" Dec 06 07:20:04 crc kubenswrapper[4809]: I1206 07:20:04.262119 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-58pxg" event={"ID":"fde5c8b4-ad5d-48ae-bc7f-3464d809dc92","Type":"ContainerStarted","Data":"27d2cb0372457e7c2300eebef04e2cec3329f84e02ba9109a5b95d946f032c49"} Dec 06 07:20:04 crc kubenswrapper[4809]: I1206 07:20:04.264511 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6p627" event={"ID":"1435a636-4c7d-44fc-ae60-c1ebc049d0c3","Type":"ContainerStarted","Data":"4ad240cc3800ccf336f10cd13b9b80ee925ffc625ab64507e49bb53adcccd5de"} Dec 06 07:20:04 crc kubenswrapper[4809]: I1206 07:20:04.266919 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sh8jm" event={"ID":"8f5f3c13-919e-490d-9796-80d82cd12a53","Type":"ContainerStarted","Data":"afa7e68136210e6e2ea57a2bbb5e73e2c7cd747569d1a6dc5095fd09043791c8"} Dec 06 07:20:04 crc kubenswrapper[4809]: I1206 07:20:04.787916 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-sh8jm" podStartSLOduration=4.912990197 podStartE2EDuration="8.787894205s" podCreationTimestamp="2025-12-06 07:19:56 +0000 UTC" firstStartedPulling="2025-12-06 07:19:59.183962903 +0000 UTC m=+5324.072945835" lastFinishedPulling="2025-12-06 07:20:03.058866901 +0000 UTC m=+5327.947849843" observedRunningTime="2025-12-06 07:20:04.785660494 +0000 UTC m=+5329.674643456" watchObservedRunningTime="2025-12-06 07:20:04.787894205 +0000 UTC m=+5329.676877147" Dec 06 07:20:07 crc kubenswrapper[4809]: I1206 07:20:07.268023 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-sh8jm" Dec 06 07:20:07 crc kubenswrapper[4809]: I1206 07:20:07.268466 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-sh8jm" Dec 06 07:20:08 crc kubenswrapper[4809]: I1206 07:20:08.320868 4809 generic.go:334] "Generic (PLEG): container finished" podID="1435a636-4c7d-44fc-ae60-c1ebc049d0c3" containerID="4ad240cc3800ccf336f10cd13b9b80ee925ffc625ab64507e49bb53adcccd5de" exitCode=0 Dec 06 07:20:08 crc kubenswrapper[4809]: I1206 07:20:08.320954 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6p627" event={"ID":"1435a636-4c7d-44fc-ae60-c1ebc049d0c3","Type":"ContainerDied","Data":"4ad240cc3800ccf336f10cd13b9b80ee925ffc625ab64507e49bb53adcccd5de"} Dec 06 07:20:08 crc kubenswrapper[4809]: I1206 07:20:08.327854 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-sh8jm" podUID="8f5f3c13-919e-490d-9796-80d82cd12a53" containerName="registry-server" probeResult="failure" output=< Dec 06 07:20:08 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Dec 06 07:20:08 crc kubenswrapper[4809]: > Dec 06 07:20:09 crc kubenswrapper[4809]: I1206 07:20:09.389093 4809 scope.go:117] "RemoveContainer" containerID="af73883c3be3c80377e0f4806f2ecfa3ce57913ac4edf4c8383f744be4172b66" Dec 06 07:20:09 crc kubenswrapper[4809]: E1206 07:20:09.389404 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:20:11 crc kubenswrapper[4809]: I1206 07:20:11.366682 4809 generic.go:334] "Generic (PLEG): container finished" podID="fde5c8b4-ad5d-48ae-bc7f-3464d809dc92" containerID="27d2cb0372457e7c2300eebef04e2cec3329f84e02ba9109a5b95d946f032c49" exitCode=0 Dec 06 07:20:11 crc kubenswrapper[4809]: I1206 07:20:11.367012 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-58pxg" event={"ID":"fde5c8b4-ad5d-48ae-bc7f-3464d809dc92","Type":"ContainerDied","Data":"27d2cb0372457e7c2300eebef04e2cec3329f84e02ba9109a5b95d946f032c49"} Dec 06 07:20:11 crc kubenswrapper[4809]: I1206 07:20:11.370144 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6p627" event={"ID":"1435a636-4c7d-44fc-ae60-c1ebc049d0c3","Type":"ContainerStarted","Data":"0b246055e4db31addf5338d300fd1da85bef62ebc4d9b4302a50c949e69825db"} Dec 06 07:20:11 crc kubenswrapper[4809]: I1206 07:20:11.415992 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6p627" podStartSLOduration=3.735156875 podStartE2EDuration="13.415977029s" podCreationTimestamp="2025-12-06 07:19:58 +0000 UTC" firstStartedPulling="2025-12-06 07:20:01.208745921 +0000 UTC m=+5326.097728863" lastFinishedPulling="2025-12-06 07:20:10.889566075 +0000 UTC m=+5335.778549017" observedRunningTime="2025-12-06 07:20:11.40828195 +0000 UTC m=+5336.297264892" watchObservedRunningTime="2025-12-06 07:20:11.415977029 +0000 UTC m=+5336.304959961" Dec 06 07:20:12 crc kubenswrapper[4809]: I1206 07:20:12.384317 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-58pxg" event={"ID":"fde5c8b4-ad5d-48ae-bc7f-3464d809dc92","Type":"ContainerStarted","Data":"5eb2377b654fa25c45f0975250841356862017683a028ecf6a522c38765039a0"} Dec 06 07:20:12 crc kubenswrapper[4809]: I1206 07:20:12.416258 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-58pxg" podStartSLOduration=2.816497854 podStartE2EDuration="13.416235443s" podCreationTimestamp="2025-12-06 07:19:59 +0000 UTC" firstStartedPulling="2025-12-06 07:20:01.205539243 +0000 UTC m=+5326.094522185" lastFinishedPulling="2025-12-06 07:20:11.805276832 +0000 UTC m=+5336.694259774" observedRunningTime="2025-12-06 07:20:12.406441538 +0000 UTC m=+5337.295424490" watchObservedRunningTime="2025-12-06 07:20:12.416235443 +0000 UTC m=+5337.305218395" Dec 06 07:20:17 crc kubenswrapper[4809]: I1206 07:20:17.337699 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-sh8jm" Dec 06 07:20:17 crc kubenswrapper[4809]: I1206 07:20:17.413548 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-sh8jm" Dec 06 07:20:19 crc kubenswrapper[4809]: I1206 07:20:19.332758 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6p627" Dec 06 07:20:19 crc kubenswrapper[4809]: I1206 07:20:19.333580 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6p627" Dec 06 07:20:19 crc kubenswrapper[4809]: I1206 07:20:19.405687 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6p627" Dec 06 07:20:19 crc kubenswrapper[4809]: I1206 07:20:19.522284 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6p627" Dec 06 07:20:19 crc kubenswrapper[4809]: I1206 07:20:19.627940 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-58pxg" Dec 06 07:20:19 crc kubenswrapper[4809]: I1206 07:20:19.628008 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-58pxg" Dec 06 07:20:20 crc kubenswrapper[4809]: I1206 07:20:20.699252 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-58pxg" podUID="fde5c8b4-ad5d-48ae-bc7f-3464d809dc92" containerName="registry-server" probeResult="failure" output=< Dec 06 07:20:20 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Dec 06 07:20:20 crc kubenswrapper[4809]: > Dec 06 07:20:21 crc kubenswrapper[4809]: I1206 07:20:21.389908 4809 scope.go:117] "RemoveContainer" containerID="af73883c3be3c80377e0f4806f2ecfa3ce57913ac4edf4c8383f744be4172b66" Dec 06 07:20:21 crc kubenswrapper[4809]: E1206 07:20:21.390574 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:20:21 crc kubenswrapper[4809]: I1206 07:20:21.778509 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sh8jm"] Dec 06 07:20:21 crc kubenswrapper[4809]: I1206 07:20:21.778972 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-sh8jm" podUID="8f5f3c13-919e-490d-9796-80d82cd12a53" containerName="registry-server" containerID="cri-o://afa7e68136210e6e2ea57a2bbb5e73e2c7cd747569d1a6dc5095fd09043791c8" gracePeriod=2 Dec 06 07:20:22 crc kubenswrapper[4809]: I1206 07:20:22.504093 4809 generic.go:334] "Generic (PLEG): container finished" podID="8f5f3c13-919e-490d-9796-80d82cd12a53" containerID="afa7e68136210e6e2ea57a2bbb5e73e2c7cd747569d1a6dc5095fd09043791c8" exitCode=0 Dec 06 07:20:22 crc kubenswrapper[4809]: I1206 07:20:22.504169 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sh8jm" event={"ID":"8f5f3c13-919e-490d-9796-80d82cd12a53","Type":"ContainerDied","Data":"afa7e68136210e6e2ea57a2bbb5e73e2c7cd747569d1a6dc5095fd09043791c8"} Dec 06 07:20:22 crc kubenswrapper[4809]: I1206 07:20:22.970510 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6p627"] Dec 06 07:20:22 crc kubenswrapper[4809]: I1206 07:20:22.971042 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6p627" podUID="1435a636-4c7d-44fc-ae60-c1ebc049d0c3" containerName="registry-server" containerID="cri-o://0b246055e4db31addf5338d300fd1da85bef62ebc4d9b4302a50c949e69825db" gracePeriod=2 Dec 06 07:20:23 crc kubenswrapper[4809]: I1206 07:20:23.217556 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sh8jm" Dec 06 07:20:23 crc kubenswrapper[4809]: I1206 07:20:23.330589 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzr2h\" (UniqueName: \"kubernetes.io/projected/8f5f3c13-919e-490d-9796-80d82cd12a53-kube-api-access-fzr2h\") pod \"8f5f3c13-919e-490d-9796-80d82cd12a53\" (UID: \"8f5f3c13-919e-490d-9796-80d82cd12a53\") " Dec 06 07:20:23 crc kubenswrapper[4809]: I1206 07:20:23.330860 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f5f3c13-919e-490d-9796-80d82cd12a53-catalog-content\") pod \"8f5f3c13-919e-490d-9796-80d82cd12a53\" (UID: \"8f5f3c13-919e-490d-9796-80d82cd12a53\") " Dec 06 07:20:23 crc kubenswrapper[4809]: I1206 07:20:23.330896 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f5f3c13-919e-490d-9796-80d82cd12a53-utilities\") pod \"8f5f3c13-919e-490d-9796-80d82cd12a53\" (UID: \"8f5f3c13-919e-490d-9796-80d82cd12a53\") " Dec 06 07:20:23 crc kubenswrapper[4809]: I1206 07:20:23.333707 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f5f3c13-919e-490d-9796-80d82cd12a53-utilities" (OuterVolumeSpecName: "utilities") pod "8f5f3c13-919e-490d-9796-80d82cd12a53" (UID: "8f5f3c13-919e-490d-9796-80d82cd12a53"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:20:23 crc kubenswrapper[4809]: I1206 07:20:23.339761 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f5f3c13-919e-490d-9796-80d82cd12a53-kube-api-access-fzr2h" (OuterVolumeSpecName: "kube-api-access-fzr2h") pod "8f5f3c13-919e-490d-9796-80d82cd12a53" (UID: "8f5f3c13-919e-490d-9796-80d82cd12a53"). InnerVolumeSpecName "kube-api-access-fzr2h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:20:23 crc kubenswrapper[4809]: I1206 07:20:23.369853 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f5f3c13-919e-490d-9796-80d82cd12a53-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8f5f3c13-919e-490d-9796-80d82cd12a53" (UID: "8f5f3c13-919e-490d-9796-80d82cd12a53"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:20:23 crc kubenswrapper[4809]: I1206 07:20:23.434478 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f5f3c13-919e-490d-9796-80d82cd12a53-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:23 crc kubenswrapper[4809]: I1206 07:20:23.434515 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f5f3c13-919e-490d-9796-80d82cd12a53-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:23 crc kubenswrapper[4809]: I1206 07:20:23.434527 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzr2h\" (UniqueName: \"kubernetes.io/projected/8f5f3c13-919e-490d-9796-80d82cd12a53-kube-api-access-fzr2h\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:23 crc kubenswrapper[4809]: I1206 07:20:23.517258 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sh8jm" event={"ID":"8f5f3c13-919e-490d-9796-80d82cd12a53","Type":"ContainerDied","Data":"0d06c24ef1bfd7fc7b45aa59961e5c70ca0f7cd0b93352ddd22197101138ac17"} Dec 06 07:20:23 crc kubenswrapper[4809]: I1206 07:20:23.517342 4809 scope.go:117] "RemoveContainer" containerID="afa7e68136210e6e2ea57a2bbb5e73e2c7cd747569d1a6dc5095fd09043791c8" Dec 06 07:20:23 crc kubenswrapper[4809]: I1206 07:20:23.517347 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sh8jm" Dec 06 07:20:23 crc kubenswrapper[4809]: I1206 07:20:23.522265 4809 generic.go:334] "Generic (PLEG): container finished" podID="1435a636-4c7d-44fc-ae60-c1ebc049d0c3" containerID="0b246055e4db31addf5338d300fd1da85bef62ebc4d9b4302a50c949e69825db" exitCode=0 Dec 06 07:20:23 crc kubenswrapper[4809]: I1206 07:20:23.522310 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6p627" event={"ID":"1435a636-4c7d-44fc-ae60-c1ebc049d0c3","Type":"ContainerDied","Data":"0b246055e4db31addf5338d300fd1da85bef62ebc4d9b4302a50c949e69825db"} Dec 06 07:20:23 crc kubenswrapper[4809]: I1206 07:20:23.522336 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6p627" event={"ID":"1435a636-4c7d-44fc-ae60-c1ebc049d0c3","Type":"ContainerDied","Data":"eb567013cbbf13d58109b66398303c9f00a949d29eb136d73b8889392f06da5a"} Dec 06 07:20:23 crc kubenswrapper[4809]: I1206 07:20:23.522347 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb567013cbbf13d58109b66398303c9f00a949d29eb136d73b8889392f06da5a" Dec 06 07:20:23 crc kubenswrapper[4809]: I1206 07:20:23.532882 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6p627" Dec 06 07:20:23 crc kubenswrapper[4809]: I1206 07:20:23.554521 4809 scope.go:117] "RemoveContainer" containerID="dad9f45d255826350ce9093c7c8f08c4c68726d32eb0e8bee06101654fb0d288" Dec 06 07:20:23 crc kubenswrapper[4809]: I1206 07:20:23.557166 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sh8jm"] Dec 06 07:20:23 crc kubenswrapper[4809]: I1206 07:20:23.571923 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-sh8jm"] Dec 06 07:20:23 crc kubenswrapper[4809]: I1206 07:20:23.591023 4809 scope.go:117] "RemoveContainer" containerID="4fca7c336c7eb96d2b9d5a117a50b6807503243c5239d55959e73cd8d339fb96" Dec 06 07:20:23 crc kubenswrapper[4809]: I1206 07:20:23.637549 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1435a636-4c7d-44fc-ae60-c1ebc049d0c3-utilities\") pod \"1435a636-4c7d-44fc-ae60-c1ebc049d0c3\" (UID: \"1435a636-4c7d-44fc-ae60-c1ebc049d0c3\") " Dec 06 07:20:23 crc kubenswrapper[4809]: I1206 07:20:23.637672 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1435a636-4c7d-44fc-ae60-c1ebc049d0c3-catalog-content\") pod \"1435a636-4c7d-44fc-ae60-c1ebc049d0c3\" (UID: \"1435a636-4c7d-44fc-ae60-c1ebc049d0c3\") " Dec 06 07:20:23 crc kubenswrapper[4809]: I1206 07:20:23.637785 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4c2c\" (UniqueName: \"kubernetes.io/projected/1435a636-4c7d-44fc-ae60-c1ebc049d0c3-kube-api-access-f4c2c\") pod \"1435a636-4c7d-44fc-ae60-c1ebc049d0c3\" (UID: \"1435a636-4c7d-44fc-ae60-c1ebc049d0c3\") " Dec 06 07:20:23 crc kubenswrapper[4809]: I1206 07:20:23.638277 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1435a636-4c7d-44fc-ae60-c1ebc049d0c3-utilities" (OuterVolumeSpecName: "utilities") pod "1435a636-4c7d-44fc-ae60-c1ebc049d0c3" (UID: "1435a636-4c7d-44fc-ae60-c1ebc049d0c3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:20:23 crc kubenswrapper[4809]: I1206 07:20:23.638425 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1435a636-4c7d-44fc-ae60-c1ebc049d0c3-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:23 crc kubenswrapper[4809]: I1206 07:20:23.641759 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1435a636-4c7d-44fc-ae60-c1ebc049d0c3-kube-api-access-f4c2c" (OuterVolumeSpecName: "kube-api-access-f4c2c") pod "1435a636-4c7d-44fc-ae60-c1ebc049d0c3" (UID: "1435a636-4c7d-44fc-ae60-c1ebc049d0c3"). InnerVolumeSpecName "kube-api-access-f4c2c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:20:23 crc kubenswrapper[4809]: I1206 07:20:23.685188 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1435a636-4c7d-44fc-ae60-c1ebc049d0c3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1435a636-4c7d-44fc-ae60-c1ebc049d0c3" (UID: "1435a636-4c7d-44fc-ae60-c1ebc049d0c3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:20:23 crc kubenswrapper[4809]: I1206 07:20:23.741096 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1435a636-4c7d-44fc-ae60-c1ebc049d0c3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:23 crc kubenswrapper[4809]: I1206 07:20:23.741155 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4c2c\" (UniqueName: \"kubernetes.io/projected/1435a636-4c7d-44fc-ae60-c1ebc049d0c3-kube-api-access-f4c2c\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:24 crc kubenswrapper[4809]: I1206 07:20:24.534715 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6p627" Dec 06 07:20:24 crc kubenswrapper[4809]: I1206 07:20:24.568423 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6p627"] Dec 06 07:20:24 crc kubenswrapper[4809]: I1206 07:20:24.579814 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6p627"] Dec 06 07:20:25 crc kubenswrapper[4809]: I1206 07:20:25.417706 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1435a636-4c7d-44fc-ae60-c1ebc049d0c3" path="/var/lib/kubelet/pods/1435a636-4c7d-44fc-ae60-c1ebc049d0c3/volumes" Dec 06 07:20:25 crc kubenswrapper[4809]: I1206 07:20:25.419444 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f5f3c13-919e-490d-9796-80d82cd12a53" path="/var/lib/kubelet/pods/8f5f3c13-919e-490d-9796-80d82cd12a53/volumes" Dec 06 07:20:29 crc kubenswrapper[4809]: I1206 07:20:29.679860 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-58pxg" Dec 06 07:20:29 crc kubenswrapper[4809]: I1206 07:20:29.732536 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-58pxg" Dec 06 07:20:30 crc kubenswrapper[4809]: I1206 07:20:30.783539 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-58pxg"] Dec 06 07:20:31 crc kubenswrapper[4809]: I1206 07:20:31.615282 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-58pxg" podUID="fde5c8b4-ad5d-48ae-bc7f-3464d809dc92" containerName="registry-server" containerID="cri-o://5eb2377b654fa25c45f0975250841356862017683a028ecf6a522c38765039a0" gracePeriod=2 Dec 06 07:20:32 crc kubenswrapper[4809]: I1206 07:20:32.243734 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-58pxg" Dec 06 07:20:32 crc kubenswrapper[4809]: I1206 07:20:32.293432 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fde5c8b4-ad5d-48ae-bc7f-3464d809dc92-utilities\") pod \"fde5c8b4-ad5d-48ae-bc7f-3464d809dc92\" (UID: \"fde5c8b4-ad5d-48ae-bc7f-3464d809dc92\") " Dec 06 07:20:32 crc kubenswrapper[4809]: I1206 07:20:32.293588 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbttt\" (UniqueName: \"kubernetes.io/projected/fde5c8b4-ad5d-48ae-bc7f-3464d809dc92-kube-api-access-rbttt\") pod \"fde5c8b4-ad5d-48ae-bc7f-3464d809dc92\" (UID: \"fde5c8b4-ad5d-48ae-bc7f-3464d809dc92\") " Dec 06 07:20:32 crc kubenswrapper[4809]: I1206 07:20:32.293705 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fde5c8b4-ad5d-48ae-bc7f-3464d809dc92-catalog-content\") pod \"fde5c8b4-ad5d-48ae-bc7f-3464d809dc92\" (UID: \"fde5c8b4-ad5d-48ae-bc7f-3464d809dc92\") " Dec 06 07:20:32 crc kubenswrapper[4809]: I1206 07:20:32.294620 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fde5c8b4-ad5d-48ae-bc7f-3464d809dc92-utilities" (OuterVolumeSpecName: "utilities") pod "fde5c8b4-ad5d-48ae-bc7f-3464d809dc92" (UID: "fde5c8b4-ad5d-48ae-bc7f-3464d809dc92"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:20:32 crc kubenswrapper[4809]: I1206 07:20:32.299395 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fde5c8b4-ad5d-48ae-bc7f-3464d809dc92-kube-api-access-rbttt" (OuterVolumeSpecName: "kube-api-access-rbttt") pod "fde5c8b4-ad5d-48ae-bc7f-3464d809dc92" (UID: "fde5c8b4-ad5d-48ae-bc7f-3464d809dc92"). InnerVolumeSpecName "kube-api-access-rbttt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:20:32 crc kubenswrapper[4809]: I1206 07:20:32.392807 4809 scope.go:117] "RemoveContainer" containerID="af73883c3be3c80377e0f4806f2ecfa3ce57913ac4edf4c8383f744be4172b66" Dec 06 07:20:32 crc kubenswrapper[4809]: E1206 07:20:32.394682 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:20:32 crc kubenswrapper[4809]: I1206 07:20:32.396525 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbttt\" (UniqueName: \"kubernetes.io/projected/fde5c8b4-ad5d-48ae-bc7f-3464d809dc92-kube-api-access-rbttt\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:32 crc kubenswrapper[4809]: I1206 07:20:32.396560 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fde5c8b4-ad5d-48ae-bc7f-3464d809dc92-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:32 crc kubenswrapper[4809]: I1206 07:20:32.427326 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fde5c8b4-ad5d-48ae-bc7f-3464d809dc92-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fde5c8b4-ad5d-48ae-bc7f-3464d809dc92" (UID: "fde5c8b4-ad5d-48ae-bc7f-3464d809dc92"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:20:32 crc kubenswrapper[4809]: I1206 07:20:32.499461 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fde5c8b4-ad5d-48ae-bc7f-3464d809dc92-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 07:20:32 crc kubenswrapper[4809]: I1206 07:20:32.637217 4809 generic.go:334] "Generic (PLEG): container finished" podID="fde5c8b4-ad5d-48ae-bc7f-3464d809dc92" containerID="5eb2377b654fa25c45f0975250841356862017683a028ecf6a522c38765039a0" exitCode=0 Dec 06 07:20:32 crc kubenswrapper[4809]: I1206 07:20:32.637277 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-58pxg" event={"ID":"fde5c8b4-ad5d-48ae-bc7f-3464d809dc92","Type":"ContainerDied","Data":"5eb2377b654fa25c45f0975250841356862017683a028ecf6a522c38765039a0"} Dec 06 07:20:32 crc kubenswrapper[4809]: I1206 07:20:32.637323 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-58pxg" event={"ID":"fde5c8b4-ad5d-48ae-bc7f-3464d809dc92","Type":"ContainerDied","Data":"bcd1d79370b2423ba7423b7caa3c83d9d38464f0456480322985038fc6d884eb"} Dec 06 07:20:32 crc kubenswrapper[4809]: I1206 07:20:32.637340 4809 scope.go:117] "RemoveContainer" containerID="5eb2377b654fa25c45f0975250841356862017683a028ecf6a522c38765039a0" Dec 06 07:20:32 crc kubenswrapper[4809]: I1206 07:20:32.637336 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-58pxg" Dec 06 07:20:32 crc kubenswrapper[4809]: I1206 07:20:32.678038 4809 scope.go:117] "RemoveContainer" containerID="27d2cb0372457e7c2300eebef04e2cec3329f84e02ba9109a5b95d946f032c49" Dec 06 07:20:32 crc kubenswrapper[4809]: I1206 07:20:32.680003 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-58pxg"] Dec 06 07:20:32 crc kubenswrapper[4809]: I1206 07:20:32.691712 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-58pxg"] Dec 06 07:20:32 crc kubenswrapper[4809]: I1206 07:20:32.713746 4809 scope.go:117] "RemoveContainer" containerID="e60d977c0f934e0236adce381ec03c98f2ff6f04a49e656ed3235d17a7d41ce9" Dec 06 07:20:32 crc kubenswrapper[4809]: I1206 07:20:32.763440 4809 scope.go:117] "RemoveContainer" containerID="5eb2377b654fa25c45f0975250841356862017683a028ecf6a522c38765039a0" Dec 06 07:20:32 crc kubenswrapper[4809]: E1206 07:20:32.763889 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5eb2377b654fa25c45f0975250841356862017683a028ecf6a522c38765039a0\": container with ID starting with 5eb2377b654fa25c45f0975250841356862017683a028ecf6a522c38765039a0 not found: ID does not exist" containerID="5eb2377b654fa25c45f0975250841356862017683a028ecf6a522c38765039a0" Dec 06 07:20:32 crc kubenswrapper[4809]: I1206 07:20:32.763922 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5eb2377b654fa25c45f0975250841356862017683a028ecf6a522c38765039a0"} err="failed to get container status \"5eb2377b654fa25c45f0975250841356862017683a028ecf6a522c38765039a0\": rpc error: code = NotFound desc = could not find container \"5eb2377b654fa25c45f0975250841356862017683a028ecf6a522c38765039a0\": container with ID starting with 5eb2377b654fa25c45f0975250841356862017683a028ecf6a522c38765039a0 not found: ID does not exist" Dec 06 07:20:32 crc kubenswrapper[4809]: I1206 07:20:32.763956 4809 scope.go:117] "RemoveContainer" containerID="27d2cb0372457e7c2300eebef04e2cec3329f84e02ba9109a5b95d946f032c49" Dec 06 07:20:32 crc kubenswrapper[4809]: E1206 07:20:32.764335 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27d2cb0372457e7c2300eebef04e2cec3329f84e02ba9109a5b95d946f032c49\": container with ID starting with 27d2cb0372457e7c2300eebef04e2cec3329f84e02ba9109a5b95d946f032c49 not found: ID does not exist" containerID="27d2cb0372457e7c2300eebef04e2cec3329f84e02ba9109a5b95d946f032c49" Dec 06 07:20:32 crc kubenswrapper[4809]: I1206 07:20:32.764360 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27d2cb0372457e7c2300eebef04e2cec3329f84e02ba9109a5b95d946f032c49"} err="failed to get container status \"27d2cb0372457e7c2300eebef04e2cec3329f84e02ba9109a5b95d946f032c49\": rpc error: code = NotFound desc = could not find container \"27d2cb0372457e7c2300eebef04e2cec3329f84e02ba9109a5b95d946f032c49\": container with ID starting with 27d2cb0372457e7c2300eebef04e2cec3329f84e02ba9109a5b95d946f032c49 not found: ID does not exist" Dec 06 07:20:32 crc kubenswrapper[4809]: I1206 07:20:32.764374 4809 scope.go:117] "RemoveContainer" containerID="e60d977c0f934e0236adce381ec03c98f2ff6f04a49e656ed3235d17a7d41ce9" Dec 06 07:20:32 crc kubenswrapper[4809]: E1206 07:20:32.764630 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e60d977c0f934e0236adce381ec03c98f2ff6f04a49e656ed3235d17a7d41ce9\": container with ID starting with e60d977c0f934e0236adce381ec03c98f2ff6f04a49e656ed3235d17a7d41ce9 not found: ID does not exist" containerID="e60d977c0f934e0236adce381ec03c98f2ff6f04a49e656ed3235d17a7d41ce9" Dec 06 07:20:32 crc kubenswrapper[4809]: I1206 07:20:32.764658 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e60d977c0f934e0236adce381ec03c98f2ff6f04a49e656ed3235d17a7d41ce9"} err="failed to get container status \"e60d977c0f934e0236adce381ec03c98f2ff6f04a49e656ed3235d17a7d41ce9\": rpc error: code = NotFound desc = could not find container \"e60d977c0f934e0236adce381ec03c98f2ff6f04a49e656ed3235d17a7d41ce9\": container with ID starting with e60d977c0f934e0236adce381ec03c98f2ff6f04a49e656ed3235d17a7d41ce9 not found: ID does not exist" Dec 06 07:20:33 crc kubenswrapper[4809]: I1206 07:20:33.409044 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fde5c8b4-ad5d-48ae-bc7f-3464d809dc92" path="/var/lib/kubelet/pods/fde5c8b4-ad5d-48ae-bc7f-3464d809dc92/volumes" Dec 06 07:20:44 crc kubenswrapper[4809]: I1206 07:20:44.389388 4809 scope.go:117] "RemoveContainer" containerID="af73883c3be3c80377e0f4806f2ecfa3ce57913ac4edf4c8383f744be4172b66" Dec 06 07:20:44 crc kubenswrapper[4809]: E1206 07:20:44.390444 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:20:59 crc kubenswrapper[4809]: I1206 07:20:59.392014 4809 scope.go:117] "RemoveContainer" containerID="af73883c3be3c80377e0f4806f2ecfa3ce57913ac4edf4c8383f744be4172b66" Dec 06 07:20:59 crc kubenswrapper[4809]: E1206 07:20:59.393228 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:21:10 crc kubenswrapper[4809]: I1206 07:21:10.392664 4809 scope.go:117] "RemoveContainer" containerID="af73883c3be3c80377e0f4806f2ecfa3ce57913ac4edf4c8383f744be4172b66" Dec 06 07:21:10 crc kubenswrapper[4809]: E1206 07:21:10.393240 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:21:21 crc kubenswrapper[4809]: I1206 07:21:21.389695 4809 scope.go:117] "RemoveContainer" containerID="af73883c3be3c80377e0f4806f2ecfa3ce57913ac4edf4c8383f744be4172b66" Dec 06 07:21:21 crc kubenswrapper[4809]: E1206 07:21:21.390579 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:21:32 crc kubenswrapper[4809]: I1206 07:21:32.388577 4809 scope.go:117] "RemoveContainer" containerID="af73883c3be3c80377e0f4806f2ecfa3ce57913ac4edf4c8383f744be4172b66" Dec 06 07:21:32 crc kubenswrapper[4809]: E1206 07:21:32.389296 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:21:43 crc kubenswrapper[4809]: I1206 07:21:43.389516 4809 scope.go:117] "RemoveContainer" containerID="af73883c3be3c80377e0f4806f2ecfa3ce57913ac4edf4c8383f744be4172b66" Dec 06 07:21:43 crc kubenswrapper[4809]: E1206 07:21:43.390366 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:21:56 crc kubenswrapper[4809]: I1206 07:21:56.388867 4809 scope.go:117] "RemoveContainer" containerID="af73883c3be3c80377e0f4806f2ecfa3ce57913ac4edf4c8383f744be4172b66" Dec 06 07:21:56 crc kubenswrapper[4809]: E1206 07:21:56.390098 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:22:07 crc kubenswrapper[4809]: I1206 07:22:07.388882 4809 scope.go:117] "RemoveContainer" containerID="af73883c3be3c80377e0f4806f2ecfa3ce57913ac4edf4c8383f744be4172b66" Dec 06 07:22:07 crc kubenswrapper[4809]: E1206 07:22:07.389817 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:22:20 crc kubenswrapper[4809]: I1206 07:22:20.388563 4809 scope.go:117] "RemoveContainer" containerID="af73883c3be3c80377e0f4806f2ecfa3ce57913ac4edf4c8383f744be4172b66" Dec 06 07:22:20 crc kubenswrapper[4809]: E1206 07:22:20.389484 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:22:31 crc kubenswrapper[4809]: I1206 07:22:31.389436 4809 scope.go:117] "RemoveContainer" containerID="af73883c3be3c80377e0f4806f2ecfa3ce57913ac4edf4c8383f744be4172b66" Dec 06 07:22:31 crc kubenswrapper[4809]: E1206 07:22:31.390478 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:22:35 crc kubenswrapper[4809]: I1206 07:22:35.519558 4809 trace.go:236] Trace[316470870]: "Calculate volume metrics of httpd-run for pod openstack/glance-default-internal-api-0" (06-Dec-2025 07:22:34.379) (total time: 1140ms): Dec 06 07:22:35 crc kubenswrapper[4809]: Trace[316470870]: [1.140232733s] [1.140232733s] END Dec 06 07:22:35 crc kubenswrapper[4809]: I1206 07:22:35.521240 4809 patch_prober.go:28] interesting pod/logging-loki-gateway-65498c4f8f-48rjq container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.53:8081/ready\": context deadline exceeded" start-of-body= Dec 06 07:22:35 crc kubenswrapper[4809]: I1206 07:22:35.527294 4809 patch_prober.go:28] interesting pod/console-operator-58897d9998-hlvjb container/console-operator namespace/openshift-console-operator: Liveness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:22:35 crc kubenswrapper[4809]: I1206 07:22:35.615368 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-65498c4f8f-48rjq" podUID="72a3ed95-3e3d-4faf-88e4-ad1731902910" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.53:8081/ready\": context deadline exceeded" Dec 06 07:22:35 crc kubenswrapper[4809]: I1206 07:22:35.617757 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console-operator/console-operator-58897d9998-hlvjb" podUID="f3a30985-f268-44de-93b9-bce0f9aebc6d" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:22:45 crc kubenswrapper[4809]: I1206 07:22:45.396787 4809 scope.go:117] "RemoveContainer" containerID="af73883c3be3c80377e0f4806f2ecfa3ce57913ac4edf4c8383f744be4172b66" Dec 06 07:22:46 crc kubenswrapper[4809]: I1206 07:22:46.733311 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" event={"ID":"cbf4cf62-024e-4703-a8b8-9aecda9cd26a","Type":"ContainerStarted","Data":"8685c71ca1e6bc659b1168658926b473eb3d43e2ebbafcbe1ef610af7804649e"} Dec 06 07:22:54 crc kubenswrapper[4809]: I1206 07:22:54.747586 4809 trace.go:236] Trace[581646009]: "Calculate volume metrics of config for pod openshift-etcd-operator/etcd-operator-b45778765-drj7v" (06-Dec-2025 07:22:53.575) (total time: 1171ms): Dec 06 07:22:54 crc kubenswrapper[4809]: Trace[581646009]: [1.171671613s] [1.171671613s] END Dec 06 07:23:34 crc kubenswrapper[4809]: I1206 07:23:34.152156 4809 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-r2lg6 container/marketplace-operator namespace/openshift-marketplace: Liveness probe status=failure output="Get \"http://10.217.0.63:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:23:34 crc kubenswrapper[4809]: I1206 07:23:34.152128 4809 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-r2lg6 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.63:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:23:34 crc kubenswrapper[4809]: I1206 07:23:34.163355 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/marketplace-operator-79b997595-r2lg6" podUID="286c8de4-0ea9-43ad-bb4f-970319e0f4d3" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.63:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:23:34 crc kubenswrapper[4809]: I1206 07:23:34.165245 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-r2lg6" podUID="286c8de4-0ea9-43ad-bb4f-970319e0f4d3" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.63:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:23:38 crc kubenswrapper[4809]: I1206 07:23:38.159235 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.620774066s: [/var/lib/containers/storage/overlay/7c95abf37ae97ac402377ac37e81439b5ebd4a29dc865c1c58c8097abef5f6fc/diff /var/log/pods/openstack_ceilometer-0_3d4b42d3-04a4-40e3-803b-11c77024f612/sg-core/0.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:23:38 crc kubenswrapper[4809]: I1206 07:23:38.200906 4809 trace.go:236] Trace[874278136]: "Calculate volume metrics of config for pod openshift-logging/logging-loki-query-frontend-84558f7c9f-9mw6j" (06-Dec-2025 07:23:36.675) (total time: 1481ms): Dec 06 07:23:38 crc kubenswrapper[4809]: Trace[874278136]: [1.481974019s] [1.481974019s] END Dec 06 07:24:15 crc kubenswrapper[4809]: I1206 07:24:15.985692 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.122451221s: [/var/lib/containers/storage/overlay/efad4a47b86230fa7fb8dfc84e01cbf95528ba82e72be4c38bc178fbdbf25480/diff /var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-697657f76c-5stmg_19fc8500-fddb-4650-839a-82166e98a567/manager/1.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:24:15 crc kubenswrapper[4809]: I1206 07:24:15.986882 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.153282376s: [/var/lib/containers/storage/overlay/d7b0ced6c801694832256ee9e51108fc436161a82f4b7a4b2a1bba08bb98ef70/diff /var/log/pods/cert-manager_cert-manager-5b446d88c5-wgc72_8bc6fec5-e0e7-416f-b7ac-0a671cac3c0b/cert-manager-controller/0.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:24:16 crc kubenswrapper[4809]: I1206 07:24:16.028598 4809 patch_prober.go:28] interesting pod/package-server-manager-789f6589d5-ch964 container/package-server-manager namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"http://10.217.0.17:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:24:16 crc kubenswrapper[4809]: I1206 07:24:16.029284 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ch964" podUID="b6b25f1e-33d4-481d-8c73-24b95c61364f" containerName="package-server-manager" probeResult="failure" output="Get \"http://10.217.0.17:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:24:16 crc kubenswrapper[4809]: I1206 07:24:16.047356 4809 patch_prober.go:28] interesting pod/package-server-manager-789f6589d5-ch964 container/package-server-manager namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"http://10.217.0.17:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:24:16 crc kubenswrapper[4809]: I1206 07:24:16.047478 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ch964" podUID="b6b25f1e-33d4-481d-8c73-24b95c61364f" containerName="package-server-manager" probeResult="failure" output="Get \"http://10.217.0.17:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:24:16 crc kubenswrapper[4809]: I1206 07:24:16.117733 4809 trace.go:236] Trace[295242394]: "Calculate volume metrics of storage for pod minio-dev/minio" (06-Dec-2025 07:24:15.114) (total time: 1000ms): Dec 06 07:24:16 crc kubenswrapper[4809]: Trace[295242394]: [1.000067621s] [1.000067621s] END Dec 06 07:24:30 crc kubenswrapper[4809]: I1206 07:24:30.364146 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-ttn9g" podUID="b483382f-c2e3-40bf-ab4b-b58c2f3f9132" containerName="frr-k8s-webhook-server" probeResult="failure" output="Get \"http://10.217.0.94:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:24:31 crc kubenswrapper[4809]: I1206 07:24:31.973416 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.176608557s: [/var/lib/containers/storage/overlay/135e0bce4f48e72d7a9948bea6faffbd06ed01c8f1a0d9814b9b3c1fd70ca0a8/diff /var/log/pods/openstack_ceilometer-0_3d4b42d3-04a4-40e3-803b-11c77024f612/ceilometer-central-agent/2.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:24:31 crc kubenswrapper[4809]: I1206 07:24:31.981365 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/speaker-5t9z6" podUID="21411d89-4a48-4556-8ad2-193f1c8be0db" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:24:31 crc kubenswrapper[4809]: I1206 07:24:31.981403 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/speaker-5t9z6" podUID="21411d89-4a48-4556-8ad2-193f1c8be0db" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:25:04 crc kubenswrapper[4809]: I1206 07:25:04.497064 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:25:04 crc kubenswrapper[4809]: I1206 07:25:04.498750 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:25:15 crc kubenswrapper[4809]: I1206 07:25:15.810473 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-9gfg2" podUID="ee834860-9609-482d-95db-82fa1a1a6941" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.104:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:25:15 crc kubenswrapper[4809]: I1206 07:25:15.842623 4809 patch_prober.go:28] interesting pod/metrics-server-556bbc5579-8wblw container/metrics-server namespace/openshift-monitoring: Liveness probe status=failure output="Get \"https://10.217.0.75:10250/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:25:15 crc kubenswrapper[4809]: I1206 07:25:15.842679 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/metrics-server-556bbc5579-8wblw" podUID="b4f96141-d70e-4715-9e30-516c9e783cc9" containerName="metrics-server" probeResult="failure" output="Get \"https://10.217.0.75:10250/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:25:34 crc kubenswrapper[4809]: I1206 07:25:34.560183 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:25:34 crc kubenswrapper[4809]: I1206 07:25:34.563257 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:25:37 crc kubenswrapper[4809]: I1206 07:25:37.450216 4809 trace.go:236] Trace[2109399485]: "Calculate volume metrics of prometheus-metric-storage-rulefiles-0 for pod openstack/prometheus-metric-storage-0" (06-Dec-2025 07:25:34.826) (total time: 2620ms): Dec 06 07:25:37 crc kubenswrapper[4809]: Trace[2109399485]: [2.620864606s] [2.620864606s] END Dec 06 07:25:37 crc kubenswrapper[4809]: I1206 07:25:37.484497 4809 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-jjf9r container/packageserver namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.34:5443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:25:37 crc kubenswrapper[4809]: E1206 07:25:37.503201 4809 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="2.112s" Dec 06 07:25:37 crc kubenswrapper[4809]: I1206 07:25:37.505687 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-controller-operator-6598f77bc9-xjfmg" podUID="2d9f04dc-3365-4b3e-b57c-604e657c6483" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.98:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:25:37 crc kubenswrapper[4809]: I1206 07:25:37.511610 4809 trace.go:236] Trace[721495269]: "Calculate volume metrics of nginx-conf for pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" (06-Dec-2025 07:25:34.636) (total time: 2806ms): Dec 06 07:25:37 crc kubenswrapper[4809]: Trace[721495269]: [2.806317223s] [2.806317223s] END Dec 06 07:25:37 crc kubenswrapper[4809]: I1206 07:25:37.512048 4809 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-gt9kb container/olm-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.19:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:25:37 crc kubenswrapper[4809]: I1206 07:25:37.512115 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gt9kb" podUID="c8d79e6f-40b0-4ecc-8501-454ba5cfe691" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.19:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:25:37 crc kubenswrapper[4809]: I1206 07:25:37.512234 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-operator-controller-operator-6598f77bc9-xjfmg" podUID="2d9f04dc-3365-4b3e-b57c-604e657c6483" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.98:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:25:37 crc kubenswrapper[4809]: I1206 07:25:37.512291 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-bhbcl" podUID="fd7dd0a5-5e39-4812-b93f-a27c39284d33" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.119:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:25:37 crc kubenswrapper[4809]: I1206 07:25:37.512419 4809 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-gt9kb container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.19:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:25:37 crc kubenswrapper[4809]: I1206 07:25:37.512440 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gt9kb" podUID="c8d79e6f-40b0-4ecc-8501-454ba5cfe691" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.19:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:25:37 crc kubenswrapper[4809]: I1206 07:25:37.484619 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jjf9r" podUID="e5e4e451-e3ee-434e-8e24-0c795925a48e" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.34:5443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:25:37 crc kubenswrapper[4809]: I1206 07:25:37.516245 4809 patch_prober.go:28] interesting pod/openshift-kube-scheduler-crc container/kube-scheduler namespace/openshift-kube-scheduler: Liveness probe status=failure output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:25:37 crc kubenswrapper[4809]: I1206 07:25:37.516295 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" probeResult="failure" output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:25:37 crc kubenswrapper[4809]: I1206 07:25:37.573415 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-ghwtq" podUID="b379307b-aafd-42da-b761-38a556a0d143" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.114:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:25:38 crc kubenswrapper[4809]: I1206 07:25:38.530891 4809 patch_prober.go:28] interesting pod/logging-loki-gateway-65498c4f8f-9s8qq container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.54:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:25:38 crc kubenswrapper[4809]: I1206 07:25:38.530992 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq" podUID="ef8f3e7e-57ea-42b8-a777-1778e5ed975b" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.54:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:25:38 crc kubenswrapper[4809]: I1206 07:25:38.534170 4809 patch_prober.go:28] interesting pod/logging-loki-gateway-65498c4f8f-48rjq container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.53:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:25:38 crc kubenswrapper[4809]: I1206 07:25:38.534216 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-65498c4f8f-48rjq" podUID="72a3ed95-3e3d-4faf-88e4-ad1731902910" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.53:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:25:38 crc kubenswrapper[4809]: I1206 07:25:38.907755 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/community-operators-4p5rc" podUID="6fa58271-edf3-4d3d-8782-83a959a691fe" containerName="registry-server" probeResult="failure" output=< Dec 06 07:25:38 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Dec 06 07:25:38 crc kubenswrapper[4809]: > Dec 06 07:25:38 crc kubenswrapper[4809]: I1206 07:25:38.908249 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/certified-operators-nqpwx" podUID="408b531c-9f2a-4ee5-ad65-db67b2313be3" containerName="registry-server" probeResult="failure" output=< Dec 06 07:25:38 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Dec 06 07:25:38 crc kubenswrapper[4809]: > Dec 06 07:25:38 crc kubenswrapper[4809]: I1206 07:25:38.908317 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/certified-operators-nqpwx" podUID="408b531c-9f2a-4ee5-ad65-db67b2313be3" containerName="registry-server" probeResult="failure" output=< Dec 06 07:25:38 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Dec 06 07:25:38 crc kubenswrapper[4809]: > Dec 06 07:25:38 crc kubenswrapper[4809]: I1206 07:25:38.908352 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/community-operators-4p5rc" podUID="6fa58271-edf3-4d3d-8782-83a959a691fe" containerName="registry-server" probeResult="failure" output=< Dec 06 07:25:38 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Dec 06 07:25:38 crc kubenswrapper[4809]: > Dec 06 07:25:38 crc kubenswrapper[4809]: I1206 07:25:38.909855 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-cell1-galera-0" podUID="e7056938-6e36-4631-adfb-01f116eba651" containerName="galera" probeResult="failure" output="command timed out" Dec 06 07:25:38 crc kubenswrapper[4809]: I1206 07:25:38.910372 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="e7056938-6e36-4631-adfb-01f116eba651" containerName="galera" probeResult="failure" output="command timed out" Dec 06 07:25:39 crc kubenswrapper[4809]: I1206 07:25:39.103777 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-zjvtm" podUID="0714d252-8cd5-4a99-9050-8e383ad64885" containerName="registry-server" probeResult="failure" output=< Dec 06 07:25:39 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Dec 06 07:25:39 crc kubenswrapper[4809]: > Dec 06 07:25:39 crc kubenswrapper[4809]: I1206 07:25:39.105171 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/redhat-marketplace-zjvtm" podUID="0714d252-8cd5-4a99-9050-8e383ad64885" containerName="registry-server" probeResult="failure" output=< Dec 06 07:25:39 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Dec 06 07:25:39 crc kubenswrapper[4809]: > Dec 06 07:25:39 crc kubenswrapper[4809]: I1206 07:25:39.778795 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="7c596e14-b96f-4e5f-a8ac-3dc66fa0276d" containerName="galera" probeResult="failure" output="command timed out" Dec 06 07:25:39 crc kubenswrapper[4809]: I1206 07:25:39.778845 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-galera-0" podUID="7c596e14-b96f-4e5f-a8ac-3dc66fa0276d" containerName="galera" probeResult="failure" output="command timed out" Dec 06 07:26:04 crc kubenswrapper[4809]: I1206 07:26:04.497463 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:26:04 crc kubenswrapper[4809]: I1206 07:26:04.499347 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:26:04 crc kubenswrapper[4809]: I1206 07:26:04.500119 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-npms2" Dec 06 07:26:04 crc kubenswrapper[4809]: I1206 07:26:04.501591 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8685c71ca1e6bc659b1168658926b473eb3d43e2ebbafcbe1ef610af7804649e"} pod="openshift-machine-config-operator/machine-config-daemon-npms2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 07:26:04 crc kubenswrapper[4809]: I1206 07:26:04.501913 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" containerID="cri-o://8685c71ca1e6bc659b1168658926b473eb3d43e2ebbafcbe1ef610af7804649e" gracePeriod=600 Dec 06 07:26:05 crc kubenswrapper[4809]: I1206 07:26:05.881366 4809 generic.go:334] "Generic (PLEG): container finished" podID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerID="8685c71ca1e6bc659b1168658926b473eb3d43e2ebbafcbe1ef610af7804649e" exitCode=0 Dec 06 07:26:05 crc kubenswrapper[4809]: I1206 07:26:05.881390 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" event={"ID":"cbf4cf62-024e-4703-a8b8-9aecda9cd26a","Type":"ContainerDied","Data":"8685c71ca1e6bc659b1168658926b473eb3d43e2ebbafcbe1ef610af7804649e"} Dec 06 07:26:05 crc kubenswrapper[4809]: I1206 07:26:05.882738 4809 scope.go:117] "RemoveContainer" containerID="af73883c3be3c80377e0f4806f2ecfa3ce57913ac4edf4c8383f744be4172b66" Dec 06 07:26:06 crc kubenswrapper[4809]: I1206 07:26:06.896427 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" event={"ID":"cbf4cf62-024e-4703-a8b8-9aecda9cd26a","Type":"ContainerStarted","Data":"26a5b8063411f48708112d11ebddb0587d0fe6fc8f6a5c0aa5728fedd5b56f1d"} Dec 06 07:26:36 crc kubenswrapper[4809]: I1206 07:26:36.582001 4809 trace.go:236] Trace[1824413183]: "Calculate volume metrics of telemetry-config for pod openshift-monitoring/cluster-monitoring-operator-6d5b84845-ln5ns" (06-Dec-2025 07:26:35.546) (total time: 1033ms): Dec 06 07:26:36 crc kubenswrapper[4809]: Trace[1824413183]: [1.03368752s] [1.03368752s] END Dec 06 07:26:39 crc kubenswrapper[4809]: I1206 07:26:39.376050 4809 scope.go:117] "RemoveContainer" containerID="11ad3f86328e9313e44ea51baf678ee472a2e294b818bd8c58fa0a6763a057e4" Dec 06 07:26:39 crc kubenswrapper[4809]: I1206 07:26:39.429336 4809 scope.go:117] "RemoveContainer" containerID="0b246055e4db31addf5338d300fd1da85bef62ebc4d9b4302a50c949e69825db" Dec 06 07:26:39 crc kubenswrapper[4809]: I1206 07:26:39.482431 4809 scope.go:117] "RemoveContainer" containerID="4ad240cc3800ccf336f10cd13b9b80ee925ffc625ab64507e49bb53adcccd5de" Dec 06 07:27:17 crc kubenswrapper[4809]: I1206 07:27:17.880820 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-galera-0" podUID="7c596e14-b96f-4e5f-a8ac-3dc66fa0276d" containerName="galera" probeResult="failure" output="command timed out" Dec 06 07:27:17 crc kubenswrapper[4809]: I1206 07:27:17.881703 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="7c596e14-b96f-4e5f-a8ac-3dc66fa0276d" containerName="galera" probeResult="failure" output="command timed out" Dec 06 07:27:19 crc kubenswrapper[4809]: I1206 07:27:19.621228 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/metallb-operator-webhook-server-765b7d7f99-8dngt" podUID="60798ea0-75be-4a0c-a8c6-9fb431ba0e67" containerName="webhook-server" probeResult="failure" output="Get \"http://10.217.0.93:7472/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:27:20 crc kubenswrapper[4809]: I1206 07:27:19.621967 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/metallb-operator-webhook-server-765b7d7f99-8dngt" podUID="60798ea0-75be-4a0c-a8c6-9fb431ba0e67" containerName="webhook-server" probeResult="failure" output="Get \"http://10.217.0.93:7472/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:27:20 crc kubenswrapper[4809]: I1206 07:27:19.778662 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-nmstate/nmstate-handler-jbb6s" podUID="57d09bf3-cccc-4745-834d-57eaad3f96a1" containerName="nmstate-handler" probeResult="failure" output="command timed out" Dec 06 07:27:20 crc kubenswrapper[4809]: I1206 07:27:20.367247 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.751653046s: [/var/lib/containers/storage/overlay/28357be912c6013aebf0697c03755fada0223caac7d057c0bd1a4b14a5b59afa/diff /var/log/pods/openstack_swift-storage-0_cfd13e5d-ca28-4061-8ca5-2c4566332bf0/container-updater/0.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:27:20 crc kubenswrapper[4809]: I1206 07:27:20.367322 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.75175907s: [/var/lib/containers/storage/overlay/8f84e3cf45f5f634e11e640ceacbb98b72d90722a0a95302846217f8c10044e3/diff /var/log/pods/openshift-kube-scheduler_openshift-kube-scheduler-crc_3dcd261975c3d6b9a6ad6367fd4facd3/kube-scheduler-recovery-controller/0.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:27:20 crc kubenswrapper[4809]: I1206 07:27:20.378223 4809 trace.go:236] Trace[129149416]: "Calculate volume metrics of v4-0-config-system-service-ca for pod openshift-authentication/oauth-openshift-55889b984c-nxw8t" (06-Dec-2025 07:27:18.469) (total time: 1904ms): Dec 06 07:27:20 crc kubenswrapper[4809]: Trace[129149416]: [1.904719209s] [1.904719209s] END Dec 06 07:27:20 crc kubenswrapper[4809]: I1206 07:27:20.385735 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.1375452s: [/var/lib/containers/storage/overlay/ea73ce0bc845f1549fc43dbe7b1959f52f9abe9ba3e4475abbbe669341bfa69c/diff /var/log/pods/openstack_placement-5fc7d5fdc6-llxdt_b8ba7416-80be-4115-b960-58d147b56e9d/placement-api/0.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:27:20 crc kubenswrapper[4809]: I1206 07:27:20.387065 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.140068708s: [/var/lib/containers/storage/overlay/4ee4cfee0766828b424604c435bdcc9c21ec79cda8cc43f59c3949613a9c9f65/diff /var/log/pods/openstack_openstack-galera-0_7c596e14-b96f-4e5f-a8ac-3dc66fa0276d/galera/0.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:27:26 crc kubenswrapper[4809]: I1206 07:27:26.097968 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tdjqn"] Dec 06 07:27:26 crc kubenswrapper[4809]: E1206 07:27:26.104298 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1435a636-4c7d-44fc-ae60-c1ebc049d0c3" containerName="extract-utilities" Dec 06 07:27:26 crc kubenswrapper[4809]: I1206 07:27:26.104331 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1435a636-4c7d-44fc-ae60-c1ebc049d0c3" containerName="extract-utilities" Dec 06 07:27:26 crc kubenswrapper[4809]: E1206 07:27:26.104573 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fde5c8b4-ad5d-48ae-bc7f-3464d809dc92" containerName="extract-content" Dec 06 07:27:26 crc kubenswrapper[4809]: I1206 07:27:26.104583 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="fde5c8b4-ad5d-48ae-bc7f-3464d809dc92" containerName="extract-content" Dec 06 07:27:26 crc kubenswrapper[4809]: E1206 07:27:26.104598 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fde5c8b4-ad5d-48ae-bc7f-3464d809dc92" containerName="extract-utilities" Dec 06 07:27:26 crc kubenswrapper[4809]: I1206 07:27:26.104620 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="fde5c8b4-ad5d-48ae-bc7f-3464d809dc92" containerName="extract-utilities" Dec 06 07:27:26 crc kubenswrapper[4809]: E1206 07:27:26.104649 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fde5c8b4-ad5d-48ae-bc7f-3464d809dc92" containerName="registry-server" Dec 06 07:27:26 crc kubenswrapper[4809]: I1206 07:27:26.104655 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="fde5c8b4-ad5d-48ae-bc7f-3464d809dc92" containerName="registry-server" Dec 06 07:27:26 crc kubenswrapper[4809]: E1206 07:27:26.104676 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f5f3c13-919e-490d-9796-80d82cd12a53" containerName="extract-utilities" Dec 06 07:27:26 crc kubenswrapper[4809]: I1206 07:27:26.104685 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f5f3c13-919e-490d-9796-80d82cd12a53" containerName="extract-utilities" Dec 06 07:27:26 crc kubenswrapper[4809]: E1206 07:27:26.104700 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f5f3c13-919e-490d-9796-80d82cd12a53" containerName="registry-server" Dec 06 07:27:26 crc kubenswrapper[4809]: I1206 07:27:26.104707 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f5f3c13-919e-490d-9796-80d82cd12a53" containerName="registry-server" Dec 06 07:27:26 crc kubenswrapper[4809]: E1206 07:27:26.104715 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1435a636-4c7d-44fc-ae60-c1ebc049d0c3" containerName="registry-server" Dec 06 07:27:26 crc kubenswrapper[4809]: I1206 07:27:26.104723 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1435a636-4c7d-44fc-ae60-c1ebc049d0c3" containerName="registry-server" Dec 06 07:27:26 crc kubenswrapper[4809]: E1206 07:27:26.104738 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f5f3c13-919e-490d-9796-80d82cd12a53" containerName="extract-content" Dec 06 07:27:26 crc kubenswrapper[4809]: I1206 07:27:26.104745 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f5f3c13-919e-490d-9796-80d82cd12a53" containerName="extract-content" Dec 06 07:27:26 crc kubenswrapper[4809]: E1206 07:27:26.104761 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1435a636-4c7d-44fc-ae60-c1ebc049d0c3" containerName="extract-content" Dec 06 07:27:26 crc kubenswrapper[4809]: I1206 07:27:26.104767 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1435a636-4c7d-44fc-ae60-c1ebc049d0c3" containerName="extract-content" Dec 06 07:27:26 crc kubenswrapper[4809]: I1206 07:27:26.105196 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="fde5c8b4-ad5d-48ae-bc7f-3464d809dc92" containerName="registry-server" Dec 06 07:27:26 crc kubenswrapper[4809]: I1206 07:27:26.105215 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f5f3c13-919e-490d-9796-80d82cd12a53" containerName="registry-server" Dec 06 07:27:26 crc kubenswrapper[4809]: I1206 07:27:26.105255 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="1435a636-4c7d-44fc-ae60-c1ebc049d0c3" containerName="registry-server" Dec 06 07:27:26 crc kubenswrapper[4809]: I1206 07:27:26.108894 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tdjqn" Dec 06 07:27:26 crc kubenswrapper[4809]: I1206 07:27:26.160592 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1afc2f7d-0519-4486-a0f5-4c110fc28e53-utilities\") pod \"certified-operators-tdjqn\" (UID: \"1afc2f7d-0519-4486-a0f5-4c110fc28e53\") " pod="openshift-marketplace/certified-operators-tdjqn" Dec 06 07:27:26 crc kubenswrapper[4809]: I1206 07:27:26.160997 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1afc2f7d-0519-4486-a0f5-4c110fc28e53-catalog-content\") pod \"certified-operators-tdjqn\" (UID: \"1afc2f7d-0519-4486-a0f5-4c110fc28e53\") " pod="openshift-marketplace/certified-operators-tdjqn" Dec 06 07:27:26 crc kubenswrapper[4809]: I1206 07:27:26.161106 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhr7d\" (UniqueName: \"kubernetes.io/projected/1afc2f7d-0519-4486-a0f5-4c110fc28e53-kube-api-access-vhr7d\") pod \"certified-operators-tdjqn\" (UID: \"1afc2f7d-0519-4486-a0f5-4c110fc28e53\") " pod="openshift-marketplace/certified-operators-tdjqn" Dec 06 07:27:26 crc kubenswrapper[4809]: I1206 07:27:26.222841 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tdjqn"] Dec 06 07:27:26 crc kubenswrapper[4809]: I1206 07:27:26.263419 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1afc2f7d-0519-4486-a0f5-4c110fc28e53-utilities\") pod \"certified-operators-tdjqn\" (UID: \"1afc2f7d-0519-4486-a0f5-4c110fc28e53\") " pod="openshift-marketplace/certified-operators-tdjqn" Dec 06 07:27:26 crc kubenswrapper[4809]: I1206 07:27:26.264232 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1afc2f7d-0519-4486-a0f5-4c110fc28e53-catalog-content\") pod \"certified-operators-tdjqn\" (UID: \"1afc2f7d-0519-4486-a0f5-4c110fc28e53\") " pod="openshift-marketplace/certified-operators-tdjqn" Dec 06 07:27:26 crc kubenswrapper[4809]: I1206 07:27:26.264480 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhr7d\" (UniqueName: \"kubernetes.io/projected/1afc2f7d-0519-4486-a0f5-4c110fc28e53-kube-api-access-vhr7d\") pod \"certified-operators-tdjqn\" (UID: \"1afc2f7d-0519-4486-a0f5-4c110fc28e53\") " pod="openshift-marketplace/certified-operators-tdjqn" Dec 06 07:27:26 crc kubenswrapper[4809]: I1206 07:27:26.265560 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1afc2f7d-0519-4486-a0f5-4c110fc28e53-utilities\") pod \"certified-operators-tdjqn\" (UID: \"1afc2f7d-0519-4486-a0f5-4c110fc28e53\") " pod="openshift-marketplace/certified-operators-tdjqn" Dec 06 07:27:26 crc kubenswrapper[4809]: I1206 07:27:26.265767 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1afc2f7d-0519-4486-a0f5-4c110fc28e53-catalog-content\") pod \"certified-operators-tdjqn\" (UID: \"1afc2f7d-0519-4486-a0f5-4c110fc28e53\") " pod="openshift-marketplace/certified-operators-tdjqn" Dec 06 07:27:26 crc kubenswrapper[4809]: I1206 07:27:26.347905 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhr7d\" (UniqueName: \"kubernetes.io/projected/1afc2f7d-0519-4486-a0f5-4c110fc28e53-kube-api-access-vhr7d\") pod \"certified-operators-tdjqn\" (UID: \"1afc2f7d-0519-4486-a0f5-4c110fc28e53\") " pod="openshift-marketplace/certified-operators-tdjqn" Dec 06 07:27:26 crc kubenswrapper[4809]: I1206 07:27:26.446918 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tdjqn" Dec 06 07:27:28 crc kubenswrapper[4809]: I1206 07:27:28.778371 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-cell1-galera-0" podUID="e7056938-6e36-4631-adfb-01f116eba651" containerName="galera" probeResult="failure" output="command timed out" Dec 06 07:27:28 crc kubenswrapper[4809]: I1206 07:27:28.869015 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tdjqn"] Dec 06 07:27:29 crc kubenswrapper[4809]: I1206 07:27:29.373645 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tdjqn" event={"ID":"1afc2f7d-0519-4486-a0f5-4c110fc28e53","Type":"ContainerStarted","Data":"819736c810c39c649ee239ca383764071e0194a255c3c9f3fbe0a1d26fec6522"} Dec 06 07:27:31 crc kubenswrapper[4809]: I1206 07:27:31.402469 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tdjqn" event={"ID":"1afc2f7d-0519-4486-a0f5-4c110fc28e53","Type":"ContainerStarted","Data":"32acaa37e8a0496a4030ad09915728ae06ccd8d0aaf33ffbe9430b54f659749e"} Dec 06 07:27:33 crc kubenswrapper[4809]: I1206 07:27:33.204853 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tdjqn" event={"ID":"1afc2f7d-0519-4486-a0f5-4c110fc28e53","Type":"ContainerDied","Data":"32acaa37e8a0496a4030ad09915728ae06ccd8d0aaf33ffbe9430b54f659749e"} Dec 06 07:27:33 crc kubenswrapper[4809]: I1206 07:27:33.207861 4809 trace.go:236] Trace[1517624457]: "Calculate volume metrics of configmap-kubelet-serving-ca-bundle for pod openshift-monitoring/metrics-server-556bbc5579-8wblw" (06-Dec-2025 07:27:32.187) (total time: 1020ms): Dec 06 07:27:33 crc kubenswrapper[4809]: Trace[1517624457]: [1.02048561s] [1.02048561s] END Dec 06 07:27:33 crc kubenswrapper[4809]: I1206 07:27:33.204571 4809 generic.go:334] "Generic (PLEG): container finished" podID="1afc2f7d-0519-4486-a0f5-4c110fc28e53" containerID="32acaa37e8a0496a4030ad09915728ae06ccd8d0aaf33ffbe9430b54f659749e" exitCode=0 Dec 06 07:27:33 crc kubenswrapper[4809]: I1206 07:27:33.212299 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.220081815s: [/var/lib/containers/storage/overlay/506098dc73f55b3f15fea7da1d75bea39a47b32f6e64cdbe0fad4f46007bea0a/diff /var/log/pods/openstack_nova-metadata-0_e0909590-844d-41ec-9443-5caacd682230/nova-metadata-metadata/0.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:27:34 crc kubenswrapper[4809]: I1206 07:27:34.223395 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 07:27:39 crc kubenswrapper[4809]: I1206 07:27:39.276441 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tdjqn" event={"ID":"1afc2f7d-0519-4486-a0f5-4c110fc28e53","Type":"ContainerStarted","Data":"6160c4cd03560d1ac5e4a708d1eadbb37c1656760d0991de463899178cdac4e8"} Dec 06 07:27:41 crc kubenswrapper[4809]: I1206 07:27:41.522940 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/infra-operator-controller-manager-57548d458d-4b9l8" podUID="29c6d2cb-27fc-4b79-969a-462be41663e3" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.106:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:27:41 crc kubenswrapper[4809]: I1206 07:27:41.522752 4809 patch_prober.go:28] interesting pod/prometheus-operator-admission-webhook-f54c54754-ljfgz container/prometheus-operator-admission-webhook namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.69:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:27:41 crc kubenswrapper[4809]: I1206 07:27:41.523518 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-ljfgz" podUID="5e0d0108-b36f-47bc-bb5c-9fbc5ed0bf92" containerName="prometheus-operator-admission-webhook" probeResult="failure" output="Get \"https://10.217.0.69:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:27:41 crc kubenswrapper[4809]: I1206 07:27:41.526635 4809 trace.go:236] Trace[1736154162]: "Calculate volume metrics of catalog-content for pod openshift-marketplace/redhat-marketplace-zjvtm" (06-Dec-2025 07:27:40.506) (total time: 1020ms): Dec 06 07:27:41 crc kubenswrapper[4809]: Trace[1736154162]: [1.020398988s] [1.020398988s] END Dec 06 07:27:41 crc kubenswrapper[4809]: I1206 07:27:41.603011 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd462lbf" podUID="6fbac224-e82d-482e-aeea-c83a9f978775" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.113:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:27:41 crc kubenswrapper[4809]: I1206 07:27:41.996242 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/speaker-5t9z6" podUID="21411d89-4a48-4556-8ad2-193f1c8be0db" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:27:41 crc kubenswrapper[4809]: I1206 07:27:41.996261 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/speaker-5t9z6" podUID="21411d89-4a48-4556-8ad2-193f1c8be0db" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:27:44 crc kubenswrapper[4809]: I1206 07:27:44.899129 4809 patch_prober.go:28] interesting pod/observability-operator-d8bb48f5d-bfc8s container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.22:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:27:44 crc kubenswrapper[4809]: I1206 07:27:44.900001 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-d8bb48f5d-bfc8s" podUID="6d4f55df-5dc2-4e72-b7f8-cee9229da476" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.22:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:27:44 crc kubenswrapper[4809]: I1206 07:27:44.899247 4809 patch_prober.go:28] interesting pod/observability-operator-d8bb48f5d-bfc8s container/operator namespace/openshift-operators: Liveness probe status=failure output="Get \"http://10.217.0.22:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:27:44 crc kubenswrapper[4809]: I1206 07:27:44.900067 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operators/observability-operator-d8bb48f5d-bfc8s" podUID="6d4f55df-5dc2-4e72-b7f8-cee9229da476" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.22:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:27:45 crc kubenswrapper[4809]: I1206 07:27:45.145600 4809 patch_prober.go:28] interesting pod/perses-operator-5446b9c989-fh2sn container/perses-operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.33:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:27:45 crc kubenswrapper[4809]: I1206 07:27:45.145721 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/perses-operator-5446b9c989-fh2sn" podUID="bb9f4b83-1aa5-48ec-ac64-2d4c7835c5af" containerName="perses-operator" probeResult="failure" output="Get \"http://10.217.0.33:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:27:45 crc kubenswrapper[4809]: I1206 07:27:45.145883 4809 patch_prober.go:28] interesting pod/perses-operator-5446b9c989-fh2sn container/perses-operator namespace/openshift-operators: Liveness probe status=failure output="Get \"http://10.217.0.33:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:27:45 crc kubenswrapper[4809]: I1206 07:27:45.145969 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operators/perses-operator-5446b9c989-fh2sn" podUID="bb9f4b83-1aa5-48ec-ac64-2d4c7835c5af" containerName="perses-operator" probeResult="failure" output="Get \"http://10.217.0.33:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:27:45 crc kubenswrapper[4809]: I1206 07:27:45.189529 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-k8s-0" podUID="869feda2-ec9d-44de-9533-27a4d7b990c1" containerName="prometheus" probeResult="failure" output="command timed out" Dec 06 07:27:45 crc kubenswrapper[4809]: I1206 07:27:45.189595 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/prometheus-k8s-0" podUID="869feda2-ec9d-44de-9533-27a4d7b990c1" containerName="prometheus" probeResult="failure" output="command timed out" Dec 06 07:27:45 crc kubenswrapper[4809]: I1206 07:27:45.190851 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-index-hv5l4" podUID="45b1df34-ffd2-42ca-85ec-885efb8ec7e5" containerName="registry-server" probeResult="failure" output=< Dec 06 07:27:45 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Dec 06 07:27:45 crc kubenswrapper[4809]: > Dec 06 07:27:45 crc kubenswrapper[4809]: I1206 07:27:45.190906 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-operator-index-hv5l4" podUID="45b1df34-ffd2-42ca-85ec-885efb8ec7e5" containerName="registry-server" probeResult="failure" output=< Dec 06 07:27:45 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Dec 06 07:27:45 crc kubenswrapper[4809]: > Dec 06 07:27:45 crc kubenswrapper[4809]: I1206 07:27:45.273111 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-l8clc" podUID="44c524db-76cf-49c3-8c6e-1ad181d553ae" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.99:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:27:45 crc kubenswrapper[4809]: I1206 07:27:45.278101 4809 patch_prober.go:28] interesting pod/nmstate-webhook-5f6d4c5ccb-pw6wx container/nmstate-webhook namespace/openshift-nmstate: Readiness probe status=failure output="Get \"https://10.217.0.87:9443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:27:45 crc kubenswrapper[4809]: I1206 07:27:45.278131 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pw6wx" podUID="50b406f3-bc91-492f-89ea-8eeae9ae132e" containerName="nmstate-webhook" probeResult="failure" output="Get \"https://10.217.0.87:9443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:27:45 crc kubenswrapper[4809]: I1206 07:27:45.338135 4809 patch_prober.go:28] interesting pod/authentication-operator-69f744f599-7qpjs container/authentication-operator namespace/openshift-authentication-operator: Liveness probe status=failure output="Get \"https://10.217.0.18:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:27:45 crc kubenswrapper[4809]: I1206 07:27:45.338200 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication-operator/authentication-operator-69f744f599-7qpjs" podUID="2a1ed741-489c-455f-a344-404bbab66c22" containerName="authentication-operator" probeResult="failure" output="Get \"https://10.217.0.18:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:27:45 crc kubenswrapper[4809]: I1206 07:27:45.419001 4809 patch_prober.go:28] interesting pod/openshift-kube-scheduler-crc container/kube-scheduler namespace/openshift-kube-scheduler: Readiness probe status=failure output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:27:45 crc kubenswrapper[4809]: I1206 07:27:45.419053 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" probeResult="failure" output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:27:45 crc kubenswrapper[4809]: I1206 07:27:45.423110 4809 patch_prober.go:28] interesting pod/metrics-server-556bbc5579-8wblw container/metrics-server namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.75:10250/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:27:45 crc kubenswrapper[4809]: I1206 07:27:45.423178 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/metrics-server-556bbc5579-8wblw" podUID="b4f96141-d70e-4715-9e30-516c9e783cc9" containerName="metrics-server" probeResult="failure" output="Get \"https://10.217.0.75:10250/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:27:45 crc kubenswrapper[4809]: I1206 07:27:45.423575 4809 patch_prober.go:28] interesting pod/metrics-server-556bbc5579-8wblw container/metrics-server namespace/openshift-monitoring: Liveness probe status=failure output="Get \"https://10.217.0.75:10250/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:27:45 crc kubenswrapper[4809]: I1206 07:27:45.423607 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/metrics-server-556bbc5579-8wblw" podUID="b4f96141-d70e-4715-9e30-516c9e783cc9" containerName="metrics-server" probeResult="failure" output="Get \"https://10.217.0.75:10250/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:27:45 crc kubenswrapper[4809]: I1206 07:27:45.431886 4809 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-pkq7h container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.23:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:27:45 crc kubenswrapper[4809]: I1206 07:27:45.432259 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pkq7h" podUID="00bf9f05-7bc3-40e1-a2e9-1af1bf93f014" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.23:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:27:45 crc kubenswrapper[4809]: I1206 07:27:45.435329 4809 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-pkq7h container/catalog-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.23:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:27:45 crc kubenswrapper[4809]: I1206 07:27:45.435402 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pkq7h" podUID="00bf9f05-7bc3-40e1-a2e9-1af1bf93f014" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.23:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:27:45 crc kubenswrapper[4809]: I1206 07:27:45.558119 4809 patch_prober.go:28] interesting pod/downloads-7954f5f757-224mv container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.37:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:27:45 crc kubenswrapper[4809]: I1206 07:27:45.558167 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-224mv" podUID="8e290142-b2fc-492e-8763-ff41224579fc" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.37:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:27:45 crc kubenswrapper[4809]: I1206 07:27:45.558393 4809 patch_prober.go:28] interesting pod/downloads-7954f5f757-224mv container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:27:45 crc kubenswrapper[4809]: I1206 07:27:45.558449 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-224mv" podUID="8e290142-b2fc-492e-8763-ff41224579fc" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.37:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:27:45 crc kubenswrapper[4809]: I1206 07:27:45.708831 4809 patch_prober.go:28] interesting pod/router-default-5444994796-n6z57 container/router namespace/openshift-ingress: Readiness probe status=failure output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:27:45 crc kubenswrapper[4809]: I1206 07:27:45.708928 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5444994796-n6z57" podUID="e5356f1d-5afd-47bd-89af-72abfa8f901f" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:27:45 crc kubenswrapper[4809]: I1206 07:27:45.709512 4809 patch_prober.go:28] interesting pod/router-default-5444994796-n6z57 container/router namespace/openshift-ingress: Liveness probe status=failure output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:27:45 crc kubenswrapper[4809]: I1206 07:27:45.709552 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-ingress/router-default-5444994796-n6z57" podUID="e5356f1d-5afd-47bd-89af-72abfa8f901f" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:27:45 crc kubenswrapper[4809]: I1206 07:27:45.709593 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7trbt" podUID="45f87079-e852-4b47-a240-6f2bcf247509" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.105:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:27:45 crc kubenswrapper[4809]: I1206 07:27:45.709630 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7trbt" podUID="45f87079-e852-4b47-a240-6f2bcf247509" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.105:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:27:45 crc kubenswrapper[4809]: I1206 07:27:45.778774 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="1360c733-da74-4d37-8842-75050a356334" containerName="ovn-northd" probeResult="failure" output="command timed out" Dec 06 07:27:45 crc kubenswrapper[4809]: I1206 07:27:45.780017 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ovn-northd-0" podUID="1360c733-da74-4d37-8842-75050a356334" containerName="ovn-northd" probeResult="failure" output="command timed out" Dec 06 07:27:45 crc kubenswrapper[4809]: I1206 07:27:45.826659 4809 patch_prober.go:28] interesting pod/monitoring-plugin-5998c84788-zjqnq container/monitoring-plugin namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.77:9443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:27:45 crc kubenswrapper[4809]: I1206 07:27:45.826766 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/monitoring-plugin-5998c84788-zjqnq" podUID="bb66c253-f1a0-4412-93f7-82b8fb03fb33" containerName="monitoring-plugin" probeResult="failure" output="Get \"https://10.217.0.77:9443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:27:46 crc kubenswrapper[4809]: I1206 07:27:46.056140 4809 patch_prober.go:28] interesting pod/package-server-manager-789f6589d5-ch964 container/package-server-manager namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"http://10.217.0.17:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:27:46 crc kubenswrapper[4809]: I1206 07:27:46.056213 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ch964" podUID="b6b25f1e-33d4-481d-8c73-24b95c61364f" containerName="package-server-manager" probeResult="failure" output="Get \"http://10.217.0.17:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:27:46 crc kubenswrapper[4809]: I1206 07:27:46.056422 4809 patch_prober.go:28] interesting pod/package-server-manager-789f6589d5-ch964 container/package-server-manager namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"http://10.217.0.17:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:27:46 crc kubenswrapper[4809]: I1206 07:27:46.056452 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ch964" podUID="b6b25f1e-33d4-481d-8c73-24b95c61364f" containerName="package-server-manager" probeResult="failure" output="Get \"http://10.217.0.17:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:27:47 crc kubenswrapper[4809]: I1206 07:27:46.438155 4809 patch_prober.go:28] interesting pod/openshift-kube-scheduler-crc container/kube-scheduler namespace/openshift-kube-scheduler: Liveness probe status=failure output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:27:47 crc kubenswrapper[4809]: I1206 07:27:46.438212 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" probeResult="failure" output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:27:47 crc kubenswrapper[4809]: I1206 07:27:46.910177 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/redhat-marketplace-zjvtm" podUID="0714d252-8cd5-4a99-9050-8e383ad64885" containerName="registry-server" probeResult="failure" output=< Dec 06 07:27:47 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Dec 06 07:27:47 crc kubenswrapper[4809]: > Dec 06 07:27:47 crc kubenswrapper[4809]: I1206 07:27:46.910177 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-zjvtm" podUID="0714d252-8cd5-4a99-9050-8e383ad64885" containerName="registry-server" probeResult="failure" output=< Dec 06 07:27:47 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Dec 06 07:27:47 crc kubenswrapper[4809]: > Dec 06 07:27:47 crc kubenswrapper[4809]: I1206 07:27:46.978488 4809 patch_prober.go:28] interesting pod/thanos-querier-d74dd8556-59sjd container/kube-rbac-proxy-web namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.74:9091/-/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:27:47 crc kubenswrapper[4809]: I1206 07:27:46.978587 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/thanos-querier-d74dd8556-59sjd" podUID="ec930236-fceb-44dd-90c3-13a25d7cd9f2" containerName="kube-rbac-proxy-web" probeResult="failure" output="Get \"https://10.217.0.74:9091/-/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:27:48 crc kubenswrapper[4809]: I1206 07:27:48.777992 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="e7056938-6e36-4631-adfb-01f116eba651" containerName="galera" probeResult="failure" output="command timed out" Dec 06 07:27:48 crc kubenswrapper[4809]: I1206 07:27:48.778273 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-cell1-galera-0" podUID="e7056938-6e36-4631-adfb-01f116eba651" containerName="galera" probeResult="failure" output="command timed out" Dec 06 07:27:48 crc kubenswrapper[4809]: I1206 07:27:48.779018 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="7c596e14-b96f-4e5f-a8ac-3dc66fa0276d" containerName="galera" probeResult="failure" output="command timed out" Dec 06 07:27:48 crc kubenswrapper[4809]: I1206 07:27:48.779032 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-galera-0" podUID="7c596e14-b96f-4e5f-a8ac-3dc66fa0276d" containerName="galera" probeResult="failure" output="command timed out" Dec 06 07:27:50 crc kubenswrapper[4809]: I1206 07:27:50.785964 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tdjqn" event={"ID":"1afc2f7d-0519-4486-a0f5-4c110fc28e53","Type":"ContainerDied","Data":"6160c4cd03560d1ac5e4a708d1eadbb37c1656760d0991de463899178cdac4e8"} Dec 06 07:27:50 crc kubenswrapper[4809]: I1206 07:27:50.785968 4809 generic.go:334] "Generic (PLEG): container finished" podID="1afc2f7d-0519-4486-a0f5-4c110fc28e53" containerID="6160c4cd03560d1ac5e4a708d1eadbb37c1656760d0991de463899178cdac4e8" exitCode=0 Dec 06 07:27:55 crc kubenswrapper[4809]: I1206 07:27:55.256223 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.272662065s: [/var/lib/containers/storage/overlay/a94bf85210c2e9b6bf289bfd780dd376a0964e8980d92b77b7a970ff037d2526/diff /var/log/pods/openstack_swift-proxy-58bb4f66d7-bdk65_d496af20-e3cf-4ce6-86f5-df841a9ba42c/proxy-server/0.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:27:55 crc kubenswrapper[4809]: I1206 07:27:55.251480 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.499410262s: [/var/lib/containers/storage/overlay/7dea7ace731e3d6dc3461e82faa7cba24577cbad99b87a8df9ed7e0c1604f68f/diff /var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-vh8d4_93397035-a1a7-4c92-bdeb-5d0091dea181/manager/0.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:27:55 crc kubenswrapper[4809]: I1206 07:27:55.262156 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.510069331s: [/var/lib/containers/storage/overlay/bee7688b6a39ee376132301bc740ddd887a64e3f5eb8734aac27dffc7e68e968/diff /var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-4s24n_35b44606-4fdd-4662-9550-adc5ed2478b3/manager/0.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:27:55 crc kubenswrapper[4809]: I1206 07:27:55.267195 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.258365619s: [/var/lib/containers/storage/overlay/ba8e2f991443a5217aa670ba3dbc7e7236b781535f60ea3194783220094f2b15/diff /var/log/pods/openstack-operators_openstack-operator-controller-manager-5c4b598d95-mbx56_4dd349b6-b792-451e-9cab-7c52a5768a3d/manager/0.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:27:55 crc kubenswrapper[4809]: I1206 07:27:55.268197 4809 trace.go:236] Trace[2006121595]: "Calculate volume metrics of manager-config for pod openshift-operators-redhat/loki-operator-controller-manager-697657f76c-5stmg" (06-Dec-2025 07:27:53.924) (total time: 1324ms): Dec 06 07:27:55 crc kubenswrapper[4809]: Trace[2006121595]: [1.32466586s] [1.32466586s] END Dec 06 07:27:55 crc kubenswrapper[4809]: I1206 07:27:55.269725 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.491794307s: [/var/lib/containers/storage/overlay/7e7d37894da4a9a045174333f3ff89a23a0bbe02f235e8185e932449b0212886/diff /var/log/pods/openstack_swift-proxy-58bb4f66d7-bdk65_d496af20-e3cf-4ce6-86f5-df841a9ba42c/proxy-httpd/0.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:27:55 crc kubenswrapper[4809]: I1206 07:27:55.270214 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.509884257s: [/var/lib/containers/storage/overlay/d2a1f1c10baddced5bb330b4e0d261a0fff2c3fa386b446ff616e8e624e3b090/diff /var/log/pods/openstack_heat-api-7849fbcbcb-msd4w_dccd4eeb-4ba8-4200-8366-28d4ead2e476/heat-api/0.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:27:55 crc kubenswrapper[4809]: I1206 07:27:55.272856 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-scheduler-0" podUID="47aec31f-059f-4f3c-baef-4f9406e8cd4c" containerName="cinder-scheduler" probeResult="failure" output="Get \"http://10.217.0.206:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:27:55 crc kubenswrapper[4809]: I1206 07:27:55.276634 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.508107968s: [/var/lib/containers/storage/overlay/4828fb2dbb120e5bc814761acb6f53fff6e26bdd0bc7f09a5c6d2b1dab44f9ad/diff /var/log/pods/openstack_heat-cfnapi-794dfc48dc-pkjpx_b09a69de-0d5f-44f7-a3c9-2e6902da17b8/heat-cfnapi/0.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:27:55 crc kubenswrapper[4809]: I1206 07:27:55.444309 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-h69gw" podUID="8be493fb-5378-467d-9fde-d5154cda3207" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.101:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:27:58 crc kubenswrapper[4809]: I1206 07:27:58.411444 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ovn-northd-0" podUID="1360c733-da74-4d37-8842-75050a356334" containerName="ovn-northd" probeResult="failure" output="command timed out" Dec 06 07:27:58 crc kubenswrapper[4809]: I1206 07:27:58.412441 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="1360c733-da74-4d37-8842-75050a356334" containerName="ovn-northd" probeResult="failure" output="command timed out" Dec 06 07:27:58 crc kubenswrapper[4809]: I1206 07:27:58.412771 4809 patch_prober.go:28] interesting pod/perses-operator-5446b9c989-fh2sn container/perses-operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.33:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:27:58 crc kubenswrapper[4809]: I1206 07:27:58.412836 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/perses-operator-5446b9c989-fh2sn" podUID="bb9f4b83-1aa5-48ec-ac64-2d4c7835c5af" containerName="perses-operator" probeResult="failure" output="Get \"http://10.217.0.33:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:27:58 crc kubenswrapper[4809]: I1206 07:27:58.413333 4809 patch_prober.go:28] interesting pod/observability-operator-d8bb48f5d-bfc8s container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.22:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:27:58 crc kubenswrapper[4809]: I1206 07:27:58.413368 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-d8bb48f5d-bfc8s" podUID="6d4f55df-5dc2-4e72-b7f8-cee9229da476" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.22:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:27:58 crc kubenswrapper[4809]: I1206 07:27:58.413410 4809 patch_prober.go:28] interesting pod/metrics-server-556bbc5579-8wblw container/metrics-server namespace/openshift-monitoring: Liveness probe status=failure output="Get \"https://10.217.0.75:10250/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:27:58 crc kubenswrapper[4809]: I1206 07:27:58.413428 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/metrics-server-556bbc5579-8wblw" podUID="b4f96141-d70e-4715-9e30-516c9e783cc9" containerName="metrics-server" probeResult="failure" output="Get \"https://10.217.0.75:10250/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:27:58 crc kubenswrapper[4809]: I1206 07:27:58.429770 4809 patch_prober.go:28] interesting pod/authentication-operator-69f744f599-7qpjs container/authentication-operator namespace/openshift-authentication-operator: Liveness probe status=failure output="Get \"https://10.217.0.18:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:27:58 crc kubenswrapper[4809]: I1206 07:27:58.429844 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication-operator/authentication-operator-69f744f599-7qpjs" podUID="2a1ed741-489c-455f-a344-404bbab66c22" containerName="authentication-operator" probeResult="failure" output="Get \"https://10.217.0.18:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:27:58 crc kubenswrapper[4809]: I1206 07:27:58.429924 4809 patch_prober.go:28] interesting pod/observability-operator-d8bb48f5d-bfc8s container/operator namespace/openshift-operators: Liveness probe status=failure output="Get \"http://10.217.0.22:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:27:58 crc kubenswrapper[4809]: I1206 07:27:58.429959 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operators/observability-operator-d8bb48f5d-bfc8s" podUID="6d4f55df-5dc2-4e72-b7f8-cee9229da476" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.22:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:27:58 crc kubenswrapper[4809]: I1206 07:27:58.429981 4809 patch_prober.go:28] interesting pod/openshift-kube-scheduler-crc container/kube-scheduler namespace/openshift-kube-scheduler: Readiness probe status=failure output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:27:58 crc kubenswrapper[4809]: I1206 07:27:58.429998 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" probeResult="failure" output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:27:58 crc kubenswrapper[4809]: I1206 07:27:58.431084 4809 patch_prober.go:28] interesting pod/nmstate-webhook-5f6d4c5ccb-pw6wx container/nmstate-webhook namespace/openshift-nmstate: Readiness probe status=failure output="Get \"https://10.217.0.87:9443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:27:58 crc kubenswrapper[4809]: I1206 07:27:58.431109 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pw6wx" podUID="50b406f3-bc91-492f-89ea-8eeae9ae132e" containerName="nmstate-webhook" probeResult="failure" output="Get \"https://10.217.0.87:9443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:27:58 crc kubenswrapper[4809]: I1206 07:27:58.437853 4809 patch_prober.go:28] interesting pod/monitoring-plugin-5998c84788-zjqnq container/monitoring-plugin namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.77:9443/health\": context deadline exceeded" start-of-body= Dec 06 07:27:58 crc kubenswrapper[4809]: I1206 07:27:58.439335 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/monitoring-plugin-5998c84788-zjqnq" podUID="bb66c253-f1a0-4412-93f7-82b8fb03fb33" containerName="monitoring-plugin" probeResult="failure" output="Get \"https://10.217.0.77:9443/health\": context deadline exceeded" Dec 06 07:27:58 crc kubenswrapper[4809]: I1206 07:27:58.440891 4809 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-pkq7h container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.23:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:27:58 crc kubenswrapper[4809]: I1206 07:27:58.441031 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pkq7h" podUID="00bf9f05-7bc3-40e1-a2e9-1af1bf93f014" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.23:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:27:58 crc kubenswrapper[4809]: I1206 07:27:58.488494 4809 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-pkq7h container/catalog-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.23:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:27:58 crc kubenswrapper[4809]: I1206 07:27:58.488567 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pkq7h" podUID="00bf9f05-7bc3-40e1-a2e9-1af1bf93f014" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.23:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:27:58 crc kubenswrapper[4809]: I1206 07:27:58.490356 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7trbt" podUID="45f87079-e852-4b47-a240-6f2bcf247509" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.105:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:27:58 crc kubenswrapper[4809]: I1206 07:27:58.490462 4809 patch_prober.go:28] interesting pod/console-operator-58897d9998-hlvjb container/console-operator namespace/openshift-console-operator: Liveness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:27:58 crc kubenswrapper[4809]: I1206 07:27:58.490492 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console-operator/console-operator-58897d9998-hlvjb" podUID="f3a30985-f268-44de-93b9-bce0f9aebc6d" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:27:58 crc kubenswrapper[4809]: I1206 07:27:58.490599 4809 patch_prober.go:28] interesting pod/console-operator-58897d9998-hlvjb container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:27:58 crc kubenswrapper[4809]: I1206 07:27:58.490624 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-hlvjb" podUID="f3a30985-f268-44de-93b9-bce0f9aebc6d" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:27:58 crc kubenswrapper[4809]: E1206 07:27:58.511839 4809 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="3.124s" Dec 06 07:27:58 crc kubenswrapper[4809]: I1206 07:27:58.704202 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/placement-operator-controller-manager-78f8948974-zz2mm" podUID="b55361fa-8a7c-46ff-92f6-9dd72dbc0536" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.116:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:27:58 crc kubenswrapper[4809]: I1206 07:27:58.783390 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:58.787120 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/octavia-operator-controller-manager-998648c74-dfs6c" podUID="2ac245f3-8f64-4b6f-a51f-09e4b5811412" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.112:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:58.910317 4809 patch_prober.go:28] interesting pod/loki-operator-controller-manager-697657f76c-5stmg container/manager namespace/openshift-operators-redhat: Readiness probe status=failure output="Get \"http://10.217.0.47:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:58.910395 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators-redhat/loki-operator-controller-manager-697657f76c-5stmg" podUID="19fc8500-fddb-4650-839a-82166e98a567" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.47:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.116688 4809 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-jjf9r container/packageserver namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.34:5443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.116744 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jjf9r" podUID="e5e4e451-e3ee-434e-8e24-0c795925a48e" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.34:5443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.157229 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-bhbcl" podUID="fd7dd0a5-5e39-4812-b93f-a27c39284d33" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.119:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.157278 4809 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-gt9kb container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.19:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.157278 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-gzv6k" podUID="ced7591a-f80d-46a0-b8ea-024c2e37ae9f" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.107:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.157340 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gt9kb" podUID="c8d79e6f-40b0-4ecc-8501-454ba5cfe691" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.19:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.157397 4809 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-gt9kb container/olm-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.19:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.157413 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gt9kb" podUID="c8d79e6f-40b0-4ecc-8501-454ba5cfe691" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.19:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.157443 4809 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-jjf9r container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:5443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.157456 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jjf9r" podUID="e5e4e451-e3ee-434e-8e24-0c795925a48e" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.34:5443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.159979 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="cert-manager/cert-manager-webhook-5655c58dd6-7vvw8" podUID="886a998b-a62e-412c-ae50-fd1ba3a4a098" containerName="cert-manager-webhook" probeResult="failure" output="Get \"http://10.217.0.44:6080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.160138 4809 patch_prober.go:28] interesting pod/downloads-7954f5f757-224mv container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.37:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.160177 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-224mv" podUID="8e290142-b2fc-492e-8763-ff41224579fc" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.37:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.160209 4809 patch_prober.go:28] interesting pod/downloads-7954f5f757-224mv container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.160227 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/telemetry-operator-controller-manager-b4644b845-zxwvr" podUID="24aed9fd-d792-4bce-82a8-3ea4d70094b4" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.117:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.160239 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-224mv" podUID="8e290142-b2fc-492e-8763-ff41224579fc" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.37:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.160287 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/test-operator-controller-manager-5854674fcc-lqkvl" podUID="b376f569-0116-41f4-8a57-b5e5769e6a10" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.118:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.160355 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-27fdn" podUID="e6e0283a-55e9-4613-b74d-81fbb840690e" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.111:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.160396 4809 patch_prober.go:28] interesting pod/image-registry-66df7c8f76-dgspm container/registry namespace/openshift-image-registry: Liveness probe status=failure output="Get \"https://10.217.0.62:5000/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.160869 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-image-registry/image-registry-66df7c8f76-dgspm" podUID="26c71fe2-c0e4-405d-a6c6-2ce933013cba" containerName="registry" probeResult="failure" output="Get \"https://10.217.0.62:5000/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.160423 4809 patch_prober.go:28] interesting pod/image-registry-66df7c8f76-dgspm container/registry namespace/openshift-image-registry: Readiness probe status=failure output="Get \"https://10.217.0.62:5000/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.160481 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-sxcr7" podUID="c37c165f-5cd1-496f-b092-41efdfcd7eca" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.110:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.160946 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-image-registry/image-registry-66df7c8f76-dgspm" podUID="26c71fe2-c0e4-405d-a6c6-2ce933013cba" containerName="registry" probeResult="failure" output="Get \"https://10.217.0.62:5000/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.160509 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-vqvvr" podUID="803d54f1-fbac-4ec5-b486-efae1981b46d" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.115:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.160441 4809 patch_prober.go:28] interesting pod/package-server-manager-789f6589d5-ch964 container/package-server-manager namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"http://10.217.0.17:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.160999 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ch964" podUID="b6b25f1e-33d4-481d-8c73-24b95c61364f" containerName="package-server-manager" probeResult="failure" output="Get \"http://10.217.0.17:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.160453 4809 patch_prober.go:28] interesting pod/router-default-5444994796-n6z57 container/router namespace/openshift-ingress: Readiness probe status=failure output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.161034 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5444994796-n6z57" podUID="e5356f1d-5afd-47bd-89af-72abfa8f901f" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.160467 4809 patch_prober.go:28] interesting pod/router-default-5444994796-n6z57 container/router namespace/openshift-ingress: Liveness probe status=failure output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.161066 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-ingress/router-default-5444994796-n6z57" podUID="e5356f1d-5afd-47bd-89af-72abfa8f901f" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.160491 4809 patch_prober.go:28] interesting pod/package-server-manager-789f6589d5-ch964 container/package-server-manager namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"http://10.217.0.17:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.161098 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ch964" podUID="b6b25f1e-33d4-481d-8c73-24b95c61364f" containerName="package-server-manager" probeResult="failure" output="Get \"http://10.217.0.17:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.512812 4809 patch_prober.go:28] interesting pod/logging-loki-gateway-65498c4f8f-9s8qq container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.54:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.512891 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq" podUID="ef8f3e7e-57ea-42b8-a777-1778e5ed975b" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.54:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.512998 4809 patch_prober.go:28] interesting pod/controller-manager-7f967c4b75-bhj49 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.61:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.513043 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-7f967c4b75-bhj49" podUID="8fa68b06-fd0e-4a09-97ee-35d77cae7b80" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.61:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.513098 4809 patch_prober.go:28] interesting pod/controller-manager-7f967c4b75-bhj49 container/controller-manager namespace/openshift-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.61:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.513114 4809 patch_prober.go:28] interesting pod/openshift-kube-scheduler-crc container/kube-scheduler namespace/openshift-kube-scheduler: Liveness probe status=failure output="Get \"https://192.168.126.11:10259/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.513237 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-controller-manager/controller-manager-7f967c4b75-bhj49" podUID="8fa68b06-fd0e-4a09-97ee-35d77cae7b80" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.61:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.513331 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" probeResult="failure" output="Get \"https://192.168.126.11:10259/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.513367 4809 patch_prober.go:28] interesting pod/logging-loki-gateway-65498c4f8f-48rjq container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.53:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.513438 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-65498c4f8f-48rjq" podUID="72a3ed95-3e3d-4faf-88e4-ad1731902910" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.53:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.514428 4809 patch_prober.go:28] interesting pod/route-controller-manager-6cd65596c4-9z8ww container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.60:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.514468 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6cd65596c4-9z8ww" podUID="8e47b3dc-260e-46a6-9ce2-2e4f4a8a5977" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.60:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.514645 4809 patch_prober.go:28] interesting pod/route-controller-manager-6cd65596c4-9z8ww container/route-controller-manager namespace/openshift-route-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.60:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.514695 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-route-controller-manager/route-controller-manager-6cd65596c4-9z8ww" podUID="8e47b3dc-260e-46a6-9ce2-2e4f4a8a5977" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.60:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.580183 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/metallb-operator-webhook-server-765b7d7f99-8dngt" podUID="60798ea0-75be-4a0c-a8c6-9fb431ba0e67" containerName="webhook-server" probeResult="failure" output="Get \"http://10.217.0.93:7472/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:27:59.778968 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-cell1-galera-0" podUID="e7056938-6e36-4631-adfb-01f116eba651" containerName="galera" probeResult="failure" output="command timed out" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:28:00.315214 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-scheduler-0" podUID="47aec31f-059f-4f3c-baef-4f9406e8cd4c" containerName="cinder-scheduler" probeResult="failure" output="Get \"http://10.217.0.206:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:28:00.397233 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="e0909590-844d-41ec-9443-5caacd682230" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.0:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:28:00.438234 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="e0909590-844d-41ec-9443-5caacd682230" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.0:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:28:00.521150 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/controller-f8648f98b-d4fgg" podUID="073938c5-bc80-47d3-91e8-1d6d7da0f000" containerName="controller" probeResult="failure" output="Get \"http://10.217.0.95:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:28:00.521317 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/controller-f8648f98b-d4fgg" podUID="073938c5-bc80-47d3-91e8-1d6d7da0f000" containerName="controller" probeResult="failure" output="Get \"http://10.217.0.95:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:28:00.778390 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="1360c733-da74-4d37-8842-75050a356334" containerName="ovn-northd" probeResult="failure" output="command timed out" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:28:00.779697 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" podUID="9024f80a-36d2-48c4-ba94-14032531ea73" containerName="sbdb" probeResult="failure" output="command timed out" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:28:00.782230 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-zjvtm" podUID="0714d252-8cd5-4a99-9050-8e383ad64885" containerName="registry-server" probeResult="failure" output="command timed out" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:28:01.022195 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-z95cs" podUID="a3b6197b-05d4-49b2-9a67-56b675046a1d" containerName="controller" probeResult="failure" output="Get \"http://127.0.0.1:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:28:01.022270 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-z95cs" podUID="a3b6197b-05d4-49b2-9a67-56b675046a1d" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:28:01.022392 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/frr-k8s-z95cs" podUID="a3b6197b-05d4-49b2-9a67-56b675046a1d" containerName="controller" probeResult="failure" output="Get \"http://127.0.0.1:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:28:01.047709 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.593695191s: [/var/lib/containers/storage/overlay/ab9e93f92a86f9c5a881c4730b20cd750c5d190c41387f95549c5c1c0711e146/diff /var/log/pods/openshift-apiserver_apiserver-76f77b778f-tphf5_09a7af3d-6bf2-4b7a-93d9-cecc73d993a1/openshift-apiserver-check-endpoints/0.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:28:01 crc kubenswrapper[4809]: E1206 07:28:01.077731 4809 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.689s" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:28:01.095157 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="e7056938-6e36-4631-adfb-01f116eba651" containerName="galera" probeResult="failure" output="command timed out" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:28:01.095245 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-galera-0" podUID="7c596e14-b96f-4e5f-a8ac-3dc66fa0276d" containerName="galera" probeResult="failure" output="command timed out" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:28:01.095292 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="7c596e14-b96f-4e5f-a8ac-3dc66fa0276d" containerName="galera" probeResult="failure" output="command timed out" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:28:01.095330 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-nmstate/nmstate-handler-jbb6s" podUID="57d09bf3-cccc-4745-834d-57eaad3f96a1" containerName="nmstate-handler" probeResult="failure" output="command timed out" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:28:01.095444 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ovn-northd-0" podUID="1360c733-da74-4d37-8842-75050a356334" containerName="ovn-northd" probeResult="failure" output="command timed out" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:28:01.095483 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-vlv76" podUID="9024f80a-36d2-48c4-ba94-14032531ea73" containerName="nbdb" probeResult="failure" output="command timed out" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:28:01.095526 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/redhat-marketplace-zjvtm" podUID="0714d252-8cd5-4a99-9050-8e383ad64885" containerName="registry-server" probeResult="failure" output="command timed out" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:28:01.174525 4809 patch_prober.go:28] interesting pod/prometheus-operator-admission-webhook-f54c54754-ljfgz container/prometheus-operator-admission-webhook namespace/openshift-monitoring: Liveness probe status=failure output="Get \"https://10.217.0.69:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:28:01.174590 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-ljfgz" podUID="5e0d0108-b36f-47bc-bb5c-9fbc5ed0bf92" containerName="prometheus-operator-admission-webhook" probeResult="failure" output="Get \"https://10.217.0.69:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:28:01.175839 4809 patch_prober.go:28] interesting pod/prometheus-operator-admission-webhook-f54c54754-ljfgz container/prometheus-operator-admission-webhook namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.69:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:28:01.175900 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-ljfgz" podUID="5e0d0108-b36f-47bc-bb5c-9fbc5ed0bf92" containerName="prometheus-operator-admission-webhook" probeResult="failure" output="Get \"https://10.217.0.69:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:01 crc kubenswrapper[4809]: I1206 07:28:01.334693 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/infra-operator-controller-manager-57548d458d-4b9l8" podUID="29c6d2cb-27fc-4b79-969a-462be41663e3" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.106:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:03 crc kubenswrapper[4809]: I1206 07:28:01.556181 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="hostpath-provisioner/csi-hostpathplugin-p2lkt" podUID="c598f580-a438-4015-9a8a-f98b3b2b0b3d" containerName="hostpath-provisioner" probeResult="failure" output="Get \"http://10.217.0.40:9898/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:03 crc kubenswrapper[4809]: I1206 07:28:01.778755 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/prometheus-k8s-0" podUID="869feda2-ec9d-44de-9533-27a4d7b990c1" containerName="prometheus" probeResult="failure" output="command timed out" Dec 06 07:28:03 crc kubenswrapper[4809]: I1206 07:28:01.780733 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-k8s-0" podUID="869feda2-ec9d-44de-9533-27a4d7b990c1" containerName="prometheus" probeResult="failure" output="command timed out" Dec 06 07:28:03 crc kubenswrapper[4809]: I1206 07:28:01.955508 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/speaker-5t9z6" podUID="21411d89-4a48-4556-8ad2-193f1c8be0db" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:03 crc kubenswrapper[4809]: I1206 07:28:02.352229 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-controller-manager-5c4b598d95-mbx56" podUID="4dd349b6-b792-451e-9cab-7c52a5768a3d" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.120:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:03 crc kubenswrapper[4809]: I1206 07:28:02.352229 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-operator-controller-manager-5c4b598d95-mbx56" podUID="4dd349b6-b792-451e-9cab-7c52a5768a3d" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.120:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:03 crc kubenswrapper[4809]: I1206 07:28:02.779866 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-nmstate/nmstate-handler-jbb6s" podUID="57d09bf3-cccc-4745-834d-57eaad3f96a1" containerName="nmstate-handler" probeResult="failure" output="command timed out" Dec 06 07:28:03 crc kubenswrapper[4809]: I1206 07:28:02.872127 4809 patch_prober.go:28] interesting pod/logging-loki-query-frontend-84558f7c9f-9mw6j container/loki-query-frontend namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.52:3101/loki/api/v1/status/buildinfo\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:03 crc kubenswrapper[4809]: I1206 07:28:02.872290 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9mw6j" podUID="142a1d98-f469-4d1c-8541-d4ceb151f74e" containerName="loki-query-frontend" probeResult="failure" output="Get \"https://10.217.0.52:3101/loki/api/v1/status/buildinfo\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:03 crc kubenswrapper[4809]: I1206 07:28:03.247044 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.024374725s: [/var/lib/containers/storage/overlay/7abf4f10e79568aeea94cb2a5f7a3b6901e85534b72e27654dc1cd7c7c5a3111/diff /var/log/pods/openstack-operators_telemetry-operator-controller-manager-b4644b845-zxwvr_24aed9fd-d792-4bce-82a8-3ea4d70094b4/manager/0.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:28:03 crc kubenswrapper[4809]: E1206 07:28:03.277861 4809 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.89s" Dec 06 07:28:03 crc kubenswrapper[4809]: I1206 07:28:03.290405 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 2.001610007s: [/var/lib/containers/storage/overlay/cb9fbf785fa7fc39b3f073e742f76731d6c85932f7d38693def427fe32cf04fe/diff /var/log/pods/openstack_aodh-0_7e16d0f3-9bd2-4887-84dd-3b8fb35d6451/aodh-api/0.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:28:03 crc kubenswrapper[4809]: I1206 07:28:03.336396 4809 patch_prober.go:28] interesting pod/logging-loki-gateway-65498c4f8f-48rjq container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.53:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:05 crc kubenswrapper[4809]: I1206 07:28:03.336473 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-65498c4f8f-48rjq" podUID="72a3ed95-3e3d-4faf-88e4-ad1731902910" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.53:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:05 crc kubenswrapper[4809]: I1206 07:28:03.394947 4809 patch_prober.go:28] interesting pod/logging-loki-gateway-65498c4f8f-9s8qq container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.54:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:05 crc kubenswrapper[4809]: I1206 07:28:03.394992 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq" podUID="ef8f3e7e-57ea-42b8-a777-1778e5ed975b" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.54:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:05 crc kubenswrapper[4809]: I1206 07:28:03.513191 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-cell1-novncproxy-0" podUID="663c54d8-faa9-466a-9126-678a92bc92f9" containerName="nova-cell1-novncproxy-novncproxy" probeResult="failure" output="Get \"https://10.217.0.249:6080/vnc_lite.html\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:05 crc kubenswrapper[4809]: I1206 07:28:03.514620 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/kube-state-metrics-0" podUID="36f27192-451f-4ce2-afda-c0e66a6028b2" containerName="kube-state-metrics" probeResult="failure" output="Get \"https://10.217.1.3:8080/livez\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:05 crc kubenswrapper[4809]: I1206 07:28:03.788669 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ovn-controller-ovs-jgqjw" podUID="92ce9234-d708-449e-9874-97ac8a4b994c" containerName="ovsdb-server" probeResult="failure" output="command timed out" Dec 06 07:28:05 crc kubenswrapper[4809]: I1206 07:28:03.788829 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-ovs-jgqjw" podUID="92ce9234-d708-449e-9874-97ac8a4b994c" containerName="ovs-vswitchd" probeResult="failure" output="command timed out" Dec 06 07:28:05 crc kubenswrapper[4809]: I1206 07:28:03.788891 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-ovs-jgqjw" podUID="92ce9234-d708-449e-9874-97ac8a4b994c" containerName="ovsdb-server" probeResult="failure" output="command timed out" Dec 06 07:28:05 crc kubenswrapper[4809]: I1206 07:28:03.788956 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-xhlzm" podUID="6b64391f-74d7-4b69-a898-f7375057c6d5" containerName="ovn-controller" probeResult="failure" output="command timed out" Dec 06 07:28:05 crc kubenswrapper[4809]: I1206 07:28:03.789020 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ovn-controller-ovs-jgqjw" podUID="92ce9234-d708-449e-9874-97ac8a4b994c" containerName="ovs-vswitchd" probeResult="failure" output="command timed out" Dec 06 07:28:05 crc kubenswrapper[4809]: I1206 07:28:03.789090 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/certified-operators-nqpwx" podUID="408b531c-9f2a-4ee5-ad65-db67b2313be3" containerName="registry-server" probeResult="failure" output="command timed out" Dec 06 07:28:05 crc kubenswrapper[4809]: I1206 07:28:03.789116 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/community-operators-4p5rc" podUID="6fa58271-edf3-4d3d-8782-83a959a691fe" containerName="registry-server" probeResult="failure" output="command timed out" Dec 06 07:28:05 crc kubenswrapper[4809]: I1206 07:28:03.789147 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/certified-operators-nqpwx" podUID="408b531c-9f2a-4ee5-ad65-db67b2313be3" containerName="registry-server" probeResult="failure" output="command timed out" Dec 06 07:28:05 crc kubenswrapper[4809]: I1206 07:28:03.792884 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ovn-controller-xhlzm" podUID="6b64391f-74d7-4b69-a898-f7375057c6d5" containerName="ovn-controller" probeResult="failure" output="command timed out" Dec 06 07:28:05 crc kubenswrapper[4809]: I1206 07:28:03.795040 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/community-operators-4p5rc" podUID="6fa58271-edf3-4d3d-8782-83a959a691fe" containerName="registry-server" probeResult="failure" output="command timed out" Dec 06 07:28:05 crc kubenswrapper[4809]: I1206 07:28:03.872985 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="hostpath-provisioner/csi-hostpathplugin-p2lkt" podUID="c598f580-a438-4015-9a8a-f98b3b2b0b3d" containerName="hostpath-provisioner" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 07:28:05 crc kubenswrapper[4809]: I1206 07:28:04.785463 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/redhat-operators-czfkp" podUID="1c5ff94a-d9b0-41f2-9d47-47f67e0a9895" containerName="registry-server" probeResult="failure" output="command timed out" Dec 06 07:28:05 crc kubenswrapper[4809]: I1206 07:28:04.785453 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/redhat-operators-czfkp" podUID="1c5ff94a-d9b0-41f2-9d47-47f67e0a9895" containerName="registry-server" probeResult="failure" output="command timed out" Dec 06 07:28:05 crc kubenswrapper[4809]: I1206 07:28:05.105144 4809 patch_prober.go:28] interesting pod/perses-operator-5446b9c989-fh2sn container/perses-operator namespace/openshift-operators: Liveness probe status=failure output="Get \"http://10.217.0.33:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:05 crc kubenswrapper[4809]: I1206 07:28:05.105204 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operators/perses-operator-5446b9c989-fh2sn" podUID="bb9f4b83-1aa5-48ec-ac64-2d4c7835c5af" containerName="perses-operator" probeResult="failure" output="Get \"http://10.217.0.33:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:05 crc kubenswrapper[4809]: I1206 07:28:05.302089 4809 patch_prober.go:28] interesting pod/authentication-operator-69f744f599-7qpjs container/authentication-operator namespace/openshift-authentication-operator: Liveness probe status=failure output="Get \"https://10.217.0.18:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:05 crc kubenswrapper[4809]: I1206 07:28:05.302135 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication-operator/authentication-operator-69f744f599-7qpjs" podUID="2a1ed741-489c-455f-a344-404bbab66c22" containerName="authentication-operator" probeResult="failure" output="Get \"https://10.217.0.18:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:05 crc kubenswrapper[4809]: I1206 07:28:05.302175 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-authentication-operator/authentication-operator-69f744f599-7qpjs" Dec 06 07:28:05 crc kubenswrapper[4809]: I1206 07:28:05.303428 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="authentication-operator" containerStatusID={"Type":"cri-o","ID":"d6349ee6d3b2d24851d9fb4c381c6461d8aade431f1bc4d59880c6989f34e131"} pod="openshift-authentication-operator/authentication-operator-69f744f599-7qpjs" containerMessage="Container authentication-operator failed liveness probe, will be restarted" Dec 06 07:28:05 crc kubenswrapper[4809]: I1206 07:28:05.305675 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication-operator/authentication-operator-69f744f599-7qpjs" podUID="2a1ed741-489c-455f-a344-404bbab66c22" containerName="authentication-operator" containerID="cri-o://d6349ee6d3b2d24851d9fb4c381c6461d8aade431f1bc4d59880c6989f34e131" gracePeriod=30 Dec 06 07:28:05 crc kubenswrapper[4809]: I1206 07:28:05.433137 4809 patch_prober.go:28] interesting pod/metrics-server-556bbc5579-8wblw container/metrics-server namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.75:10250/livez\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:05 crc kubenswrapper[4809]: I1206 07:28:05.433353 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/metrics-server-556bbc5579-8wblw" podUID="b4f96141-d70e-4715-9e30-516c9e783cc9" containerName="metrics-server" probeResult="failure" output="Get \"https://10.217.0.75:10250/livez\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:05 crc kubenswrapper[4809]: I1206 07:28:05.433507 4809 patch_prober.go:28] interesting pod/metrics-server-556bbc5579-8wblw container/metrics-server namespace/openshift-monitoring: Liveness probe status=failure output="Get \"https://10.217.0.75:10250/livez\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:05 crc kubenswrapper[4809]: I1206 07:28:05.433536 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/metrics-server-556bbc5579-8wblw" podUID="b4f96141-d70e-4715-9e30-516c9e783cc9" containerName="metrics-server" probeResult="failure" output="Get \"https://10.217.0.75:10250/livez\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:05 crc kubenswrapper[4809]: I1206 07:28:05.433634 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-monitoring/metrics-server-556bbc5579-8wblw" Dec 06 07:28:05 crc kubenswrapper[4809]: I1206 07:28:05.435201 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="metrics-server" containerStatusID={"Type":"cri-o","ID":"cac60d4c52fd5a110675ebf867d8dabeb845a5056ba0753e52a1a1f592f0790d"} pod="openshift-monitoring/metrics-server-556bbc5579-8wblw" containerMessage="Container metrics-server failed liveness probe, will be restarted" Dec 06 07:28:05 crc kubenswrapper[4809]: I1206 07:28:05.435349 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-monitoring/metrics-server-556bbc5579-8wblw" podUID="b4f96141-d70e-4715-9e30-516c9e783cc9" containerName="metrics-server" containerID="cri-o://cac60d4c52fd5a110675ebf867d8dabeb845a5056ba0753e52a1a1f592f0790d" gracePeriod=170 Dec 06 07:28:05 crc kubenswrapper[4809]: I1206 07:28:05.486122 4809 patch_prober.go:28] interesting pod/console-operator-58897d9998-hlvjb container/console-operator namespace/openshift-console-operator: Liveness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:05 crc kubenswrapper[4809]: I1206 07:28:05.486189 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console-operator/console-operator-58897d9998-hlvjb" podUID="f3a30985-f268-44de-93b9-bce0f9aebc6d" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:05 crc kubenswrapper[4809]: I1206 07:28:05.486072 4809 patch_prober.go:28] interesting pod/console-operator-58897d9998-hlvjb container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:05 crc kubenswrapper[4809]: I1206 07:28:05.486508 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-hlvjb" podUID="f3a30985-f268-44de-93b9-bce0f9aebc6d" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:05 crc kubenswrapper[4809]: I1206 07:28:05.709407 4809 patch_prober.go:28] interesting pod/router-default-5444994796-n6z57 container/router namespace/openshift-ingress: Liveness probe status=failure output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:05 crc kubenswrapper[4809]: I1206 07:28:05.709727 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-ingress/router-default-5444994796-n6z57" podUID="e5356f1d-5afd-47bd-89af-72abfa8f901f" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:06 crc kubenswrapper[4809]: I1206 07:28:05.709773 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-ingress/router-default-5444994796-n6z57" Dec 06 07:28:06 crc kubenswrapper[4809]: I1206 07:28:05.710116 4809 patch_prober.go:28] interesting pod/router-default-5444994796-n6z57 container/router namespace/openshift-ingress: Readiness probe status=failure output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:06 crc kubenswrapper[4809]: I1206 07:28:05.710184 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5444994796-n6z57" podUID="e5356f1d-5afd-47bd-89af-72abfa8f901f" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:06 crc kubenswrapper[4809]: I1206 07:28:05.710293 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-n6z57" Dec 06 07:28:06 crc kubenswrapper[4809]: I1206 07:28:05.721634 4809 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-jjf9r container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:5443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:06 crc kubenswrapper[4809]: I1206 07:28:05.721692 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jjf9r" podUID="e5e4e451-e3ee-434e-8e24-0c795925a48e" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.34:5443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:06 crc kubenswrapper[4809]: I1206 07:28:05.721762 4809 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-jjf9r container/packageserver namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.34:5443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:06 crc kubenswrapper[4809]: I1206 07:28:05.721774 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jjf9r" podUID="e5e4e451-e3ee-434e-8e24-0c795925a48e" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.34:5443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:06 crc kubenswrapper[4809]: I1206 07:28:06.095302 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="router" containerStatusID={"Type":"cri-o","ID":"e061b12da080bb0dc2c8e03d029012d9f630feb21d1998eb7a5ed7f369824e29"} pod="openshift-ingress/router-default-5444994796-n6z57" containerMessage="Container router failed liveness probe, will be restarted" Dec 06 07:28:06 crc kubenswrapper[4809]: I1206 07:28:06.095410 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ingress/router-default-5444994796-n6z57" podUID="e5356f1d-5afd-47bd-89af-72abfa8f901f" containerName="router" containerID="cri-o://e061b12da080bb0dc2c8e03d029012d9f630feb21d1998eb7a5ed7f369824e29" gracePeriod=10 Dec 06 07:28:06 crc kubenswrapper[4809]: I1206 07:28:06.103629 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-ghwtq" podUID="b379307b-aafd-42da-b761-38a556a0d143" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.114:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:06 crc kubenswrapper[4809]: I1206 07:28:06.779015 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/prometheus-k8s-0" podUID="869feda2-ec9d-44de-9533-27a4d7b990c1" containerName="prometheus" probeResult="failure" output="command timed out" Dec 06 07:28:06 crc kubenswrapper[4809]: I1206 07:28:06.779121 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-k8s-0" podUID="869feda2-ec9d-44de-9533-27a4d7b990c1" containerName="prometheus" probeResult="failure" output="command timed out" Dec 06 07:28:06 crc kubenswrapper[4809]: I1206 07:28:06.782982 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Dec 06 07:28:06 crc kubenswrapper[4809]: I1206 07:28:06.977356 4809 patch_prober.go:28] interesting pod/thanos-querier-d74dd8556-59sjd container/kube-rbac-proxy-web namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.74:9091/-/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:06 crc kubenswrapper[4809]: I1206 07:28:06.977408 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/thanos-querier-d74dd8556-59sjd" podUID="ec930236-fceb-44dd-90c3-13a25d7cd9f2" containerName="kube-rbac-proxy-web" probeResult="failure" output="Get \"https://10.217.0.74:9091/-/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:07 crc kubenswrapper[4809]: I1206 07:28:07.779303 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-galera-0" podUID="7c596e14-b96f-4e5f-a8ac-3dc66fa0276d" containerName="galera" probeResult="failure" output="command timed out" Dec 06 07:28:07 crc kubenswrapper[4809]: I1206 07:28:07.779688 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="7c596e14-b96f-4e5f-a8ac-3dc66fa0276d" containerName="galera" probeResult="failure" output="command timed out" Dec 06 07:28:07 crc kubenswrapper[4809]: I1206 07:28:07.779698 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/openstack-galera-0" Dec 06 07:28:07 crc kubenswrapper[4809]: I1206 07:28:07.779916 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 06 07:28:07 crc kubenswrapper[4809]: I1206 07:28:07.781387 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="galera" containerStatusID={"Type":"cri-o","ID":"a33b2b043d42437efe270d1e3ec31c3d840ad9b765f7e171518bdd9f6ff5ad80"} pod="openstack/openstack-galera-0" containerMessage="Container galera failed liveness probe, will be restarted" Dec 06 07:28:08 crc kubenswrapper[4809]: I1206 07:28:08.238766 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-7qpjs" event={"ID":"2a1ed741-489c-455f-a344-404bbab66c22","Type":"ContainerDied","Data":"d6349ee6d3b2d24851d9fb4c381c6461d8aade431f1bc4d59880c6989f34e131"} Dec 06 07:28:08 crc kubenswrapper[4809]: I1206 07:28:08.242121 4809 generic.go:334] "Generic (PLEG): container finished" podID="2a1ed741-489c-455f-a344-404bbab66c22" containerID="d6349ee6d3b2d24851d9fb4c381c6461d8aade431f1bc4d59880c6989f34e131" exitCode=0 Dec 06 07:28:08 crc kubenswrapper[4809]: I1206 07:28:08.334560 4809 patch_prober.go:28] interesting pod/logging-loki-gateway-65498c4f8f-48rjq container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.53:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:08 crc kubenswrapper[4809]: I1206 07:28:08.334633 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-65498c4f8f-48rjq" podUID="72a3ed95-3e3d-4faf-88e4-ad1731902910" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.53:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:08 crc kubenswrapper[4809]: I1206 07:28:08.335876 4809 patch_prober.go:28] interesting pod/logging-loki-gateway-65498c4f8f-48rjq container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.53:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:08 crc kubenswrapper[4809]: I1206 07:28:08.335967 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-65498c4f8f-48rjq" podUID="72a3ed95-3e3d-4faf-88e4-ad1731902910" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.53:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:08 crc kubenswrapper[4809]: I1206 07:28:08.395100 4809 patch_prober.go:28] interesting pod/logging-loki-gateway-65498c4f8f-9s8qq container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.54:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:08 crc kubenswrapper[4809]: I1206 07:28:08.395162 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq" podUID="ef8f3e7e-57ea-42b8-a777-1778e5ed975b" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.54:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:08 crc kubenswrapper[4809]: I1206 07:28:08.396175 4809 patch_prober.go:28] interesting pod/logging-loki-gateway-65498c4f8f-9s8qq container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.54:8083/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:08 crc kubenswrapper[4809]: I1206 07:28:08.396279 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq" podUID="ef8f3e7e-57ea-42b8-a777-1778e5ed975b" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.54:8083/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:08 crc kubenswrapper[4809]: I1206 07:28:08.778775 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-cell1-galera-0" podUID="e7056938-6e36-4631-adfb-01f116eba651" containerName="galera" probeResult="failure" output="command timed out" Dec 06 07:28:08 crc kubenswrapper[4809]: I1206 07:28:08.779153 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="e7056938-6e36-4631-adfb-01f116eba651" containerName="galera" probeResult="failure" output="command timed out" Dec 06 07:28:08 crc kubenswrapper[4809]: I1206 07:28:08.778775 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="7c596e14-b96f-4e5f-a8ac-3dc66fa0276d" containerName="galera" probeResult="failure" output="command timed out" Dec 06 07:28:08 crc kubenswrapper[4809]: I1206 07:28:08.779203 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 06 07:28:08 crc kubenswrapper[4809]: I1206 07:28:08.779294 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 06 07:28:08 crc kubenswrapper[4809]: I1206 07:28:08.902038 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="galera" containerStatusID={"Type":"cri-o","ID":"c5e7932399fd9c59936d31395b31c6f5cc98b66aad28f9a1faffee5aa4d0028c"} pod="openstack/openstack-cell1-galera-0" containerMessage="Container galera failed liveness probe, will be restarted" Dec 06 07:28:09 crc kubenswrapper[4809]: I1206 07:28:09.289405 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tdjqn" event={"ID":"1afc2f7d-0519-4486-a0f5-4c110fc28e53","Type":"ContainerStarted","Data":"aa8a15cd520fe4652ff7fa9c351a8feac89f3ba113cfdb0ff068ddb24f3ead61"} Dec 06 07:28:09 crc kubenswrapper[4809]: I1206 07:28:09.298666 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-7qpjs" event={"ID":"2a1ed741-489c-455f-a344-404bbab66c22","Type":"ContainerStarted","Data":"4e1fbadae559e71985e3cd99ba5040ef13528186bda4ee77d6c171462f3ed57c"} Dec 06 07:28:09 crc kubenswrapper[4809]: I1206 07:28:09.372575 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tdjqn" podStartSLOduration=10.810684509 podStartE2EDuration="44.368843706s" podCreationTimestamp="2025-12-06 07:27:25 +0000 UTC" firstStartedPulling="2025-12-06 07:27:34.22187897 +0000 UTC m=+5779.110861912" lastFinishedPulling="2025-12-06 07:28:07.780038167 +0000 UTC m=+5812.669021109" observedRunningTime="2025-12-06 07:28:09.361922929 +0000 UTC m=+5814.250905871" watchObservedRunningTime="2025-12-06 07:28:09.368843706 +0000 UTC m=+5814.257826648" Dec 06 07:28:09 crc kubenswrapper[4809]: I1206 07:28:09.523185 4809 trace.go:236] Trace[49520776]: "Calculate volume metrics of registry-storage for pod openshift-image-registry/image-registry-66df7c8f76-dgspm" (06-Dec-2025 07:28:07.579) (total time: 1942ms): Dec 06 07:28:09 crc kubenswrapper[4809]: Trace[49520776]: [1.942795686s] [1.942795686s] END Dec 06 07:28:10 crc kubenswrapper[4809]: I1206 07:28:10.406731 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-ttn9g" podUID="b483382f-c2e3-40bf-ab4b-b58c2f3f9132" containerName="frr-k8s-webhook-server" probeResult="failure" output="Get \"http://10.217.0.94:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:10 crc kubenswrapper[4809]: I1206 07:28:10.409480 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-ttn9g" podUID="b483382f-c2e3-40bf-ab4b-b58c2f3f9132" containerName="frr-k8s-webhook-server" probeResult="failure" output="Get \"http://10.217.0.94:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:10 crc kubenswrapper[4809]: I1206 07:28:10.625603 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/kube-state-metrics-0" podUID="36f27192-451f-4ce2-afda-c0e66a6028b2" containerName="kube-state-metrics" probeResult="failure" output="Get \"https://10.217.1.3:8081/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:10 crc kubenswrapper[4809]: I1206 07:28:10.625655 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/kube-state-metrics-0" podUID="36f27192-451f-4ce2-afda-c0e66a6028b2" containerName="kube-state-metrics" probeResult="failure" output="Get \"https://10.217.1.3:8080/livez\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:10 crc kubenswrapper[4809]: I1206 07:28:10.781110 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="e7056938-6e36-4631-adfb-01f116eba651" containerName="galera" probeResult="failure" output="command timed out" Dec 06 07:28:11 crc kubenswrapper[4809]: I1206 07:28:11.021203 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-z95cs" podUID="a3b6197b-05d4-49b2-9a67-56b675046a1d" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:11 crc kubenswrapper[4809]: I1206 07:28:11.021559 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/frr-k8s-z95cs" podUID="a3b6197b-05d4-49b2-9a67-56b675046a1d" containerName="controller" probeResult="failure" output="Get \"http://127.0.0.1:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:11 crc kubenswrapper[4809]: I1206 07:28:11.021298 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-z95cs" podUID="a3b6197b-05d4-49b2-9a67-56b675046a1d" containerName="controller" probeResult="failure" output="Get \"http://127.0.0.1:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:11 crc kubenswrapper[4809]: I1206 07:28:11.174773 4809 patch_prober.go:28] interesting pod/prometheus-operator-admission-webhook-f54c54754-ljfgz container/prometheus-operator-admission-webhook namespace/openshift-monitoring: Liveness probe status=failure output="Get \"https://10.217.0.69:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:11 crc kubenswrapper[4809]: I1206 07:28:11.174842 4809 patch_prober.go:28] interesting pod/prometheus-operator-admission-webhook-f54c54754-ljfgz container/prometheus-operator-admission-webhook namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.69:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:11 crc kubenswrapper[4809]: I1206 07:28:11.174851 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-ljfgz" podUID="5e0d0108-b36f-47bc-bb5c-9fbc5ed0bf92" containerName="prometheus-operator-admission-webhook" probeResult="failure" output="Get \"https://10.217.0.69:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:11 crc kubenswrapper[4809]: I1206 07:28:11.174901 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-ljfgz" podUID="5e0d0108-b36f-47bc-bb5c-9fbc5ed0bf92" containerName="prometheus-operator-admission-webhook" probeResult="failure" output="Get \"https://10.217.0.69:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:11 crc kubenswrapper[4809]: I1206 07:28:11.686686 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd462lbf" podUID="6fbac224-e82d-482e-aeea-c83a9f978775" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.113:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:12 crc kubenswrapper[4809]: I1206 07:28:12.618889 4809 patch_prober.go:28] interesting pod/logging-loki-distributor-76cc67bf56-r5p25 container/loki-distributor namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.50:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:12 crc kubenswrapper[4809]: I1206 07:28:12.619339 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-distributor-76cc67bf56-r5p25" podUID="2ae7975b-9830-47bf-af4f-6f7fe5e47894" containerName="loki-distributor" probeResult="failure" output="Get \"https://10.217.0.50:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:12 crc kubenswrapper[4809]: I1206 07:28:12.738572 4809 patch_prober.go:28] interesting pod/logging-loki-querier-5895d59bb8-q5rsl container/loki-querier namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.51:3101/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:12 crc kubenswrapper[4809]: I1206 07:28:12.738645 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-querier-5895d59bb8-q5rsl" podUID="c4ce3880-869a-4fef-b76c-e669273700e1" containerName="loki-querier" probeResult="failure" output="Get \"https://10.217.0.51:3101/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:12 crc kubenswrapper[4809]: I1206 07:28:12.782762 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Dec 06 07:28:12 crc kubenswrapper[4809]: I1206 07:28:12.783207 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/ceilometer-0" Dec 06 07:28:12 crc kubenswrapper[4809]: I1206 07:28:12.784441 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="ceilometer-central-agent" containerStatusID={"Type":"cri-o","ID":"1ed2ac2be19d734f28ddb67018abd7ca0033f58827c7e4ab6c9c7cfe52be430c"} pod="openstack/ceilometer-0" containerMessage="Container ceilometer-central-agent failed liveness probe, will be restarted" Dec 06 07:28:12 crc kubenswrapper[4809]: I1206 07:28:12.784713 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerName="ceilometer-central-agent" containerID="cri-o://1ed2ac2be19d734f28ddb67018abd7ca0033f58827c7e4ab6c9c7cfe52be430c" gracePeriod=30 Dec 06 07:28:12 crc kubenswrapper[4809]: I1206 07:28:12.872445 4809 patch_prober.go:28] interesting pod/logging-loki-query-frontend-84558f7c9f-9mw6j container/loki-query-frontend namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.52:3101/loki/api/v1/status/buildinfo\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:12 crc kubenswrapper[4809]: I1206 07:28:12.872518 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9mw6j" podUID="142a1d98-f469-4d1c-8541-d4ceb151f74e" containerName="loki-query-frontend" probeResult="failure" output="Get \"https://10.217.0.52:3101/loki/api/v1/status/buildinfo\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:13 crc kubenswrapper[4809]: I1206 07:28:13.336134 4809 patch_prober.go:28] interesting pod/logging-loki-gateway-65498c4f8f-48rjq container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.53:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:13 crc kubenswrapper[4809]: I1206 07:28:13.336208 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-65498c4f8f-48rjq" podUID="72a3ed95-3e3d-4faf-88e4-ad1731902910" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.53:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:13 crc kubenswrapper[4809]: I1206 07:28:13.394706 4809 patch_prober.go:28] interesting pod/logging-loki-gateway-65498c4f8f-9s8qq container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.54:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:13 crc kubenswrapper[4809]: I1206 07:28:13.394761 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq" podUID="ef8f3e7e-57ea-42b8-a777-1778e5ed975b" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.54:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:13 crc kubenswrapper[4809]: I1206 07:28:13.788653 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerName="ceilometer-notification-agent" probeResult="failure" output="command timed out" Dec 06 07:28:14 crc kubenswrapper[4809]: I1206 07:28:14.898269 4809 patch_prober.go:28] interesting pod/observability-operator-d8bb48f5d-bfc8s container/operator namespace/openshift-operators: Liveness probe status=failure output="Get \"http://10.217.0.22:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:15 crc kubenswrapper[4809]: I1206 07:28:14.898657 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operators/observability-operator-d8bb48f5d-bfc8s" podUID="6d4f55df-5dc2-4e72-b7f8-cee9229da476" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.22:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:15 crc kubenswrapper[4809]: I1206 07:28:15.144396 4809 patch_prober.go:28] interesting pod/observability-operator-d8bb48f5d-bfc8s container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.22:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:15 crc kubenswrapper[4809]: I1206 07:28:15.144489 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-d8bb48f5d-bfc8s" podUID="6d4f55df-5dc2-4e72-b7f8-cee9229da476" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.22:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:15 crc kubenswrapper[4809]: I1206 07:28:15.318192 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-l8clc" podUID="44c524db-76cf-49c3-8c6e-1ad181d553ae" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.99:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:15 crc kubenswrapper[4809]: I1206 07:28:15.359184 4809 patch_prober.go:28] interesting pod/nmstate-webhook-5f6d4c5ccb-pw6wx container/nmstate-webhook namespace/openshift-nmstate: Readiness probe status=failure output="Get \"https://10.217.0.87:9443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:15 crc kubenswrapper[4809]: I1206 07:28:15.359301 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pw6wx" podUID="50b406f3-bc91-492f-89ea-8eeae9ae132e" containerName="nmstate-webhook" probeResult="failure" output="Get \"https://10.217.0.87:9443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:15 crc kubenswrapper[4809]: I1206 07:28:15.359385 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-h69gw" podUID="8be493fb-5378-467d-9fde-d5154cda3207" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.101:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:15 crc kubenswrapper[4809]: I1206 07:28:15.359427 4809 patch_prober.go:28] interesting pod/perses-operator-5446b9c989-fh2sn container/perses-operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.33:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:15 crc kubenswrapper[4809]: I1206 07:28:15.359447 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/perses-operator-5446b9c989-fh2sn" podUID="bb9f4b83-1aa5-48ec-ac64-2d4c7835c5af" containerName="perses-operator" probeResult="failure" output="Get \"http://10.217.0.33:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:15 crc kubenswrapper[4809]: I1206 07:28:15.413444 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-9gfg2" podUID="ee834860-9609-482d-95db-82fa1a1a6941" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.104:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:15 crc kubenswrapper[4809]: I1206 07:28:15.418491 4809 patch_prober.go:28] interesting pod/openshift-kube-scheduler-crc container/kube-scheduler namespace/openshift-kube-scheduler: Readiness probe status=failure output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:15 crc kubenswrapper[4809]: I1206 07:28:15.418573 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" probeResult="failure" output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:15 crc kubenswrapper[4809]: I1206 07:28:15.430660 4809 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-pkq7h container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.23:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:15 crc kubenswrapper[4809]: I1206 07:28:15.430714 4809 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-pkq7h container/catalog-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.23:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:15 crc kubenswrapper[4809]: I1206 07:28:15.430724 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pkq7h" podUID="00bf9f05-7bc3-40e1-a2e9-1af1bf93f014" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.23:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:15 crc kubenswrapper[4809]: I1206 07:28:15.430772 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pkq7h" podUID="00bf9f05-7bc3-40e1-a2e9-1af1bf93f014" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.23:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:15 crc kubenswrapper[4809]: I1206 07:28:15.527309 4809 patch_prober.go:28] interesting pod/console-operator-58897d9998-hlvjb container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:15 crc kubenswrapper[4809]: I1206 07:28:15.527389 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-hlvjb" podUID="f3a30985-f268-44de-93b9-bce0f9aebc6d" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:15 crc kubenswrapper[4809]: I1206 07:28:15.527801 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-hlvjb" Dec 06 07:28:15 crc kubenswrapper[4809]: I1206 07:28:15.609160 4809 patch_prober.go:28] interesting pod/console-operator-58897d9998-hlvjb container/console-operator namespace/openshift-console-operator: Liveness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:15 crc kubenswrapper[4809]: I1206 07:28:15.609180 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="cert-manager/cert-manager-webhook-5655c58dd6-7vvw8" podUID="886a998b-a62e-412c-ae50-fd1ba3a4a098" containerName="cert-manager-webhook" probeResult="failure" output="Get \"http://10.217.0.44:6080/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:15 crc kubenswrapper[4809]: I1206 07:28:15.609233 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console-operator/console-operator-58897d9998-hlvjb" podUID="f3a30985-f268-44de-93b9-bce0f9aebc6d" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:15 crc kubenswrapper[4809]: I1206 07:28:15.609294 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console-operator/console-operator-58897d9998-hlvjb" Dec 06 07:28:15 crc kubenswrapper[4809]: I1206 07:28:15.734188 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-j4c62" podUID="7233cfab-96cb-4459-a507-3031cac37720" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.100:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:15 crc kubenswrapper[4809]: I1206 07:28:15.734764 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="cert-manager/cert-manager-webhook-5655c58dd6-7vvw8" podUID="886a998b-a62e-412c-ae50-fd1ba3a4a098" containerName="cert-manager-webhook" probeResult="failure" output="Get \"http://10.217.0.44:6080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:15 crc kubenswrapper[4809]: I1206 07:28:15.736752 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="console-operator" containerStatusID={"Type":"cri-o","ID":"14009f9cf4e890d82d10c1569da8ae672118c24e2c0f3c5f4fb375548f3e2114"} pod="openshift-console-operator/console-operator-58897d9998-hlvjb" containerMessage="Container console-operator failed liveness probe, will be restarted" Dec 06 07:28:15 crc kubenswrapper[4809]: I1206 07:28:15.739486 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console-operator/console-operator-58897d9998-hlvjb" podUID="f3a30985-f268-44de-93b9-bce0f9aebc6d" containerName="console-operator" containerID="cri-o://14009f9cf4e890d82d10c1569da8ae672118c24e2c0f3c5f4fb375548f3e2114" gracePeriod=30 Dec 06 07:28:15 crc kubenswrapper[4809]: I1206 07:28:15.816166 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4s24n" podUID="35b44606-4fdd-4662-9550-adc5ed2478b3" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.103:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:15 crc kubenswrapper[4809]: I1206 07:28:15.857254 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7trbt" podUID="45f87079-e852-4b47-a240-6f2bcf247509" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.105:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:15 crc kubenswrapper[4809]: I1206 07:28:15.941131 4809 patch_prober.go:28] interesting pod/downloads-7954f5f757-224mv container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.37:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:15 crc kubenswrapper[4809]: I1206 07:28:15.941158 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-operator-controller-operator-6598f77bc9-xjfmg" podUID="2d9f04dc-3365-4b3e-b57c-604e657c6483" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.98:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:15 crc kubenswrapper[4809]: I1206 07:28:15.941208 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-224mv" podUID="8e290142-b2fc-492e-8763-ff41224579fc" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.37:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:15 crc kubenswrapper[4809]: I1206 07:28:15.941268 4809 patch_prober.go:28] interesting pod/downloads-7954f5f757-224mv container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:15 crc kubenswrapper[4809]: I1206 07:28:15.941291 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-224mv" podUID="8e290142-b2fc-492e-8763-ff41224579fc" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.37:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:15 crc kubenswrapper[4809]: I1206 07:28:15.982071 4809 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-jjf9r container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:15 crc kubenswrapper[4809]: I1206 07:28:15.982115 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-gzv6k" podUID="ced7591a-f80d-46a0-b8ea-024c2e37ae9f" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.107:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:15 crc kubenswrapper[4809]: I1206 07:28:15.982141 4809 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-jjf9r container/packageserver namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.34:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:15 crc kubenswrapper[4809]: I1206 07:28:15.982444 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jjf9r" podUID="e5e4e451-e3ee-434e-8e24-0c795925a48e" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.34:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:15 crc kubenswrapper[4809]: I1206 07:28:15.982396 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jjf9r" podUID="e5e4e451-e3ee-434e-8e24-0c795925a48e" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.34:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:15 crc kubenswrapper[4809]: I1206 07:28:15.982485 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jjf9r" Dec 06 07:28:15 crc kubenswrapper[4809]: I1206 07:28:15.982539 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jjf9r" Dec 06 07:28:15 crc kubenswrapper[4809]: I1206 07:28:15.983803 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="packageserver" containerStatusID={"Type":"cri-o","ID":"17e00e064844bbaa3aeb76dd9b61ea649de4b3f1cdb9d6b7b0238d8f5b30b9a1"} pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jjf9r" containerMessage="Container packageserver failed liveness probe, will be restarted" Dec 06 07:28:15 crc kubenswrapper[4809]: I1206 07:28:15.983851 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jjf9r" podUID="e5e4e451-e3ee-434e-8e24-0c795925a48e" containerName="packageserver" containerID="cri-o://17e00e064844bbaa3aeb76dd9b61ea649de4b3f1cdb9d6b7b0238d8f5b30b9a1" gracePeriod=30 Dec 06 07:28:16 crc kubenswrapper[4809]: I1206 07:28:16.023261 4809 patch_prober.go:28] interesting pod/router-default-5444994796-n6z57 container/router namespace/openshift-ingress: Readiness probe status=failure output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:16 crc kubenswrapper[4809]: I1206 07:28:16.023329 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5444994796-n6z57" podUID="e5356f1d-5afd-47bd-89af-72abfa8f901f" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:16 crc kubenswrapper[4809]: I1206 07:28:16.023330 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-vh8d4" podUID="93397035-a1a7-4c92-bdeb-5d0091dea181" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.108:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:16 crc kubenswrapper[4809]: I1206 07:28:16.072155 4809 patch_prober.go:28] interesting pod/monitoring-plugin-5998c84788-zjqnq container/monitoring-plugin namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.77:9443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:16 crc kubenswrapper[4809]: I1206 07:28:16.072225 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/monitoring-plugin-5998c84788-zjqnq" podUID="bb66c253-f1a0-4412-93f7-82b8fb03fb33" containerName="monitoring-plugin" probeResult="failure" output="Get \"https://10.217.0.77:9443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:16 crc kubenswrapper[4809]: I1206 07:28:16.072282 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-kcp8k" podUID="f736a8a4-ff7e-4179-951f-66405b1fd933" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.109:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:16 crc kubenswrapper[4809]: I1206 07:28:16.115180 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-controller-operator-6598f77bc9-xjfmg" podUID="2d9f04dc-3365-4b3e-b57c-604e657c6483" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.98:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:16 crc kubenswrapper[4809]: I1206 07:28:16.157289 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/octavia-operator-controller-manager-998648c74-dfs6c" podUID="2ac245f3-8f64-4b6f-a51f-09e4b5811412" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.112:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:16 crc kubenswrapper[4809]: I1206 07:28:16.157483 4809 patch_prober.go:28] interesting pod/console-operator-58897d9998-hlvjb container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 06 07:28:16 crc kubenswrapper[4809]: [+]log ok Dec 06 07:28:16 crc kubenswrapper[4809]: [+]poststarthook/max-in-flight-filter ok Dec 06 07:28:16 crc kubenswrapper[4809]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 06 07:28:16 crc kubenswrapper[4809]: [-]shutdown failed: reason withheld Dec 06 07:28:16 crc kubenswrapper[4809]: readyz check failed Dec 06 07:28:16 crc kubenswrapper[4809]: I1206 07:28:16.157536 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-hlvjb" podUID="f3a30985-f268-44de-93b9-bce0f9aebc6d" containerName="console-operator" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 07:28:16 crc kubenswrapper[4809]: I1206 07:28:16.240157 4809 patch_prober.go:28] interesting pod/package-server-manager-789f6589d5-ch964 container/package-server-manager namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"http://10.217.0.17:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:16 crc kubenswrapper[4809]: I1206 07:28:16.240240 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ch964" podUID="b6b25f1e-33d4-481d-8c73-24b95c61364f" containerName="package-server-manager" probeResult="failure" output="Get \"http://10.217.0.17:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:16 crc kubenswrapper[4809]: I1206 07:28:16.323123 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-ghwtq" podUID="b379307b-aafd-42da-b761-38a556a0d143" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.114:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:16 crc kubenswrapper[4809]: I1206 07:28:16.406116 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-27fdn" podUID="e6e0283a-55e9-4613-b74d-81fbb840690e" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.111:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:16 crc kubenswrapper[4809]: I1206 07:28:16.406300 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/telemetry-operator-controller-manager-b4644b845-zxwvr" podUID="24aed9fd-d792-4bce-82a8-3ea4d70094b4" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.117:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:16 crc kubenswrapper[4809]: I1206 07:28:16.406533 4809 patch_prober.go:28] interesting pod/console-operator-58897d9998-hlvjb container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/readyz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Dec 06 07:28:16 crc kubenswrapper[4809]: I1206 07:28:16.406562 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-hlvjb" podUID="f3a30985-f268-44de-93b9-bce0f9aebc6d" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/readyz\": dial tcp 10.217.0.24:8443: connect: connection refused" Dec 06 07:28:16 crc kubenswrapper[4809]: I1206 07:28:16.447268 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/test-operator-controller-manager-5854674fcc-lqkvl" podUID="b376f569-0116-41f4-8a57-b5e5769e6a10" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.118:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:16 crc kubenswrapper[4809]: I1206 07:28:16.447374 4809 patch_prober.go:28] interesting pod/package-server-manager-789f6589d5-ch964 container/package-server-manager namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"http://10.217.0.17:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:16 crc kubenswrapper[4809]: I1206 07:28:16.447398 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ch964" podUID="b6b25f1e-33d4-481d-8c73-24b95c61364f" containerName="package-server-manager" probeResult="failure" output="Get \"http://10.217.0.17:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:16 crc kubenswrapper[4809]: I1206 07:28:16.447399 4809 patch_prober.go:28] interesting pod/loki-operator-controller-manager-697657f76c-5stmg container/manager namespace/openshift-operators-redhat: Readiness probe status=failure output="Get \"http://10.217.0.47:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:16 crc kubenswrapper[4809]: I1206 07:28:16.447449 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators-redhat/loki-operator-controller-manager-697657f76c-5stmg" podUID="19fc8500-fddb-4650-839a-82166e98a567" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.47:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:16 crc kubenswrapper[4809]: I1206 07:28:16.447787 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tdjqn" Dec 06 07:28:16 crc kubenswrapper[4809]: I1206 07:28:16.447792 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-vqvvr" podUID="803d54f1-fbac-4ec5-b486-efae1981b46d" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.115:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:16 crc kubenswrapper[4809]: I1206 07:28:16.447844 4809 patch_prober.go:28] interesting pod/openshift-kube-scheduler-crc container/kube-scheduler namespace/openshift-kube-scheduler: Liveness probe status=failure output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:16 crc kubenswrapper[4809]: I1206 07:28:16.447843 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/placement-operator-controller-manager-78f8948974-zz2mm" podUID="b55361fa-8a7c-46ff-92f6-9dd72dbc0536" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.116:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:16 crc kubenswrapper[4809]: I1206 07:28:16.447870 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" probeResult="failure" output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:16 crc kubenswrapper[4809]: I1206 07:28:16.448050 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tdjqn" Dec 06 07:28:16 crc kubenswrapper[4809]: I1206 07:28:16.772038 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress_router-default-5444994796-n6z57_e5356f1d-5afd-47bd-89af-72abfa8f901f/router/0.log" Dec 06 07:28:16 crc kubenswrapper[4809]: I1206 07:28:16.773150 4809 generic.go:334] "Generic (PLEG): container finished" podID="e5356f1d-5afd-47bd-89af-72abfa8f901f" containerID="e061b12da080bb0dc2c8e03d029012d9f630feb21d1998eb7a5ed7f369824e29" exitCode=137 Dec 06 07:28:16 crc kubenswrapper[4809]: I1206 07:28:16.773222 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-n6z57" event={"ID":"e5356f1d-5afd-47bd-89af-72abfa8f901f","Type":"ContainerDied","Data":"e061b12da080bb0dc2c8e03d029012d9f630feb21d1998eb7a5ed7f369824e29"} Dec 06 07:28:16 crc kubenswrapper[4809]: I1206 07:28:16.781413 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-58897d9998-hlvjb_f3a30985-f268-44de-93b9-bce0f9aebc6d/console-operator/0.log" Dec 06 07:28:16 crc kubenswrapper[4809]: I1206 07:28:16.781473 4809 generic.go:334] "Generic (PLEG): container finished" podID="f3a30985-f268-44de-93b9-bce0f9aebc6d" containerID="14009f9cf4e890d82d10c1569da8ae672118c24e2c0f3c5f4fb375548f3e2114" exitCode=1 Dec 06 07:28:16 crc kubenswrapper[4809]: I1206 07:28:16.781575 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-hlvjb" event={"ID":"f3a30985-f268-44de-93b9-bce0f9aebc6d","Type":"ContainerDied","Data":"14009f9cf4e890d82d10c1569da8ae672118c24e2c0f3c5f4fb375548f3e2114"} Dec 06 07:28:16 crc kubenswrapper[4809]: I1206 07:28:16.982910 4809 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-jjf9r container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:5443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:16 crc kubenswrapper[4809]: I1206 07:28:16.982998 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jjf9r" podUID="e5e4e451-e3ee-434e-8e24-0c795925a48e" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.34:5443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:17 crc kubenswrapper[4809]: I1206 07:28:17.779193 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="7c596e14-b96f-4e5f-a8ac-3dc66fa0276d" containerName="galera" probeResult="failure" output="command timed out" Dec 06 07:28:19 crc kubenswrapper[4809]: I1206 07:28:17.956149 4809 patch_prober.go:28] interesting pod/oauth-openshift-55889b984c-nxw8t container/oauth-openshift namespace/openshift-authentication: Liveness probe status=failure output="Get \"https://10.217.0.56:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:19 crc kubenswrapper[4809]: I1206 07:28:17.956763 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" podUID="404214e0-30aa-44ff-b2d0-fb3c127068f9" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.56:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:19 crc kubenswrapper[4809]: I1206 07:28:17.956168 4809 patch_prober.go:28] interesting pod/oauth-openshift-55889b984c-nxw8t container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.56:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:19 crc kubenswrapper[4809]: I1206 07:28:17.956859 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" podUID="404214e0-30aa-44ff-b2d0-fb3c127068f9" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.56:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:19 crc kubenswrapper[4809]: I1206 07:28:18.335030 4809 patch_prober.go:28] interesting pod/logging-loki-gateway-65498c4f8f-48rjq container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.53:8083/ready\": context deadline exceeded" start-of-body= Dec 06 07:28:19 crc kubenswrapper[4809]: I1206 07:28:18.335096 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-65498c4f8f-48rjq" podUID="72a3ed95-3e3d-4faf-88e4-ad1731902910" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.53:8083/ready\": context deadline exceeded" Dec 06 07:28:19 crc kubenswrapper[4809]: I1206 07:28:18.335451 4809 patch_prober.go:28] interesting pod/logging-loki-gateway-65498c4f8f-48rjq container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.53:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:19 crc kubenswrapper[4809]: I1206 07:28:18.335497 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-65498c4f8f-48rjq" podUID="72a3ed95-3e3d-4faf-88e4-ad1731902910" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.53:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:19 crc kubenswrapper[4809]: I1206 07:28:18.394359 4809 patch_prober.go:28] interesting pod/logging-loki-gateway-65498c4f8f-9s8qq container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.54:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:19 crc kubenswrapper[4809]: I1206 07:28:18.394406 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq" podUID="ef8f3e7e-57ea-42b8-a777-1778e5ed975b" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.54:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:19 crc kubenswrapper[4809]: I1206 07:28:18.396606 4809 patch_prober.go:28] interesting pod/logging-loki-gateway-65498c4f8f-9s8qq container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.54:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:19 crc kubenswrapper[4809]: I1206 07:28:18.396630 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq" podUID="ef8f3e7e-57ea-42b8-a777-1778e5ed975b" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.54:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:19 crc kubenswrapper[4809]: I1206 07:28:18.403219 4809 patch_prober.go:28] interesting pod/route-controller-manager-6cd65596c4-9z8ww container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.60:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:19 crc kubenswrapper[4809]: I1206 07:28:18.403281 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6cd65596c4-9z8ww" podUID="8e47b3dc-260e-46a6-9ce2-2e4f4a8a5977" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.60:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:19 crc kubenswrapper[4809]: I1206 07:28:18.403237 4809 patch_prober.go:28] interesting pod/route-controller-manager-6cd65596c4-9z8ww container/route-controller-manager namespace/openshift-route-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.60:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:19 crc kubenswrapper[4809]: I1206 07:28:18.403401 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-route-controller-manager/route-controller-manager-6cd65596c4-9z8ww" podUID="8e47b3dc-260e-46a6-9ce2-2e4f4a8a5977" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.60:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:19 crc kubenswrapper[4809]: I1206 07:28:18.408754 4809 patch_prober.go:28] interesting pod/controller-manager-7f967c4b75-bhj49 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.61:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:19 crc kubenswrapper[4809]: I1206 07:28:18.408826 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-7f967c4b75-bhj49" podUID="8fa68b06-fd0e-4a09-97ee-35d77cae7b80" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.61:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:19 crc kubenswrapper[4809]: I1206 07:28:18.409960 4809 patch_prober.go:28] interesting pod/controller-manager-7f967c4b75-bhj49 container/controller-manager namespace/openshift-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.61:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:19 crc kubenswrapper[4809]: I1206 07:28:18.410044 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-controller-manager/controller-manager-7f967c4b75-bhj49" podUID="8fa68b06-fd0e-4a09-97ee-35d77cae7b80" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.61:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:19 crc kubenswrapper[4809]: I1206 07:28:18.725102 4809 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-qqv6x container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.30:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:19 crc kubenswrapper[4809]: I1206 07:28:18.725165 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qqv6x" podUID="8a225b5c-4ad5-4f08-a0cb-ae38a8fa83d4" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.30:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:19 crc kubenswrapper[4809]: I1206 07:28:18.725102 4809 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-qqv6x container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.30:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:19 crc kubenswrapper[4809]: I1206 07:28:18.725235 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qqv6x" podUID="8a225b5c-4ad5-4f08-a0cb-ae38a8fa83d4" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.30:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:19 crc kubenswrapper[4809]: I1206 07:28:18.778223 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-k8s-0" podUID="869feda2-ec9d-44de-9533-27a4d7b990c1" containerName="prometheus" probeResult="failure" output="command timed out" Dec 06 07:28:19 crc kubenswrapper[4809]: I1206 07:28:18.779600 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/prometheus-k8s-0" podUID="869feda2-ec9d-44de-9533-27a4d7b990c1" containerName="prometheus" probeResult="failure" output="command timed out" Dec 06 07:28:19 crc kubenswrapper[4809]: I1206 07:28:18.779597 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="e7056938-6e36-4631-adfb-01f116eba651" containerName="galera" probeResult="failure" output="command timed out" Dec 06 07:28:19 crc kubenswrapper[4809]: I1206 07:28:18.782259 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-engine-7fcd5d8964-kw6sz" podUID="a71b0d8f-059d-4801-a2e4-6dc8ec44c78a" containerName="heat-engine" probeResult="failure" output="command timed out" Dec 06 07:28:19 crc kubenswrapper[4809]: I1206 07:28:18.782595 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/heat-engine-7fcd5d8964-kw6sz" podUID="a71b0d8f-059d-4801-a2e4-6dc8ec44c78a" containerName="heat-engine" probeResult="failure" output="command timed out" Dec 06 07:28:19 crc kubenswrapper[4809]: I1206 07:28:18.855587 4809 patch_prober.go:28] interesting pod/image-registry-66df7c8f76-dgspm container/registry namespace/openshift-image-registry: Liveness probe status=failure output="Get \"https://10.217.0.62:5000/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:19 crc kubenswrapper[4809]: I1206 07:28:18.855655 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-image-registry/image-registry-66df7c8f76-dgspm" podUID="26c71fe2-c0e4-405d-a6c6-2ce933013cba" containerName="registry" probeResult="failure" output="Get \"https://10.217.0.62:5000/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:19 crc kubenswrapper[4809]: I1206 07:28:18.855657 4809 patch_prober.go:28] interesting pod/image-registry-66df7c8f76-dgspm container/registry namespace/openshift-image-registry: Readiness probe status=failure output="Get \"https://10.217.0.62:5000/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:19 crc kubenswrapper[4809]: I1206 07:28:18.855754 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-image-registry/image-registry-66df7c8f76-dgspm" podUID="26c71fe2-c0e4-405d-a6c6-2ce933013cba" containerName="registry" probeResult="failure" output="Get \"https://10.217.0.62:5000/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:19 crc kubenswrapper[4809]: I1206 07:28:18.991204 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/metallb-operator-controller-manager-6f667b6558-c67wt" podUID="69484422-8317-4050-945c-9c9b127552b1" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.91:8080/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:19 crc kubenswrapper[4809]: E1206 07:28:19.421354 4809 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="2.034s" Dec 06 07:28:19 crc kubenswrapper[4809]: I1206 07:28:19.443286 4809 trace.go:236] Trace[1636563243]: "Calculate volume metrics of config-data for pod openstack/cinder-api-0" (06-Dec-2025 07:28:18.334) (total time: 1082ms): Dec 06 07:28:19 crc kubenswrapper[4809]: Trace[1636563243]: [1.082423963s] [1.082423963s] END Dec 06 07:28:19 crc kubenswrapper[4809]: I1206 07:28:19.446964 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.562213651s: [/var/lib/containers/storage/overlay/5c1abc02f4a525f89d7e58a96747fa5f46e3d6f9365299c0c4fb630e1bb526ad/diff /var/log/pods/openstack-operators_openstack-operator-index-hv5l4_45b1df34-ffd2-42ca-85ec-885efb8ec7e5/registry-server/0.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:28:19 crc kubenswrapper[4809]: I1206 07:28:19.446027 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.560623998s: [/var/lib/containers/storage/overlay/5f72b826028cd34501041610a68de257c31a0b97eeaa82638ffeabe81b2fe169/diff /var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager-recovery-controller/0.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:28:19 crc kubenswrapper[4809]: I1206 07:28:19.613025 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-cell1-galera-0" podUID="e7056938-6e36-4631-adfb-01f116eba651" containerName="galera" containerID="cri-o://c5e7932399fd9c59936d31395b31c6f5cc98b66aad28f9a1faffee5aa4d0028c" gracePeriod=20 Dec 06 07:28:19 crc kubenswrapper[4809]: I1206 07:28:19.620172 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/metallb-operator-webhook-server-765b7d7f99-8dngt" podUID="60798ea0-75be-4a0c-a8c6-9fb431ba0e67" containerName="webhook-server" probeResult="failure" output="Get \"http://10.217.0.93:7472/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:19 crc kubenswrapper[4809]: I1206 07:28:19.620426 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/metallb-operator-webhook-server-765b7d7f99-8dngt" podUID="60798ea0-75be-4a0c-a8c6-9fb431ba0e67" containerName="webhook-server" probeResult="failure" output="Get \"http://10.217.0.93:7472/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:19 crc kubenswrapper[4809]: I1206 07:28:19.630484 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-scheduler-0" podUID="47aec31f-059f-4f3c-baef-4f9406e8cd4c" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 07:28:19 crc kubenswrapper[4809]: I1206 07:28:19.638998 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-galera-0" podUID="7c596e14-b96f-4e5f-a8ac-3dc66fa0276d" containerName="galera" containerID="cri-o://a33b2b043d42437efe270d1e3ec31c3d840ad9b765f7e171518bdd9f6ff5ad80" gracePeriod=19 Dec 06 07:28:19 crc kubenswrapper[4809]: I1206 07:28:19.782595 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/redhat-marketplace-zjvtm" podUID="0714d252-8cd5-4a99-9050-8e383ad64885" containerName="registry-server" probeResult="failure" output="command timed out" Dec 06 07:28:19 crc kubenswrapper[4809]: I1206 07:28:19.783556 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-zjvtm" podUID="0714d252-8cd5-4a99-9050-8e383ad64885" containerName="registry-server" probeResult="failure" output="command timed out" Dec 06 07:28:20 crc kubenswrapper[4809]: I1206 07:28:20.510186 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/controller-f8648f98b-d4fgg" podUID="073938c5-bc80-47d3-91e8-1d6d7da0f000" containerName="controller" probeResult="failure" output="Get \"http://10.217.0.95:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:20 crc kubenswrapper[4809]: I1206 07:28:20.510307 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/controller-f8648f98b-d4fgg" podUID="073938c5-bc80-47d3-91e8-1d6d7da0f000" containerName="controller" probeResult="failure" output="Get \"http://10.217.0.95:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:20 crc kubenswrapper[4809]: I1206 07:28:20.625796 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/kube-state-metrics-0" podUID="36f27192-451f-4ce2-afda-c0e66a6028b2" containerName="kube-state-metrics" probeResult="failure" output="Get \"https://10.217.1.3:8081/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:20 crc kubenswrapper[4809]: I1206 07:28:20.625884 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/kube-state-metrics-0" podUID="36f27192-451f-4ce2-afda-c0e66a6028b2" containerName="kube-state-metrics" probeResult="failure" output="Get \"https://10.217.1.3:8080/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:20 crc kubenswrapper[4809]: I1206 07:28:20.626674 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/kube-state-metrics-0" Dec 06 07:28:20 crc kubenswrapper[4809]: I1206 07:28:20.627829 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-state-metrics" containerStatusID={"Type":"cri-o","ID":"b2f3204f4908e566576640527cae5ec93f5e360f0125a482c46c8b15ed867b25"} pod="openstack/kube-state-metrics-0" containerMessage="Container kube-state-metrics failed liveness probe, will be restarted" Dec 06 07:28:20 crc kubenswrapper[4809]: I1206 07:28:20.627888 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="36f27192-451f-4ce2-afda-c0e66a6028b2" containerName="kube-state-metrics" containerID="cri-o://b2f3204f4908e566576640527cae5ec93f5e360f0125a482c46c8b15ed867b25" gracePeriod=30 Dec 06 07:28:21 crc kubenswrapper[4809]: I1206 07:28:20.832664 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-58897d9998-hlvjb_f3a30985-f268-44de-93b9-bce0f9aebc6d/console-operator/0.log" Dec 06 07:28:21 crc kubenswrapper[4809]: I1206 07:28:20.832766 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-hlvjb" event={"ID":"f3a30985-f268-44de-93b9-bce0f9aebc6d","Type":"ContainerStarted","Data":"eaab16e1003b3e7833afe03786dc4841a92a9e5dc3e49fb0fc1fe43963f75fb6"} Dec 06 07:28:21 crc kubenswrapper[4809]: I1206 07:28:20.833037 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-hlvjb" Dec 06 07:28:21 crc kubenswrapper[4809]: I1206 07:28:20.833284 4809 patch_prober.go:28] interesting pod/console-operator-58897d9998-hlvjb container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/readyz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Dec 06 07:28:21 crc kubenswrapper[4809]: I1206 07:28:20.833320 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-hlvjb" podUID="f3a30985-f268-44de-93b9-bce0f9aebc6d" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/readyz\": dial tcp 10.217.0.24:8443: connect: connection refused" Dec 06 07:28:21 crc kubenswrapper[4809]: I1206 07:28:21.004229 4809 generic.go:334] "Generic (PLEG): container finished" podID="e5e4e451-e3ee-434e-8e24-0c795925a48e" containerID="17e00e064844bbaa3aeb76dd9b61ea649de4b3f1cdb9d6b7b0238d8f5b30b9a1" exitCode=0 Dec 06 07:28:21 crc kubenswrapper[4809]: I1206 07:28:21.004311 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jjf9r" event={"ID":"e5e4e451-e3ee-434e-8e24-0c795925a48e","Type":"ContainerDied","Data":"17e00e064844bbaa3aeb76dd9b61ea649de4b3f1cdb9d6b7b0238d8f5b30b9a1"} Dec 06 07:28:21 crc kubenswrapper[4809]: I1206 07:28:21.313684 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-tdjqn" podUID="1afc2f7d-0519-4486-a0f5-4c110fc28e53" containerName="registry-server" probeResult="failure" output=< Dec 06 07:28:21 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Dec 06 07:28:21 crc kubenswrapper[4809]: > Dec 06 07:28:21 crc kubenswrapper[4809]: I1206 07:28:21.372832 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/infra-operator-controller-manager-57548d458d-4b9l8" podUID="29c6d2cb-27fc-4b79-969a-462be41663e3" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.106:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:21 crc kubenswrapper[4809]: I1206 07:28:21.373151 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/infra-operator-controller-manager-57548d458d-4b9l8" podUID="29c6d2cb-27fc-4b79-969a-462be41663e3" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.106:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:21 crc kubenswrapper[4809]: I1206 07:28:21.373239 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/infra-operator-controller-manager-57548d458d-4b9l8" Dec 06 07:28:21 crc kubenswrapper[4809]: I1206 07:28:21.375439 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="manager" containerStatusID={"Type":"cri-o","ID":"58b3abc20e667a1fcfa7e8dd2925566cfe5ccf91513884a14368e407258af682"} pod="openstack-operators/infra-operator-controller-manager-57548d458d-4b9l8" containerMessage="Container manager failed liveness probe, will be restarted" Dec 06 07:28:21 crc kubenswrapper[4809]: I1206 07:28:21.375527 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/infra-operator-controller-manager-57548d458d-4b9l8" podUID="29c6d2cb-27fc-4b79-969a-462be41663e3" containerName="manager" containerID="cri-o://58b3abc20e667a1fcfa7e8dd2925566cfe5ccf91513884a14368e407258af682" gracePeriod=10 Dec 06 07:28:22 crc kubenswrapper[4809]: I1206 07:28:22.013523 4809 patch_prober.go:28] interesting pod/console-operator-58897d9998-hlvjb container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/readyz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Dec 06 07:28:22 crc kubenswrapper[4809]: I1206 07:28:22.013920 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-hlvjb" podUID="f3a30985-f268-44de-93b9-bce0f9aebc6d" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/readyz\": dial tcp 10.217.0.24:8443: connect: connection refused" Dec 06 07:28:23 crc kubenswrapper[4809]: I1206 07:28:23.033341 4809 generic.go:334] "Generic (PLEG): container finished" podID="36f27192-451f-4ce2-afda-c0e66a6028b2" containerID="b2f3204f4908e566576640527cae5ec93f5e360f0125a482c46c8b15ed867b25" exitCode=2 Dec 06 07:28:23 crc kubenswrapper[4809]: I1206 07:28:23.033546 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"36f27192-451f-4ce2-afda-c0e66a6028b2","Type":"ContainerDied","Data":"b2f3204f4908e566576640527cae5ec93f5e360f0125a482c46c8b15ed867b25"} Dec 06 07:28:23 crc kubenswrapper[4809]: I1206 07:28:23.046233 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-scheduler-0" podUID="47aec31f-059f-4f3c-baef-4f9406e8cd4c" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 07:28:23 crc kubenswrapper[4809]: I1206 07:28:23.049381 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress_router-default-5444994796-n6z57_e5356f1d-5afd-47bd-89af-72abfa8f901f/router/0.log" Dec 06 07:28:24 crc kubenswrapper[4809]: I1206 07:28:24.060883 4809 generic.go:334] "Generic (PLEG): container finished" podID="29c6d2cb-27fc-4b79-969a-462be41663e3" containerID="58b3abc20e667a1fcfa7e8dd2925566cfe5ccf91513884a14368e407258af682" exitCode=0 Dec 06 07:28:24 crc kubenswrapper[4809]: I1206 07:28:24.061315 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-4b9l8" event={"ID":"29c6d2cb-27fc-4b79-969a-462be41663e3","Type":"ContainerDied","Data":"58b3abc20e667a1fcfa7e8dd2925566cfe5ccf91513884a14368e407258af682"} Dec 06 07:28:24 crc kubenswrapper[4809]: I1206 07:28:24.063372 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress_router-default-5444994796-n6z57_e5356f1d-5afd-47bd-89af-72abfa8f901f/router/0.log" Dec 06 07:28:24 crc kubenswrapper[4809]: I1206 07:28:24.063489 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-n6z57" event={"ID":"e5356f1d-5afd-47bd-89af-72abfa8f901f","Type":"ContainerStarted","Data":"613de4ddeb8e23739a93690607e88af5b98637aabb65463998fba8b3470ad3ac"} Dec 06 07:28:24 crc kubenswrapper[4809]: I1206 07:28:24.065683 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jjf9r" event={"ID":"e5e4e451-e3ee-434e-8e24-0c795925a48e","Type":"ContainerStarted","Data":"e9eb7fc1647acf7f8c7ab69e5e62e826151cc96d65627bc831f7829acead5a58"} Dec 06 07:28:24 crc kubenswrapper[4809]: I1206 07:28:24.066200 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jjf9r" Dec 06 07:28:24 crc kubenswrapper[4809]: I1206 07:28:24.401565 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jjf9r" Dec 06 07:28:24 crc kubenswrapper[4809]: I1206 07:28:24.449996 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-hlvjb" Dec 06 07:28:24 crc kubenswrapper[4809]: I1206 07:28:24.617439 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-n6z57" Dec 06 07:28:24 crc kubenswrapper[4809]: I1206 07:28:24.617735 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-n6z57" Dec 06 07:28:24 crc kubenswrapper[4809]: I1206 07:28:24.624791 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-n6z57" Dec 06 07:28:25 crc kubenswrapper[4809]: I1206 07:28:25.079415 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-n6z57" Dec 06 07:28:25 crc kubenswrapper[4809]: I1206 07:28:25.594333 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-scheduler-0" podUID="47aec31f-059f-4f3c-baef-4f9406e8cd4c" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 07:28:25 crc kubenswrapper[4809]: I1206 07:28:25.594726 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 06 07:28:25 crc kubenswrapper[4809]: I1206 07:28:25.595674 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cinder-scheduler" containerStatusID={"Type":"cri-o","ID":"fefd478b5d0c758d8e4f6cc5ea01120bb084cf57e583dab89c1587610263b6bf"} pod="openstack/cinder-scheduler-0" containerMessage="Container cinder-scheduler failed liveness probe, will be restarted" Dec 06 07:28:25 crc kubenswrapper[4809]: I1206 07:28:25.595742 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="47aec31f-059f-4f3c-baef-4f9406e8cd4c" containerName="cinder-scheduler" containerID="cri-o://fefd478b5d0c758d8e4f6cc5ea01120bb084cf57e583dab89c1587610263b6bf" gracePeriod=30 Dec 06 07:28:25 crc kubenswrapper[4809]: I1206 07:28:25.625589 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/kube-state-metrics-0" podUID="36f27192-451f-4ce2-afda-c0e66a6028b2" containerName="kube-state-metrics" probeResult="failure" output="Get \"https://10.217.1.3:8081/readyz\": dial tcp 10.217.1.3:8081: connect: connection refused" Dec 06 07:28:25 crc kubenswrapper[4809]: I1206 07:28:25.626310 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 06 07:28:25 crc kubenswrapper[4809]: E1206 07:28:25.937834 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a33b2b043d42437efe270d1e3ec31c3d840ad9b765f7e171518bdd9f6ff5ad80" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Dec 06 07:28:25 crc kubenswrapper[4809]: E1206 07:28:25.939539 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a33b2b043d42437efe270d1e3ec31c3d840ad9b765f7e171518bdd9f6ff5ad80" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Dec 06 07:28:25 crc kubenswrapper[4809]: E1206 07:28:25.941144 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a33b2b043d42437efe270d1e3ec31c3d840ad9b765f7e171518bdd9f6ff5ad80" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Dec 06 07:28:25 crc kubenswrapper[4809]: E1206 07:28:25.941271 4809 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="7c596e14-b96f-4e5f-a8ac-3dc66fa0276d" containerName="galera" Dec 06 07:28:26 crc kubenswrapper[4809]: I1206 07:28:26.095632 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-4b9l8" event={"ID":"29c6d2cb-27fc-4b79-969a-462be41663e3","Type":"ContainerStarted","Data":"c3d7b07a35f99380c13ce497031c8c066744e88b952fd03d344dad54b65be169"} Dec 06 07:28:26 crc kubenswrapper[4809]: I1206 07:28:26.103389 4809 generic.go:334] "Generic (PLEG): container finished" podID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerID="1ed2ac2be19d734f28ddb67018abd7ca0033f58827c7e4ab6c9c7cfe52be430c" exitCode=0 Dec 06 07:28:26 crc kubenswrapper[4809]: I1206 07:28:26.103480 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3d4b42d3-04a4-40e3-803b-11c77024f612","Type":"ContainerDied","Data":"1ed2ac2be19d734f28ddb67018abd7ca0033f58827c7e4ab6c9c7cfe52be430c"} Dec 06 07:28:26 crc kubenswrapper[4809]: I1206 07:28:26.103767 4809 scope.go:117] "RemoveContainer" containerID="92f800a6ae2ed5e64c1afe14ad830a380ec334e44b69d5e189323d25b3c22531" Dec 06 07:28:26 crc kubenswrapper[4809]: I1206 07:28:26.564224 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:28:27 crc kubenswrapper[4809]: I1206 07:28:27.117801 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-4b9l8" Dec 06 07:28:27 crc kubenswrapper[4809]: E1206 07:28:27.330322 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c5e7932399fd9c59936d31395b31c6f5cc98b66aad28f9a1faffee5aa4d0028c" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Dec 06 07:28:28 crc kubenswrapper[4809]: E1206 07:28:27.331830 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c5e7932399fd9c59936d31395b31c6f5cc98b66aad28f9a1faffee5aa4d0028c" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Dec 06 07:28:28 crc kubenswrapper[4809]: E1206 07:28:27.333147 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c5e7932399fd9c59936d31395b31c6f5cc98b66aad28f9a1faffee5aa4d0028c" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Dec 06 07:28:28 crc kubenswrapper[4809]: E1206 07:28:27.333200 4809 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="e7056938-6e36-4631-adfb-01f116eba651" containerName="galera" Dec 06 07:28:28 crc kubenswrapper[4809]: I1206 07:28:27.503750 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-tdjqn" podUID="1afc2f7d-0519-4486-a0f5-4c110fc28e53" containerName="registry-server" probeResult="failure" output=< Dec 06 07:28:28 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Dec 06 07:28:28 crc kubenswrapper[4809]: > Dec 06 07:28:28 crc kubenswrapper[4809]: I1206 07:28:28.414165 4809 patch_prober.go:28] interesting pod/route-controller-manager-6cd65596c4-9z8ww container/route-controller-manager namespace/openshift-route-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.60:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:28 crc kubenswrapper[4809]: I1206 07:28:28.414233 4809 patch_prober.go:28] interesting pod/route-controller-manager-6cd65596c4-9z8ww container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.60:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:28 crc kubenswrapper[4809]: I1206 07:28:28.414296 4809 patch_prober.go:28] interesting pod/controller-manager-7f967c4b75-bhj49 container/controller-manager namespace/openshift-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.61:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:28 crc kubenswrapper[4809]: I1206 07:28:28.414311 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6cd65596c4-9z8ww" podUID="8e47b3dc-260e-46a6-9ce2-2e4f4a8a5977" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.60:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:28 crc kubenswrapper[4809]: I1206 07:28:28.414355 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-controller-manager/controller-manager-7f967c4b75-bhj49" podUID="8fa68b06-fd0e-4a09-97ee-35d77cae7b80" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.61:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:28 crc kubenswrapper[4809]: I1206 07:28:28.414232 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-route-controller-manager/route-controller-manager-6cd65596c4-9z8ww" podUID="8e47b3dc-260e-46a6-9ce2-2e4f4a8a5977" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.60:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:28 crc kubenswrapper[4809]: I1206 07:28:28.455064 4809 patch_prober.go:28] interesting pod/controller-manager-7f967c4b75-bhj49 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.61:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:28 crc kubenswrapper[4809]: I1206 07:28:28.455119 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-7f967c4b75-bhj49" podUID="8fa68b06-fd0e-4a09-97ee-35d77cae7b80" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.61:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:28 crc kubenswrapper[4809]: I1206 07:28:28.867118 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.01344674s: [/var/lib/containers/storage/overlay/c702a860905547bf5efddf30aea5fb720f6263bdea245e32145b98e5f8164715/diff /var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_71bb4a3aecc4ba5b26c4b7318770ce13/kube-apiserver-check-endpoints/0.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:28:28 crc kubenswrapper[4809]: I1206 07:28:28.867171 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.377061466s: [/var/lib/containers/storage/overlay/0adab0c53071e96981e778d5d3ed2d92d680727259aee9d7110480771915fd54/diff /var/log/pods/openstack_ovn-northd-0_1360c733-da74-4d37-8842-75050a356334/openstack-network-exporter/0.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:28:28 crc kubenswrapper[4809]: I1206 07:28:28.981218 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.749086952s: [/var/lib/containers/storage/overlay/1de327ae8706779b9c5821fbd96eee6addd4aaed4b6108c9539100d37e03001d/diff /var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_71bb4a3aecc4ba5b26c4b7318770ce13/kube-apiserver-cert-regeneration-controller/0.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:28:29 crc kubenswrapper[4809]: I1206 07:28:29.085179 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="hostpath-provisioner/csi-hostpathplugin-p2lkt" podUID="c598f580-a438-4015-9a8a-f98b3b2b0b3d" containerName="hostpath-provisioner" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 07:28:30 crc kubenswrapper[4809]: I1206 07:28:30.299613 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-4b9l8" Dec 06 07:28:34 crc kubenswrapper[4809]: I1206 07:28:34.496438 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:28:34 crc kubenswrapper[4809]: I1206 07:28:34.497498 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:28:34 crc kubenswrapper[4809]: I1206 07:28:34.856365 4809 patch_prober.go:28] interesting pod/observability-operator-d8bb48f5d-bfc8s container/operator namespace/openshift-operators: Liveness probe status=failure output="Get \"http://10.217.0.22:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:28:34 crc kubenswrapper[4809]: I1206 07:28:34.857029 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operators/observability-operator-d8bb48f5d-bfc8s" podUID="6d4f55df-5dc2-4e72-b7f8-cee9229da476" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.22:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:35 crc kubenswrapper[4809]: I1206 07:28:35.625533 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/kube-state-metrics-0" podUID="36f27192-451f-4ce2-afda-c0e66a6028b2" containerName="kube-state-metrics" probeResult="failure" output="Get \"https://10.217.1.3:8081/readyz\": dial tcp 10.217.1.3:8081: connect: connection refused" Dec 06 07:28:35 crc kubenswrapper[4809]: E1206 07:28:35.937088 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a33b2b043d42437efe270d1e3ec31c3d840ad9b765f7e171518bdd9f6ff5ad80" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Dec 06 07:28:35 crc kubenswrapper[4809]: E1206 07:28:35.943606 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a33b2b043d42437efe270d1e3ec31c3d840ad9b765f7e171518bdd9f6ff5ad80" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Dec 06 07:28:35 crc kubenswrapper[4809]: E1206 07:28:35.948788 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a33b2b043d42437efe270d1e3ec31c3d840ad9b765f7e171518bdd9f6ff5ad80" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Dec 06 07:28:35 crc kubenswrapper[4809]: E1206 07:28:35.948874 4809 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="7c596e14-b96f-4e5f-a8ac-3dc66fa0276d" containerName="galera" Dec 06 07:28:36 crc kubenswrapper[4809]: I1206 07:28:36.232747 4809 generic.go:334] "Generic (PLEG): container finished" podID="973f75f0-3e24-4648-b57b-864065deb9b1" containerID="54afc4c460a902cc89e61598f063fb9f68c0ca80c46e96e0e702615a45bab812" exitCode=1 Dec 06 07:28:36 crc kubenswrapper[4809]: I1206 07:28:36.232817 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"973f75f0-3e24-4648-b57b-864065deb9b1","Type":"ContainerDied","Data":"54afc4c460a902cc89e61598f063fb9f68c0ca80c46e96e0e702615a45bab812"} Dec 06 07:28:36 crc kubenswrapper[4809]: I1206 07:28:36.237860 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"36f27192-451f-4ce2-afda-c0e66a6028b2","Type":"ContainerStarted","Data":"34e6315e93153fe76bd0df1e545866c105c90c780754185f4256176a0ac9a970"} Dec 06 07:28:36 crc kubenswrapper[4809]: I1206 07:28:36.637259 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tdjqn" Dec 06 07:28:36 crc kubenswrapper[4809]: I1206 07:28:36.691580 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tdjqn" Dec 06 07:28:36 crc kubenswrapper[4809]: I1206 07:28:36.880623 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tdjqn"] Dec 06 07:28:37 crc kubenswrapper[4809]: E1206 07:28:37.329403 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c5e7932399fd9c59936d31395b31c6f5cc98b66aad28f9a1faffee5aa4d0028c" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Dec 06 07:28:37 crc kubenswrapper[4809]: E1206 07:28:37.330572 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c5e7932399fd9c59936d31395b31c6f5cc98b66aad28f9a1faffee5aa4d0028c" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Dec 06 07:28:37 crc kubenswrapper[4809]: E1206 07:28:37.331983 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c5e7932399fd9c59936d31395b31c6f5cc98b66aad28f9a1faffee5aa4d0028c" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Dec 06 07:28:37 crc kubenswrapper[4809]: E1206 07:28:37.332018 4809 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="e7056938-6e36-4631-adfb-01f116eba651" containerName="galera" Dec 06 07:28:37 crc kubenswrapper[4809]: I1206 07:28:37.759863 4809 generic.go:334] "Generic (PLEG): container finished" podID="47aec31f-059f-4f3c-baef-4f9406e8cd4c" containerID="fefd478b5d0c758d8e4f6cc5ea01120bb084cf57e583dab89c1587610263b6bf" exitCode=0 Dec 06 07:28:37 crc kubenswrapper[4809]: I1206 07:28:37.760351 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"47aec31f-059f-4f3c-baef-4f9406e8cd4c","Type":"ContainerDied","Data":"fefd478b5d0c758d8e4f6cc5ea01120bb084cf57e583dab89c1587610263b6bf"} Dec 06 07:28:38 crc kubenswrapper[4809]: I1206 07:28:38.776853 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"973f75f0-3e24-4648-b57b-864065deb9b1","Type":"ContainerDied","Data":"1ce794f012a2205d331438d97153b6f6aeee06635e5a44bc945d8ad86fc0fbb2"} Dec 06 07:28:38 crc kubenswrapper[4809]: I1206 07:28:38.777243 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-tdjqn" podUID="1afc2f7d-0519-4486-a0f5-4c110fc28e53" containerName="registry-server" containerID="cri-o://aa8a15cd520fe4652ff7fa9c351a8feac89f3ba113cfdb0ff068ddb24f3ead61" gracePeriod=2 Dec 06 07:28:38 crc kubenswrapper[4809]: I1206 07:28:38.777532 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ce794f012a2205d331438d97153b6f6aeee06635e5a44bc945d8ad86fc0fbb2" Dec 06 07:28:38 crc kubenswrapper[4809]: I1206 07:28:38.779982 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 06 07:28:38 crc kubenswrapper[4809]: I1206 07:28:38.900644 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 06 07:28:39 crc kubenswrapper[4809]: I1206 07:28:39.099410 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/973f75f0-3e24-4648-b57b-864065deb9b1-ca-certs\") pod \"973f75f0-3e24-4648-b57b-864065deb9b1\" (UID: \"973f75f0-3e24-4648-b57b-864065deb9b1\") " Dec 06 07:28:39 crc kubenswrapper[4809]: I1206 07:28:39.099507 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/973f75f0-3e24-4648-b57b-864065deb9b1-openstack-config-secret\") pod \"973f75f0-3e24-4648-b57b-864065deb9b1\" (UID: \"973f75f0-3e24-4648-b57b-864065deb9b1\") " Dec 06 07:28:39 crc kubenswrapper[4809]: I1206 07:28:39.099733 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/973f75f0-3e24-4648-b57b-864065deb9b1-ssh-key\") pod \"973f75f0-3e24-4648-b57b-864065deb9b1\" (UID: \"973f75f0-3e24-4648-b57b-864065deb9b1\") " Dec 06 07:28:39 crc kubenswrapper[4809]: I1206 07:28:39.099760 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"973f75f0-3e24-4648-b57b-864065deb9b1\" (UID: \"973f75f0-3e24-4648-b57b-864065deb9b1\") " Dec 06 07:28:39 crc kubenswrapper[4809]: I1206 07:28:39.099783 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/973f75f0-3e24-4648-b57b-864065deb9b1-test-operator-ephemeral-workdir\") pod \"973f75f0-3e24-4648-b57b-864065deb9b1\" (UID: \"973f75f0-3e24-4648-b57b-864065deb9b1\") " Dec 06 07:28:39 crc kubenswrapper[4809]: I1206 07:28:39.099861 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/973f75f0-3e24-4648-b57b-864065deb9b1-test-operator-ephemeral-temporary\") pod \"973f75f0-3e24-4648-b57b-864065deb9b1\" (UID: \"973f75f0-3e24-4648-b57b-864065deb9b1\") " Dec 06 07:28:39 crc kubenswrapper[4809]: I1206 07:28:39.099959 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bkqlh\" (UniqueName: \"kubernetes.io/projected/973f75f0-3e24-4648-b57b-864065deb9b1-kube-api-access-bkqlh\") pod \"973f75f0-3e24-4648-b57b-864065deb9b1\" (UID: \"973f75f0-3e24-4648-b57b-864065deb9b1\") " Dec 06 07:28:39 crc kubenswrapper[4809]: I1206 07:28:39.099980 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/973f75f0-3e24-4648-b57b-864065deb9b1-openstack-config\") pod \"973f75f0-3e24-4648-b57b-864065deb9b1\" (UID: \"973f75f0-3e24-4648-b57b-864065deb9b1\") " Dec 06 07:28:39 crc kubenswrapper[4809]: I1206 07:28:39.100031 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/973f75f0-3e24-4648-b57b-864065deb9b1-config-data\") pod \"973f75f0-3e24-4648-b57b-864065deb9b1\" (UID: \"973f75f0-3e24-4648-b57b-864065deb9b1\") " Dec 06 07:28:39 crc kubenswrapper[4809]: I1206 07:28:39.103762 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/973f75f0-3e24-4648-b57b-864065deb9b1-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "973f75f0-3e24-4648-b57b-864065deb9b1" (UID: "973f75f0-3e24-4648-b57b-864065deb9b1"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:28:39 crc kubenswrapper[4809]: I1206 07:28:39.104192 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/973f75f0-3e24-4648-b57b-864065deb9b1-config-data" (OuterVolumeSpecName: "config-data") pod "973f75f0-3e24-4648-b57b-864065deb9b1" (UID: "973f75f0-3e24-4648-b57b-864065deb9b1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:28:39 crc kubenswrapper[4809]: I1206 07:28:39.109445 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "test-operator-logs") pod "973f75f0-3e24-4648-b57b-864065deb9b1" (UID: "973f75f0-3e24-4648-b57b-864065deb9b1"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 07:28:39 crc kubenswrapper[4809]: I1206 07:28:39.112671 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/973f75f0-3e24-4648-b57b-864065deb9b1-kube-api-access-bkqlh" (OuterVolumeSpecName: "kube-api-access-bkqlh") pod "973f75f0-3e24-4648-b57b-864065deb9b1" (UID: "973f75f0-3e24-4648-b57b-864065deb9b1"). InnerVolumeSpecName "kube-api-access-bkqlh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:28:39 crc kubenswrapper[4809]: I1206 07:28:39.113175 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/973f75f0-3e24-4648-b57b-864065deb9b1-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "973f75f0-3e24-4648-b57b-864065deb9b1" (UID: "973f75f0-3e24-4648-b57b-864065deb9b1"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:28:39 crc kubenswrapper[4809]: I1206 07:28:39.147017 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/973f75f0-3e24-4648-b57b-864065deb9b1-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "973f75f0-3e24-4648-b57b-864065deb9b1" (UID: "973f75f0-3e24-4648-b57b-864065deb9b1"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:28:39 crc kubenswrapper[4809]: I1206 07:28:39.148394 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/973f75f0-3e24-4648-b57b-864065deb9b1-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "973f75f0-3e24-4648-b57b-864065deb9b1" (UID: "973f75f0-3e24-4648-b57b-864065deb9b1"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:28:39 crc kubenswrapper[4809]: I1206 07:28:39.150183 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/973f75f0-3e24-4648-b57b-864065deb9b1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "973f75f0-3e24-4648-b57b-864065deb9b1" (UID: "973f75f0-3e24-4648-b57b-864065deb9b1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:28:39 crc kubenswrapper[4809]: I1206 07:28:39.203097 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/973f75f0-3e24-4648-b57b-864065deb9b1-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:28:39 crc kubenswrapper[4809]: I1206 07:28:39.203344 4809 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/973f75f0-3e24-4648-b57b-864065deb9b1-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 06 07:28:39 crc kubenswrapper[4809]: I1206 07:28:39.203354 4809 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/973f75f0-3e24-4648-b57b-864065deb9b1-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 06 07:28:39 crc kubenswrapper[4809]: I1206 07:28:39.203364 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/973f75f0-3e24-4648-b57b-864065deb9b1-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 07:28:39 crc kubenswrapper[4809]: I1206 07:28:39.203399 4809 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Dec 06 07:28:39 crc kubenswrapper[4809]: I1206 07:28:39.203414 4809 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/973f75f0-3e24-4648-b57b-864065deb9b1-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 06 07:28:39 crc kubenswrapper[4809]: I1206 07:28:39.203424 4809 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/973f75f0-3e24-4648-b57b-864065deb9b1-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 06 07:28:39 crc kubenswrapper[4809]: I1206 07:28:39.203434 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bkqlh\" (UniqueName: \"kubernetes.io/projected/973f75f0-3e24-4648-b57b-864065deb9b1-kube-api-access-bkqlh\") on node \"crc\" DevicePath \"\"" Dec 06 07:28:39 crc kubenswrapper[4809]: I1206 07:28:39.915307 4809 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Dec 06 07:28:39 crc kubenswrapper[4809]: I1206 07:28:39.922136 4809 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Dec 06 07:28:41 crc kubenswrapper[4809]: I1206 07:28:40.939159 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-z95cs" podUID="a3b6197b-05d4-49b2-9a67-56b675046a1d" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:28:41 crc kubenswrapper[4809]: I1206 07:28:41.064110 4809 trace.go:236] Trace[697178761]: "Calculate volume metrics of env-overrides for pod openshift-ovn-kubernetes/ovnkube-node-vlv76" (06-Dec-2025 07:28:39.824) (total time: 1239ms): Dec 06 07:28:41 crc kubenswrapper[4809]: Trace[697178761]: [1.239810187s] [1.239810187s] END Dec 06 07:28:41 crc kubenswrapper[4809]: I1206 07:28:41.064193 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.685600395s: [/var/lib/containers/storage/overlay/e61778f8609182bf939721a218e98ad9c331c04fd4f399b160e28be62d6f7477/diff /var/log/pods/openstack_ceilometer-0_3d4b42d3-04a4-40e3-803b-11c77024f612/proxy-httpd/0.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:28:41 crc kubenswrapper[4809]: I1206 07:28:41.064842 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/973f75f0-3e24-4648-b57b-864065deb9b1-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "973f75f0-3e24-4648-b57b-864065deb9b1" (UID: "973f75f0-3e24-4648-b57b-864065deb9b1"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:28:41 crc kubenswrapper[4809]: I1206 07:28:41.112786 4809 generic.go:334] "Generic (PLEG): container finished" podID="7c596e14-b96f-4e5f-a8ac-3dc66fa0276d" containerID="a33b2b043d42437efe270d1e3ec31c3d840ad9b765f7e171518bdd9f6ff5ad80" exitCode=137 Dec 06 07:28:41 crc kubenswrapper[4809]: I1206 07:28:41.112843 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"7c596e14-b96f-4e5f-a8ac-3dc66fa0276d","Type":"ContainerDied","Data":"a33b2b043d42437efe270d1e3ec31c3d840ad9b765f7e171518bdd9f6ff5ad80"} Dec 06 07:28:41 crc kubenswrapper[4809]: I1206 07:28:41.113012 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 06 07:28:41 crc kubenswrapper[4809]: I1206 07:28:41.146642 4809 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/973f75f0-3e24-4648-b57b-864065deb9b1-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 06 07:28:41 crc kubenswrapper[4809]: I1206 07:28:41.343375 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 06 07:28:42 crc kubenswrapper[4809]: I1206 07:28:42.130049 4809 generic.go:334] "Generic (PLEG): container finished" podID="1afc2f7d-0519-4486-a0f5-4c110fc28e53" containerID="aa8a15cd520fe4652ff7fa9c351a8feac89f3ba113cfdb0ff068ddb24f3ead61" exitCode=0 Dec 06 07:28:42 crc kubenswrapper[4809]: I1206 07:28:42.130128 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tdjqn" event={"ID":"1afc2f7d-0519-4486-a0f5-4c110fc28e53","Type":"ContainerDied","Data":"aa8a15cd520fe4652ff7fa9c351a8feac89f3ba113cfdb0ff068ddb24f3ead61"} Dec 06 07:28:43 crc kubenswrapper[4809]: I1206 07:28:43.878266 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tdjqn" Dec 06 07:28:44 crc kubenswrapper[4809]: I1206 07:28:44.035734 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1afc2f7d-0519-4486-a0f5-4c110fc28e53-catalog-content\") pod \"1afc2f7d-0519-4486-a0f5-4c110fc28e53\" (UID: \"1afc2f7d-0519-4486-a0f5-4c110fc28e53\") " Dec 06 07:28:44 crc kubenswrapper[4809]: I1206 07:28:44.035904 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1afc2f7d-0519-4486-a0f5-4c110fc28e53-utilities\") pod \"1afc2f7d-0519-4486-a0f5-4c110fc28e53\" (UID: \"1afc2f7d-0519-4486-a0f5-4c110fc28e53\") " Dec 06 07:28:44 crc kubenswrapper[4809]: I1206 07:28:44.036180 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vhr7d\" (UniqueName: \"kubernetes.io/projected/1afc2f7d-0519-4486-a0f5-4c110fc28e53-kube-api-access-vhr7d\") pod \"1afc2f7d-0519-4486-a0f5-4c110fc28e53\" (UID: \"1afc2f7d-0519-4486-a0f5-4c110fc28e53\") " Dec 06 07:28:44 crc kubenswrapper[4809]: I1206 07:28:44.036627 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1afc2f7d-0519-4486-a0f5-4c110fc28e53-utilities" (OuterVolumeSpecName: "utilities") pod "1afc2f7d-0519-4486-a0f5-4c110fc28e53" (UID: "1afc2f7d-0519-4486-a0f5-4c110fc28e53"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:28:44 crc kubenswrapper[4809]: I1206 07:28:44.038278 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1afc2f7d-0519-4486-a0f5-4c110fc28e53-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 07:28:44 crc kubenswrapper[4809]: I1206 07:28:44.043011 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1afc2f7d-0519-4486-a0f5-4c110fc28e53-kube-api-access-vhr7d" (OuterVolumeSpecName: "kube-api-access-vhr7d") pod "1afc2f7d-0519-4486-a0f5-4c110fc28e53" (UID: "1afc2f7d-0519-4486-a0f5-4c110fc28e53"). InnerVolumeSpecName "kube-api-access-vhr7d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:28:44 crc kubenswrapper[4809]: I1206 07:28:44.119588 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1afc2f7d-0519-4486-a0f5-4c110fc28e53-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1afc2f7d-0519-4486-a0f5-4c110fc28e53" (UID: "1afc2f7d-0519-4486-a0f5-4c110fc28e53"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:28:44 crc kubenswrapper[4809]: I1206 07:28:44.141003 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vhr7d\" (UniqueName: \"kubernetes.io/projected/1afc2f7d-0519-4486-a0f5-4c110fc28e53-kube-api-access-vhr7d\") on node \"crc\" DevicePath \"\"" Dec 06 07:28:44 crc kubenswrapper[4809]: I1206 07:28:44.141048 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1afc2f7d-0519-4486-a0f5-4c110fc28e53-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 07:28:44 crc kubenswrapper[4809]: I1206 07:28:44.161960 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"7c596e14-b96f-4e5f-a8ac-3dc66fa0276d","Type":"ContainerStarted","Data":"44169e42ff49b18586bf406c2262a67d7808d4c679837e87f453ab403acd409c"} Dec 06 07:28:44 crc kubenswrapper[4809]: I1206 07:28:44.167150 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tdjqn" event={"ID":"1afc2f7d-0519-4486-a0f5-4c110fc28e53","Type":"ContainerDied","Data":"819736c810c39c649ee239ca383764071e0194a255c3c9f3fbe0a1d26fec6522"} Dec 06 07:28:44 crc kubenswrapper[4809]: I1206 07:28:44.167225 4809 scope.go:117] "RemoveContainer" containerID="aa8a15cd520fe4652ff7fa9c351a8feac89f3ba113cfdb0ff068ddb24f3ead61" Dec 06 07:28:44 crc kubenswrapper[4809]: I1206 07:28:44.168060 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tdjqn" Dec 06 07:28:44 crc kubenswrapper[4809]: I1206 07:28:44.228267 4809 scope.go:117] "RemoveContainer" containerID="6160c4cd03560d1ac5e4a708d1eadbb37c1656760d0991de463899178cdac4e8" Dec 06 07:28:44 crc kubenswrapper[4809]: I1206 07:28:44.253495 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tdjqn"] Dec 06 07:28:44 crc kubenswrapper[4809]: I1206 07:28:44.284286 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-tdjqn"] Dec 06 07:28:44 crc kubenswrapper[4809]: I1206 07:28:44.325983 4809 scope.go:117] "RemoveContainer" containerID="32acaa37e8a0496a4030ad09915728ae06ccd8d0aaf33ffbe9430b54f659749e" Dec 06 07:28:45 crc kubenswrapper[4809]: I1206 07:28:45.184641 4809 generic.go:334] "Generic (PLEG): container finished" podID="e7056938-6e36-4631-adfb-01f116eba651" containerID="c5e7932399fd9c59936d31395b31c6f5cc98b66aad28f9a1faffee5aa4d0028c" exitCode=137 Dec 06 07:28:45 crc kubenswrapper[4809]: I1206 07:28:45.188540 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e7056938-6e36-4631-adfb-01f116eba651","Type":"ContainerDied","Data":"c5e7932399fd9c59936d31395b31c6f5cc98b66aad28f9a1faffee5aa4d0028c"} Dec 06 07:28:45 crc kubenswrapper[4809]: I1206 07:28:45.201575 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3d4b42d3-04a4-40e3-803b-11c77024f612","Type":"ContainerStarted","Data":"31bab77bd42d875e0d08430f5790f3d44e3d6ffa21e0b32a8c7679a432bb1cc8"} Dec 06 07:28:45 crc kubenswrapper[4809]: I1206 07:28:45.202401 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerName="ceilometer-central-agent" containerID="cri-o://31bab77bd42d875e0d08430f5790f3d44e3d6ffa21e0b32a8c7679a432bb1cc8" gracePeriod=30 Dec 06 07:28:45 crc kubenswrapper[4809]: I1206 07:28:45.202448 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerName="proxy-httpd" containerID="cri-o://f7069d7aabb93d228c5974d5a56a8727514829ecffaab6e9caf693226d4fdde2" gracePeriod=30 Dec 06 07:28:45 crc kubenswrapper[4809]: I1206 07:28:45.202487 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerName="sg-core" containerID="cri-o://493dde793f5dce6b38eb58b439b1d65c77ff6c4a3277b1a17ad6b0e5a2f1c252" gracePeriod=30 Dec 06 07:28:45 crc kubenswrapper[4809]: I1206 07:28:45.202401 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerName="ceilometer-notification-agent" containerID="cri-o://10189cd18cf124194d1cd5f8103894926b7f3965e6baefbf9c437f773512728e" gracePeriod=30 Dec 06 07:28:45 crc kubenswrapper[4809]: I1206 07:28:45.404659 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1afc2f7d-0519-4486-a0f5-4c110fc28e53" path="/var/lib/kubelet/pods/1afc2f7d-0519-4486-a0f5-4c110fc28e53/volumes" Dec 06 07:28:45 crc kubenswrapper[4809]: I1206 07:28:45.950410 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 06 07:28:45 crc kubenswrapper[4809]: I1206 07:28:45.950482 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 06 07:28:47 crc kubenswrapper[4809]: I1206 07:28:47.224401 4809 generic.go:334] "Generic (PLEG): container finished" podID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerID="493dde793f5dce6b38eb58b439b1d65c77ff6c4a3277b1a17ad6b0e5a2f1c252" exitCode=2 Dec 06 07:28:47 crc kubenswrapper[4809]: I1206 07:28:47.224485 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3d4b42d3-04a4-40e3-803b-11c77024f612","Type":"ContainerDied","Data":"493dde793f5dce6b38eb58b439b1d65c77ff6c4a3277b1a17ad6b0e5a2f1c252"} Dec 06 07:28:47 crc kubenswrapper[4809]: E1206 07:28:47.328210 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c5e7932399fd9c59936d31395b31c6f5cc98b66aad28f9a1faffee5aa4d0028c is running failed: container process not found" containerID="c5e7932399fd9c59936d31395b31c6f5cc98b66aad28f9a1faffee5aa4d0028c" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Dec 06 07:28:47 crc kubenswrapper[4809]: E1206 07:28:47.328797 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c5e7932399fd9c59936d31395b31c6f5cc98b66aad28f9a1faffee5aa4d0028c is running failed: container process not found" containerID="c5e7932399fd9c59936d31395b31c6f5cc98b66aad28f9a1faffee5aa4d0028c" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Dec 06 07:28:47 crc kubenswrapper[4809]: E1206 07:28:47.329292 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c5e7932399fd9c59936d31395b31c6f5cc98b66aad28f9a1faffee5aa4d0028c is running failed: container process not found" containerID="c5e7932399fd9c59936d31395b31c6f5cc98b66aad28f9a1faffee5aa4d0028c" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Dec 06 07:28:47 crc kubenswrapper[4809]: E1206 07:28:47.329335 4809 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c5e7932399fd9c59936d31395b31c6f5cc98b66aad28f9a1faffee5aa4d0028c is running failed: container process not found" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="e7056938-6e36-4631-adfb-01f116eba651" containerName="galera" Dec 06 07:28:49 crc kubenswrapper[4809]: I1206 07:28:49.250218 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e7056938-6e36-4631-adfb-01f116eba651","Type":"ContainerStarted","Data":"94abe5ee3044a9cdaac5a19c7e43e83663f839759f665f598fb983a7ee0ea6ba"} Dec 06 07:28:49 crc kubenswrapper[4809]: I1206 07:28:49.254213 4809 generic.go:334] "Generic (PLEG): container finished" podID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerID="f7069d7aabb93d228c5974d5a56a8727514829ecffaab6e9caf693226d4fdde2" exitCode=0 Dec 06 07:28:49 crc kubenswrapper[4809]: I1206 07:28:49.254609 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3d4b42d3-04a4-40e3-803b-11c77024f612","Type":"ContainerDied","Data":"f7069d7aabb93d228c5974d5a56a8727514829ecffaab6e9caf693226d4fdde2"} Dec 06 07:28:50 crc kubenswrapper[4809]: I1206 07:28:50.640183 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.699326346s: [/var/lib/containers/storage/overlay/d73fc1db1aa512f87bd3804f83b1aac73063744550bfdcd27626a630c9052392/diff /var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-l8clc_44c524db-76cf-49c3-8c6e-1ad181d553ae/manager/0.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:28:50 crc kubenswrapper[4809]: I1206 07:28:50.703523 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 06 07:28:50 crc kubenswrapper[4809]: E1206 07:28:50.704349 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="973f75f0-3e24-4648-b57b-864065deb9b1" containerName="tempest-tests-tempest-tests-runner" Dec 06 07:28:50 crc kubenswrapper[4809]: I1206 07:28:50.704389 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="973f75f0-3e24-4648-b57b-864065deb9b1" containerName="tempest-tests-tempest-tests-runner" Dec 06 07:28:50 crc kubenswrapper[4809]: E1206 07:28:50.704438 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1afc2f7d-0519-4486-a0f5-4c110fc28e53" containerName="extract-utilities" Dec 06 07:28:50 crc kubenswrapper[4809]: I1206 07:28:50.704450 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1afc2f7d-0519-4486-a0f5-4c110fc28e53" containerName="extract-utilities" Dec 06 07:28:50 crc kubenswrapper[4809]: E1206 07:28:50.704495 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1afc2f7d-0519-4486-a0f5-4c110fc28e53" containerName="registry-server" Dec 06 07:28:50 crc kubenswrapper[4809]: I1206 07:28:50.704506 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1afc2f7d-0519-4486-a0f5-4c110fc28e53" containerName="registry-server" Dec 06 07:28:50 crc kubenswrapper[4809]: E1206 07:28:50.704536 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1afc2f7d-0519-4486-a0f5-4c110fc28e53" containerName="extract-content" Dec 06 07:28:50 crc kubenswrapper[4809]: I1206 07:28:50.704547 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1afc2f7d-0519-4486-a0f5-4c110fc28e53" containerName="extract-content" Dec 06 07:28:50 crc kubenswrapper[4809]: I1206 07:28:50.704903 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="973f75f0-3e24-4648-b57b-864065deb9b1" containerName="tempest-tests-tempest-tests-runner" Dec 06 07:28:50 crc kubenswrapper[4809]: I1206 07:28:50.706670 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="1afc2f7d-0519-4486-a0f5-4c110fc28e53" containerName="registry-server" Dec 06 07:28:50 crc kubenswrapper[4809]: I1206 07:28:50.708880 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 07:28:50 crc kubenswrapper[4809]: I1206 07:28:50.713015 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-cfdpf" Dec 06 07:28:50 crc kubenswrapper[4809]: I1206 07:28:50.773626 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 06 07:28:50 crc kubenswrapper[4809]: I1206 07:28:50.836877 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"25b967d8-604b-4b77-9414-9f3922a56bdc\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 07:28:50 crc kubenswrapper[4809]: I1206 07:28:50.837356 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99v4h\" (UniqueName: \"kubernetes.io/projected/25b967d8-604b-4b77-9414-9f3922a56bdc-kube-api-access-99v4h\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"25b967d8-604b-4b77-9414-9f3922a56bdc\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 07:28:50 crc kubenswrapper[4809]: I1206 07:28:50.939546 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99v4h\" (UniqueName: \"kubernetes.io/projected/25b967d8-604b-4b77-9414-9f3922a56bdc-kube-api-access-99v4h\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"25b967d8-604b-4b77-9414-9f3922a56bdc\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 07:28:50 crc kubenswrapper[4809]: I1206 07:28:50.939738 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"25b967d8-604b-4b77-9414-9f3922a56bdc\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 07:28:50 crc kubenswrapper[4809]: I1206 07:28:50.940520 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"25b967d8-604b-4b77-9414-9f3922a56bdc\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 07:28:50 crc kubenswrapper[4809]: I1206 07:28:50.998836 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99v4h\" (UniqueName: \"kubernetes.io/projected/25b967d8-604b-4b77-9414-9f3922a56bdc-kube-api-access-99v4h\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"25b967d8-604b-4b77-9414-9f3922a56bdc\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 07:28:50 crc kubenswrapper[4809]: I1206 07:28:50.999719 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"25b967d8-604b-4b77-9414-9f3922a56bdc\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 07:28:51 crc kubenswrapper[4809]: I1206 07:28:51.030326 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 07:28:51 crc kubenswrapper[4809]: I1206 07:28:51.608967 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 06 07:28:51 crc kubenswrapper[4809]: W1206 07:28:51.614670 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod25b967d8_604b_4b77_9414_9f3922a56bdc.slice/crio-3a5b5741e23b65d963e481f7ce187d512fda50a715276e4cfcb8c82f38f48bbd WatchSource:0}: Error finding container 3a5b5741e23b65d963e481f7ce187d512fda50a715276e4cfcb8c82f38f48bbd: Status 404 returned error can't find the container with id 3a5b5741e23b65d963e481f7ce187d512fda50a715276e4cfcb8c82f38f48bbd Dec 06 07:28:52 crc kubenswrapper[4809]: I1206 07:28:52.290826 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"25b967d8-604b-4b77-9414-9f3922a56bdc","Type":"ContainerStarted","Data":"3a5b5741e23b65d963e481f7ce187d512fda50a715276e4cfcb8c82f38f48bbd"} Dec 06 07:28:53 crc kubenswrapper[4809]: I1206 07:28:53.321568 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"47aec31f-059f-4f3c-baef-4f9406e8cd4c","Type":"ContainerStarted","Data":"5152e7a5d6d7ec682522b6cd5cfcced47273433161bf2ece5cd384d07c46c8ed"} Dec 06 07:28:59 crc kubenswrapper[4809]: I1206 07:28:59.310384 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 4.038529715s: [/var/lib/containers/storage/overlay/bee7688b6a39ee376132301bc740ddd887a64e3f5eb8734aac27dffc7e68e968/diff /var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-4s24n_35b44606-4fdd-4662-9550-adc5ed2478b3/manager/0.log]; will not log again for this container unless duration exceeds 3s Dec 06 07:28:59 crc kubenswrapper[4809]: I1206 07:28:59.339268 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 4.067670333s: [/var/lib/containers/storage/overlay/7dea7ace731e3d6dc3461e82faa7cba24577cbad99b87a8df9ed7e0c1604f68f/diff /var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-vh8d4_93397035-a1a7-4c92-bdeb-5d0091dea181/manager/0.log]; will not log again for this container unless duration exceeds 3s Dec 06 07:28:59 crc kubenswrapper[4809]: I1206 07:28:59.344671 4809 generic.go:334] "Generic (PLEG): container finished" podID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerID="10189cd18cf124194d1cd5f8103894926b7f3965e6baefbf9c437f773512728e" exitCode=-1 Dec 06 07:28:59 crc kubenswrapper[4809]: I1206 07:28:59.357495 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 4.093611545s: [/var/lib/containers/storage/overlay/bbecdf4a78b8288c4447514716ca28ff09bc7f3bea5a14922ba6a024fa0ea0cc/diff /var/log/pods/openshift-apiserver_apiserver-76f77b778f-tphf5_09a7af3d-6bf2-4b7a-93d9-cecc73d993a1/openshift-apiserver/0.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:28:59 crc kubenswrapper[4809]: E1206 07:28:59.445872 4809 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="4.057s" Dec 06 07:28:59 crc kubenswrapper[4809]: I1206 07:28:59.502989 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 06 07:28:59 crc kubenswrapper[4809]: I1206 07:28:59.503040 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3d4b42d3-04a4-40e3-803b-11c77024f612","Type":"ContainerDied","Data":"10189cd18cf124194d1cd5f8103894926b7f3965e6baefbf9c437f773512728e"} Dec 06 07:28:59 crc kubenswrapper[4809]: I1206 07:28:59.503081 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 06 07:29:00 crc kubenswrapper[4809]: I1206 07:29:00.180850 4809 trace.go:236] Trace[1287297862]: "Calculate volume metrics of trusted-ca for pod openshift-logging/collector-hmjlt" (06-Dec-2025 07:28:56.943) (total time: 3237ms): Dec 06 07:29:00 crc kubenswrapper[4809]: Trace[1287297862]: [3.237441535s] [3.237441535s] END Dec 06 07:29:00 crc kubenswrapper[4809]: I1206 07:29:00.183178 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 4.909862604s: [/var/lib/containers/storage/overlay/699679aa451488029a361e8a55a602bbec560f10abe3fa635d15f638f3127413/diff /var/log/pods/openstack_cinder-scheduler-0_47aec31f-059f-4f3c-baef-4f9406e8cd4c/probe/0.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:29:00 crc kubenswrapper[4809]: I1206 07:29:00.184157 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.696983673s: [/var/lib/containers/storage/overlay/ed7d0da6a08ea40ae11baf2aca33e01eb16e3bf197c6c2c91709921930b93773/diff /var/log/pods/openshift-authentication_oauth-openshift-55889b984c-nxw8t_404214e0-30aa-44ff-b2d0-fb3c127068f9/oauth-openshift/0.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:29:00 crc kubenswrapper[4809]: I1206 07:29:00.184562 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 4.913210395s: [/var/lib/containers/storage/overlay/a94bf85210c2e9b6bf289bfd780dd376a0964e8980d92b77b7a970ff037d2526/diff /var/log/pods/openstack_swift-proxy-58bb4f66d7-bdk65_d496af20-e3cf-4ce6-86f5-df841a9ba42c/proxy-server/0.log]; will not log again for this container unless duration exceeds 3s Dec 06 07:29:00 crc kubenswrapper[4809]: I1206 07:29:00.185582 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 4.912584067s: [/var/lib/containers/storage/overlay/d2a1f1c10baddced5bb330b4e0d261a0fff2c3fa386b446ff616e8e624e3b090/diff /var/log/pods/openstack_heat-api-7849fbcbcb-msd4w_dccd4eeb-4ba8-4200-8366-28d4ead2e476/heat-api/0.log]; will not log again for this container unless duration exceeds 3s Dec 06 07:29:00 crc kubenswrapper[4809]: I1206 07:29:00.202478 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 4.924930732s: [/var/lib/containers/storage/overlay/4828fb2dbb120e5bc814761acb6f53fff6e26bdd0bc7f09a5c6d2b1dab44f9ad/diff /var/log/pods/openstack_heat-cfnapi-794dfc48dc-pkjpx_b09a69de-0d5f-44f7-a3c9-2e6902da17b8/heat-cfnapi/0.log]; will not log again for this container unless duration exceeds 3s Dec 06 07:29:02 crc kubenswrapper[4809]: I1206 07:29:02.576648 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 06 07:29:02 crc kubenswrapper[4809]: I1206 07:29:02.595501 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="47aec31f-059f-4f3c-baef-4f9406e8cd4c" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 07:29:04 crc kubenswrapper[4809]: I1206 07:29:04.496617 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:29:04 crc kubenswrapper[4809]: I1206 07:29:04.497354 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:29:07 crc kubenswrapper[4809]: I1206 07:29:07.593728 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="47aec31f-059f-4f3c-baef-4f9406e8cd4c" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 07:29:08 crc kubenswrapper[4809]: I1206 07:29:08.010483 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.1.8:3000/\": dial tcp 10.217.1.8:3000: connect: connection refused" Dec 06 07:29:08 crc kubenswrapper[4809]: I1206 07:29:08.820308 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 06 07:29:08 crc kubenswrapper[4809]: I1206 07:29:08.937483 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="e7056938-6e36-4631-adfb-01f116eba651" containerName="galera" probeResult="failure" output=< Dec 06 07:29:08 crc kubenswrapper[4809]: wsrep_local_state_comment (Joined) differs from Synced Dec 06 07:29:08 crc kubenswrapper[4809]: > Dec 06 07:29:09 crc kubenswrapper[4809]: I1206 07:29:09.511736 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 06 07:29:09 crc kubenswrapper[4809]: I1206 07:29:09.623989 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 06 07:29:10 crc kubenswrapper[4809]: I1206 07:29:10.468882 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"25b967d8-604b-4b77-9414-9f3922a56bdc","Type":"ContainerStarted","Data":"81945f05503c7815e86c054cd329f99d8cea7c0511f209618c3ecc157deb6e0f"} Dec 06 07:29:10 crc kubenswrapper[4809]: I1206 07:29:10.491344 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.778365366 podStartE2EDuration="20.491325749s" podCreationTimestamp="2025-12-06 07:28:50 +0000 UTC" firstStartedPulling="2025-12-06 07:28:51.618281983 +0000 UTC m=+5856.507264925" lastFinishedPulling="2025-12-06 07:29:09.331242366 +0000 UTC m=+5874.220225308" observedRunningTime="2025-12-06 07:29:10.480307631 +0000 UTC m=+5875.369290573" watchObservedRunningTime="2025-12-06 07:29:10.491325749 +0000 UTC m=+5875.380308691" Dec 06 07:29:12 crc kubenswrapper[4809]: I1206 07:29:12.793347 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 06 07:29:15 crc kubenswrapper[4809]: I1206 07:29:15.898079 4809 generic.go:334] "Generic (PLEG): container finished" podID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerID="31bab77bd42d875e0d08430f5790f3d44e3d6ffa21e0b32a8c7679a432bb1cc8" exitCode=137 Dec 06 07:29:15 crc kubenswrapper[4809]: I1206 07:29:15.898740 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3d4b42d3-04a4-40e3-803b-11c77024f612","Type":"ContainerDied","Data":"31bab77bd42d875e0d08430f5790f3d44e3d6ffa21e0b32a8c7679a432bb1cc8"} Dec 06 07:29:15 crc kubenswrapper[4809]: I1206 07:29:15.898789 4809 scope.go:117] "RemoveContainer" containerID="1ed2ac2be19d734f28ddb67018abd7ca0033f58827c7e4ab6c9c7cfe52be430c" Dec 06 07:29:17 crc kubenswrapper[4809]: I1206 07:29:17.345191 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 07:29:17 crc kubenswrapper[4809]: I1206 07:29:17.434509 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d4b42d3-04a4-40e3-803b-11c77024f612-combined-ca-bundle\") pod \"3d4b42d3-04a4-40e3-803b-11c77024f612\" (UID: \"3d4b42d3-04a4-40e3-803b-11c77024f612\") " Dec 06 07:29:17 crc kubenswrapper[4809]: I1206 07:29:17.434687 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3d4b42d3-04a4-40e3-803b-11c77024f612-sg-core-conf-yaml\") pod \"3d4b42d3-04a4-40e3-803b-11c77024f612\" (UID: \"3d4b42d3-04a4-40e3-803b-11c77024f612\") " Dec 06 07:29:17 crc kubenswrapper[4809]: I1206 07:29:17.434720 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvzk9\" (UniqueName: \"kubernetes.io/projected/3d4b42d3-04a4-40e3-803b-11c77024f612-kube-api-access-pvzk9\") pod \"3d4b42d3-04a4-40e3-803b-11c77024f612\" (UID: \"3d4b42d3-04a4-40e3-803b-11c77024f612\") " Dec 06 07:29:17 crc kubenswrapper[4809]: I1206 07:29:17.434821 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3d4b42d3-04a4-40e3-803b-11c77024f612-run-httpd\") pod \"3d4b42d3-04a4-40e3-803b-11c77024f612\" (UID: \"3d4b42d3-04a4-40e3-803b-11c77024f612\") " Dec 06 07:29:17 crc kubenswrapper[4809]: I1206 07:29:17.434919 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3d4b42d3-04a4-40e3-803b-11c77024f612-log-httpd\") pod \"3d4b42d3-04a4-40e3-803b-11c77024f612\" (UID: \"3d4b42d3-04a4-40e3-803b-11c77024f612\") " Dec 06 07:29:17 crc kubenswrapper[4809]: I1206 07:29:17.434996 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d4b42d3-04a4-40e3-803b-11c77024f612-config-data\") pod \"3d4b42d3-04a4-40e3-803b-11c77024f612\" (UID: \"3d4b42d3-04a4-40e3-803b-11c77024f612\") " Dec 06 07:29:17 crc kubenswrapper[4809]: I1206 07:29:17.435047 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d4b42d3-04a4-40e3-803b-11c77024f612-scripts\") pod \"3d4b42d3-04a4-40e3-803b-11c77024f612\" (UID: \"3d4b42d3-04a4-40e3-803b-11c77024f612\") " Dec 06 07:29:17 crc kubenswrapper[4809]: I1206 07:29:17.435129 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d4b42d3-04a4-40e3-803b-11c77024f612-ceilometer-tls-certs\") pod \"3d4b42d3-04a4-40e3-803b-11c77024f612\" (UID: \"3d4b42d3-04a4-40e3-803b-11c77024f612\") " Dec 06 07:29:17 crc kubenswrapper[4809]: I1206 07:29:17.453056 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d4b42d3-04a4-40e3-803b-11c77024f612-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3d4b42d3-04a4-40e3-803b-11c77024f612" (UID: "3d4b42d3-04a4-40e3-803b-11c77024f612"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:29:17 crc kubenswrapper[4809]: I1206 07:29:17.456859 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d4b42d3-04a4-40e3-803b-11c77024f612-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3d4b42d3-04a4-40e3-803b-11c77024f612" (UID: "3d4b42d3-04a4-40e3-803b-11c77024f612"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:29:17 crc kubenswrapper[4809]: I1206 07:29:17.498837 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d4b42d3-04a4-40e3-803b-11c77024f612-kube-api-access-pvzk9" (OuterVolumeSpecName: "kube-api-access-pvzk9") pod "3d4b42d3-04a4-40e3-803b-11c77024f612" (UID: "3d4b42d3-04a4-40e3-803b-11c77024f612"). InnerVolumeSpecName "kube-api-access-pvzk9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:29:17 crc kubenswrapper[4809]: I1206 07:29:17.499552 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d4b42d3-04a4-40e3-803b-11c77024f612-scripts" (OuterVolumeSpecName: "scripts") pod "3d4b42d3-04a4-40e3-803b-11c77024f612" (UID: "3d4b42d3-04a4-40e3-803b-11c77024f612"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:29:17 crc kubenswrapper[4809]: I1206 07:29:17.537964 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvzk9\" (UniqueName: \"kubernetes.io/projected/3d4b42d3-04a4-40e3-803b-11c77024f612-kube-api-access-pvzk9\") on node \"crc\" DevicePath \"\"" Dec 06 07:29:17 crc kubenswrapper[4809]: I1206 07:29:17.538004 4809 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3d4b42d3-04a4-40e3-803b-11c77024f612-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 07:29:17 crc kubenswrapper[4809]: I1206 07:29:17.538018 4809 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3d4b42d3-04a4-40e3-803b-11c77024f612-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 07:29:17 crc kubenswrapper[4809]: I1206 07:29:17.538029 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d4b42d3-04a4-40e3-803b-11c77024f612-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:29:17 crc kubenswrapper[4809]: I1206 07:29:17.579630 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d4b42d3-04a4-40e3-803b-11c77024f612-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3d4b42d3-04a4-40e3-803b-11c77024f612" (UID: "3d4b42d3-04a4-40e3-803b-11c77024f612"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:29:17 crc kubenswrapper[4809]: I1206 07:29:17.604881 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 06 07:29:17 crc kubenswrapper[4809]: I1206 07:29:17.643413 4809 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3d4b42d3-04a4-40e3-803b-11c77024f612-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 07:29:17 crc kubenswrapper[4809]: I1206 07:29:17.715457 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d4b42d3-04a4-40e3-803b-11c77024f612-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "3d4b42d3-04a4-40e3-803b-11c77024f612" (UID: "3d4b42d3-04a4-40e3-803b-11c77024f612"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:29:17 crc kubenswrapper[4809]: I1206 07:29:17.740030 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d4b42d3-04a4-40e3-803b-11c77024f612-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3d4b42d3-04a4-40e3-803b-11c77024f612" (UID: "3d4b42d3-04a4-40e3-803b-11c77024f612"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:29:17 crc kubenswrapper[4809]: I1206 07:29:17.747054 4809 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d4b42d3-04a4-40e3-803b-11c77024f612-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 07:29:17 crc kubenswrapper[4809]: I1206 07:29:17.747727 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d4b42d3-04a4-40e3-803b-11c77024f612-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:29:17 crc kubenswrapper[4809]: I1206 07:29:17.753073 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d4b42d3-04a4-40e3-803b-11c77024f612-config-data" (OuterVolumeSpecName: "config-data") pod "3d4b42d3-04a4-40e3-803b-11c77024f612" (UID: "3d4b42d3-04a4-40e3-803b-11c77024f612"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:29:17 crc kubenswrapper[4809]: I1206 07:29:17.851589 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d4b42d3-04a4-40e3-803b-11c77024f612-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:29:17 crc kubenswrapper[4809]: I1206 07:29:17.933394 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3d4b42d3-04a4-40e3-803b-11c77024f612","Type":"ContainerDied","Data":"a8230327f933146f837ab8150b34c63856481bed9f0cfc001a394c94f516fded"} Dec 06 07:29:17 crc kubenswrapper[4809]: I1206 07:29:17.933450 4809 scope.go:117] "RemoveContainer" containerID="31bab77bd42d875e0d08430f5790f3d44e3d6ffa21e0b32a8c7679a432bb1cc8" Dec 06 07:29:17 crc kubenswrapper[4809]: I1206 07:29:17.933453 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 07:29:17 crc kubenswrapper[4809]: I1206 07:29:17.961442 4809 scope.go:117] "RemoveContainer" containerID="f7069d7aabb93d228c5974d5a56a8727514829ecffaab6e9caf693226d4fdde2" Dec 06 07:29:17 crc kubenswrapper[4809]: I1206 07:29:17.979477 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:29:17 crc kubenswrapper[4809]: I1206 07:29:17.992699 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.009005 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:29:18 crc kubenswrapper[4809]: E1206 07:29:18.009533 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerName="ceilometer-central-agent" Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.009554 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerName="ceilometer-central-agent" Dec 06 07:29:18 crc kubenswrapper[4809]: E1206 07:29:18.009567 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerName="ceilometer-central-agent" Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.009575 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerName="ceilometer-central-agent" Dec 06 07:29:18 crc kubenswrapper[4809]: E1206 07:29:18.009601 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerName="ceilometer-notification-agent" Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.009611 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerName="ceilometer-notification-agent" Dec 06 07:29:18 crc kubenswrapper[4809]: E1206 07:29:18.009661 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerName="proxy-httpd" Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.009669 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerName="proxy-httpd" Dec 06 07:29:18 crc kubenswrapper[4809]: E1206 07:29:18.009680 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerName="sg-core" Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.009690 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerName="sg-core" Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.010025 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerName="proxy-httpd" Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.010052 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerName="ceilometer-central-agent" Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.010065 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerName="ceilometer-central-agent" Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.010085 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerName="ceilometer-notification-agent" Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.010103 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerName="sg-core" Dec 06 07:29:18 crc kubenswrapper[4809]: E1206 07:29:18.010387 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerName="ceilometer-central-agent" Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.010407 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerName="ceilometer-central-agent" Dec 06 07:29:18 crc kubenswrapper[4809]: E1206 07:29:18.010423 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerName="ceilometer-central-agent" Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.010430 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerName="ceilometer-central-agent" Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.010725 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerName="ceilometer-central-agent" Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.010752 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d4b42d3-04a4-40e3-803b-11c77024f612" containerName="ceilometer-central-agent" Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.013088 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.015276 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.015711 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.015886 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.023519 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.034102 4809 scope.go:117] "RemoveContainer" containerID="493dde793f5dce6b38eb58b439b1d65c77ff6c4a3277b1a17ad6b0e5a2f1c252" Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.065006 4809 scope.go:117] "RemoveContainer" containerID="10189cd18cf124194d1cd5f8103894926b7f3965e6baefbf9c437f773512728e" Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.158210 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f868ba4b-d293-4711-a358-fcc5f65a43c7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f868ba4b-d293-4711-a358-fcc5f65a43c7\") " pod="openstack/ceilometer-0" Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.158257 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmvmf\" (UniqueName: \"kubernetes.io/projected/f868ba4b-d293-4711-a358-fcc5f65a43c7-kube-api-access-hmvmf\") pod \"ceilometer-0\" (UID: \"f868ba4b-d293-4711-a358-fcc5f65a43c7\") " pod="openstack/ceilometer-0" Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.158328 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f868ba4b-d293-4711-a358-fcc5f65a43c7-run-httpd\") pod \"ceilometer-0\" (UID: \"f868ba4b-d293-4711-a358-fcc5f65a43c7\") " pod="openstack/ceilometer-0" Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.158373 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f868ba4b-d293-4711-a358-fcc5f65a43c7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f868ba4b-d293-4711-a358-fcc5f65a43c7\") " pod="openstack/ceilometer-0" Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.158396 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f868ba4b-d293-4711-a358-fcc5f65a43c7-scripts\") pod \"ceilometer-0\" (UID: \"f868ba4b-d293-4711-a358-fcc5f65a43c7\") " pod="openstack/ceilometer-0" Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.158457 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f868ba4b-d293-4711-a358-fcc5f65a43c7-config-data\") pod \"ceilometer-0\" (UID: \"f868ba4b-d293-4711-a358-fcc5f65a43c7\") " pod="openstack/ceilometer-0" Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.158522 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f868ba4b-d293-4711-a358-fcc5f65a43c7-log-httpd\") pod \"ceilometer-0\" (UID: \"f868ba4b-d293-4711-a358-fcc5f65a43c7\") " pod="openstack/ceilometer-0" Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.158637 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f868ba4b-d293-4711-a358-fcc5f65a43c7-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f868ba4b-d293-4711-a358-fcc5f65a43c7\") " pod="openstack/ceilometer-0" Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.261205 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f868ba4b-d293-4711-a358-fcc5f65a43c7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f868ba4b-d293-4711-a358-fcc5f65a43c7\") " pod="openstack/ceilometer-0" Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.261252 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmvmf\" (UniqueName: \"kubernetes.io/projected/f868ba4b-d293-4711-a358-fcc5f65a43c7-kube-api-access-hmvmf\") pod \"ceilometer-0\" (UID: \"f868ba4b-d293-4711-a358-fcc5f65a43c7\") " pod="openstack/ceilometer-0" Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.261307 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f868ba4b-d293-4711-a358-fcc5f65a43c7-run-httpd\") pod \"ceilometer-0\" (UID: \"f868ba4b-d293-4711-a358-fcc5f65a43c7\") " pod="openstack/ceilometer-0" Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.261339 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f868ba4b-d293-4711-a358-fcc5f65a43c7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f868ba4b-d293-4711-a358-fcc5f65a43c7\") " pod="openstack/ceilometer-0" Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.261364 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f868ba4b-d293-4711-a358-fcc5f65a43c7-scripts\") pod \"ceilometer-0\" (UID: \"f868ba4b-d293-4711-a358-fcc5f65a43c7\") " pod="openstack/ceilometer-0" Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.262036 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f868ba4b-d293-4711-a358-fcc5f65a43c7-run-httpd\") pod \"ceilometer-0\" (UID: \"f868ba4b-d293-4711-a358-fcc5f65a43c7\") " pod="openstack/ceilometer-0" Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.262324 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f868ba4b-d293-4711-a358-fcc5f65a43c7-config-data\") pod \"ceilometer-0\" (UID: \"f868ba4b-d293-4711-a358-fcc5f65a43c7\") " pod="openstack/ceilometer-0" Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.262429 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f868ba4b-d293-4711-a358-fcc5f65a43c7-log-httpd\") pod \"ceilometer-0\" (UID: \"f868ba4b-d293-4711-a358-fcc5f65a43c7\") " pod="openstack/ceilometer-0" Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.262522 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f868ba4b-d293-4711-a358-fcc5f65a43c7-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f868ba4b-d293-4711-a358-fcc5f65a43c7\") " pod="openstack/ceilometer-0" Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.262725 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f868ba4b-d293-4711-a358-fcc5f65a43c7-log-httpd\") pod \"ceilometer-0\" (UID: \"f868ba4b-d293-4711-a358-fcc5f65a43c7\") " pod="openstack/ceilometer-0" Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.266475 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f868ba4b-d293-4711-a358-fcc5f65a43c7-scripts\") pod \"ceilometer-0\" (UID: \"f868ba4b-d293-4711-a358-fcc5f65a43c7\") " pod="openstack/ceilometer-0" Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.267774 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f868ba4b-d293-4711-a358-fcc5f65a43c7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f868ba4b-d293-4711-a358-fcc5f65a43c7\") " pod="openstack/ceilometer-0" Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.268080 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f868ba4b-d293-4711-a358-fcc5f65a43c7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f868ba4b-d293-4711-a358-fcc5f65a43c7\") " pod="openstack/ceilometer-0" Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.268814 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f868ba4b-d293-4711-a358-fcc5f65a43c7-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f868ba4b-d293-4711-a358-fcc5f65a43c7\") " pod="openstack/ceilometer-0" Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.274426 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f868ba4b-d293-4711-a358-fcc5f65a43c7-config-data\") pod \"ceilometer-0\" (UID: \"f868ba4b-d293-4711-a358-fcc5f65a43c7\") " pod="openstack/ceilometer-0" Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.281914 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmvmf\" (UniqueName: \"kubernetes.io/projected/f868ba4b-d293-4711-a358-fcc5f65a43c7-kube-api-access-hmvmf\") pod \"ceilometer-0\" (UID: \"f868ba4b-d293-4711-a358-fcc5f65a43c7\") " pod="openstack/ceilometer-0" Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.335332 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.820682 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:29:18 crc kubenswrapper[4809]: W1206 07:29:18.821664 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf868ba4b_d293_4711_a358_fcc5f65a43c7.slice/crio-93b75310cfdeca03b0bffb7c77139606dcaa52d97c8143fd72accc41d91d7374 WatchSource:0}: Error finding container 93b75310cfdeca03b0bffb7c77139606dcaa52d97c8143fd72accc41d91d7374: Status 404 returned error can't find the container with id 93b75310cfdeca03b0bffb7c77139606dcaa52d97c8143fd72accc41d91d7374 Dec 06 07:29:18 crc kubenswrapper[4809]: I1206 07:29:18.948260 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f868ba4b-d293-4711-a358-fcc5f65a43c7","Type":"ContainerStarted","Data":"93b75310cfdeca03b0bffb7c77139606dcaa52d97c8143fd72accc41d91d7374"} Dec 06 07:29:19 crc kubenswrapper[4809]: I1206 07:29:19.402087 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d4b42d3-04a4-40e3-803b-11c77024f612" path="/var/lib/kubelet/pods/3d4b42d3-04a4-40e3-803b-11c77024f612/volumes" Dec 06 07:29:20 crc kubenswrapper[4809]: I1206 07:29:20.511634 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:29:20 crc kubenswrapper[4809]: I1206 07:29:20.971445 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f868ba4b-d293-4711-a358-fcc5f65a43c7","Type":"ContainerStarted","Data":"c87adb2a97041adac1124cea46a19e5d3e671d8dd6655cffaf2e8c68ccdc7875"} Dec 06 07:29:22 crc kubenswrapper[4809]: I1206 07:29:22.366782 4809 trace.go:236] Trace[596730519]: "Calculate volume metrics of tmpfs for pod openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jjf9r" (06-Dec-2025 07:29:21.223) (total time: 1143ms): Dec 06 07:29:22 crc kubenswrapper[4809]: Trace[596730519]: [1.143451182s] [1.143451182s] END Dec 06 07:29:27 crc kubenswrapper[4809]: I1206 07:29:27.033686 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f868ba4b-d293-4711-a358-fcc5f65a43c7","Type":"ContainerStarted","Data":"a113a401da13312d789f8223882475d9b908dfe12899d2e1fb8f765d1c7d3e67"} Dec 06 07:29:31 crc kubenswrapper[4809]: I1206 07:29:31.085894 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f868ba4b-d293-4711-a358-fcc5f65a43c7","Type":"ContainerStarted","Data":"27b9db1d148e7ae45c884dc8a98752dc92038f48b37deb5a2be0f826db6ee929"} Dec 06 07:29:33 crc kubenswrapper[4809]: I1206 07:29:33.114751 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f868ba4b-d293-4711-a358-fcc5f65a43c7","Type":"ContainerStarted","Data":"8158e56f978214b8752daa3ed7efdcb3d0d4aec4898be5ea663dad82186ccc7e"} Dec 06 07:29:33 crc kubenswrapper[4809]: I1206 07:29:33.115370 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 07:29:33 crc kubenswrapper[4809]: I1206 07:29:33.115147 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f868ba4b-d293-4711-a358-fcc5f65a43c7" containerName="proxy-httpd" containerID="cri-o://8158e56f978214b8752daa3ed7efdcb3d0d4aec4898be5ea663dad82186ccc7e" gracePeriod=30 Dec 06 07:29:33 crc kubenswrapper[4809]: I1206 07:29:33.114894 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f868ba4b-d293-4711-a358-fcc5f65a43c7" containerName="ceilometer-central-agent" containerID="cri-o://c87adb2a97041adac1124cea46a19e5d3e671d8dd6655cffaf2e8c68ccdc7875" gracePeriod=30 Dec 06 07:29:33 crc kubenswrapper[4809]: I1206 07:29:33.115160 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f868ba4b-d293-4711-a358-fcc5f65a43c7" containerName="sg-core" containerID="cri-o://27b9db1d148e7ae45c884dc8a98752dc92038f48b37deb5a2be0f826db6ee929" gracePeriod=30 Dec 06 07:29:33 crc kubenswrapper[4809]: I1206 07:29:33.115171 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f868ba4b-d293-4711-a358-fcc5f65a43c7" containerName="ceilometer-notification-agent" containerID="cri-o://a113a401da13312d789f8223882475d9b908dfe12899d2e1fb8f765d1c7d3e67" gracePeriod=30 Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.113268 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.127513 4809 generic.go:334] "Generic (PLEG): container finished" podID="f868ba4b-d293-4711-a358-fcc5f65a43c7" containerID="8158e56f978214b8752daa3ed7efdcb3d0d4aec4898be5ea663dad82186ccc7e" exitCode=0 Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.127558 4809 generic.go:334] "Generic (PLEG): container finished" podID="f868ba4b-d293-4711-a358-fcc5f65a43c7" containerID="27b9db1d148e7ae45c884dc8a98752dc92038f48b37deb5a2be0f826db6ee929" exitCode=2 Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.127568 4809 generic.go:334] "Generic (PLEG): container finished" podID="f868ba4b-d293-4711-a358-fcc5f65a43c7" containerID="a113a401da13312d789f8223882475d9b908dfe12899d2e1fb8f765d1c7d3e67" exitCode=0 Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.127578 4809 generic.go:334] "Generic (PLEG): container finished" podID="f868ba4b-d293-4711-a358-fcc5f65a43c7" containerID="c87adb2a97041adac1124cea46a19e5d3e671d8dd6655cffaf2e8c68ccdc7875" exitCode=0 Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.127602 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f868ba4b-d293-4711-a358-fcc5f65a43c7","Type":"ContainerDied","Data":"8158e56f978214b8752daa3ed7efdcb3d0d4aec4898be5ea663dad82186ccc7e"} Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.127644 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f868ba4b-d293-4711-a358-fcc5f65a43c7","Type":"ContainerDied","Data":"27b9db1d148e7ae45c884dc8a98752dc92038f48b37deb5a2be0f826db6ee929"} Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.127658 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f868ba4b-d293-4711-a358-fcc5f65a43c7","Type":"ContainerDied","Data":"a113a401da13312d789f8223882475d9b908dfe12899d2e1fb8f765d1c7d3e67"} Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.127672 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f868ba4b-d293-4711-a358-fcc5f65a43c7","Type":"ContainerDied","Data":"c87adb2a97041adac1124cea46a19e5d3e671d8dd6655cffaf2e8c68ccdc7875"} Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.127676 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.127699 4809 scope.go:117] "RemoveContainer" containerID="8158e56f978214b8752daa3ed7efdcb3d0d4aec4898be5ea663dad82186ccc7e" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.127683 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f868ba4b-d293-4711-a358-fcc5f65a43c7","Type":"ContainerDied","Data":"93b75310cfdeca03b0bffb7c77139606dcaa52d97c8143fd72accc41d91d7374"} Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.162530 4809 scope.go:117] "RemoveContainer" containerID="27b9db1d148e7ae45c884dc8a98752dc92038f48b37deb5a2be0f826db6ee929" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.211402 4809 scope.go:117] "RemoveContainer" containerID="a113a401da13312d789f8223882475d9b908dfe12899d2e1fb8f765d1c7d3e67" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.234974 4809 scope.go:117] "RemoveContainer" containerID="c87adb2a97041adac1124cea46a19e5d3e671d8dd6655cffaf2e8c68ccdc7875" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.292139 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f868ba4b-d293-4711-a358-fcc5f65a43c7-config-data\") pod \"f868ba4b-d293-4711-a358-fcc5f65a43c7\" (UID: \"f868ba4b-d293-4711-a358-fcc5f65a43c7\") " Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.292229 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f868ba4b-d293-4711-a358-fcc5f65a43c7-sg-core-conf-yaml\") pod \"f868ba4b-d293-4711-a358-fcc5f65a43c7\" (UID: \"f868ba4b-d293-4711-a358-fcc5f65a43c7\") " Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.292324 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f868ba4b-d293-4711-a358-fcc5f65a43c7-combined-ca-bundle\") pod \"f868ba4b-d293-4711-a358-fcc5f65a43c7\" (UID: \"f868ba4b-d293-4711-a358-fcc5f65a43c7\") " Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.292355 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f868ba4b-d293-4711-a358-fcc5f65a43c7-log-httpd\") pod \"f868ba4b-d293-4711-a358-fcc5f65a43c7\" (UID: \"f868ba4b-d293-4711-a358-fcc5f65a43c7\") " Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.292379 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f868ba4b-d293-4711-a358-fcc5f65a43c7-ceilometer-tls-certs\") pod \"f868ba4b-d293-4711-a358-fcc5f65a43c7\" (UID: \"f868ba4b-d293-4711-a358-fcc5f65a43c7\") " Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.292466 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f868ba4b-d293-4711-a358-fcc5f65a43c7-run-httpd\") pod \"f868ba4b-d293-4711-a358-fcc5f65a43c7\" (UID: \"f868ba4b-d293-4711-a358-fcc5f65a43c7\") " Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.292552 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hmvmf\" (UniqueName: \"kubernetes.io/projected/f868ba4b-d293-4711-a358-fcc5f65a43c7-kube-api-access-hmvmf\") pod \"f868ba4b-d293-4711-a358-fcc5f65a43c7\" (UID: \"f868ba4b-d293-4711-a358-fcc5f65a43c7\") " Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.292633 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f868ba4b-d293-4711-a358-fcc5f65a43c7-scripts\") pod \"f868ba4b-d293-4711-a358-fcc5f65a43c7\" (UID: \"f868ba4b-d293-4711-a358-fcc5f65a43c7\") " Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.293640 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f868ba4b-d293-4711-a358-fcc5f65a43c7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f868ba4b-d293-4711-a358-fcc5f65a43c7" (UID: "f868ba4b-d293-4711-a358-fcc5f65a43c7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.294308 4809 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f868ba4b-d293-4711-a358-fcc5f65a43c7-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.294567 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f868ba4b-d293-4711-a358-fcc5f65a43c7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f868ba4b-d293-4711-a358-fcc5f65a43c7" (UID: "f868ba4b-d293-4711-a358-fcc5f65a43c7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.298845 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f868ba4b-d293-4711-a358-fcc5f65a43c7-kube-api-access-hmvmf" (OuterVolumeSpecName: "kube-api-access-hmvmf") pod "f868ba4b-d293-4711-a358-fcc5f65a43c7" (UID: "f868ba4b-d293-4711-a358-fcc5f65a43c7"). InnerVolumeSpecName "kube-api-access-hmvmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.299132 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f868ba4b-d293-4711-a358-fcc5f65a43c7-scripts" (OuterVolumeSpecName: "scripts") pod "f868ba4b-d293-4711-a358-fcc5f65a43c7" (UID: "f868ba4b-d293-4711-a358-fcc5f65a43c7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.307056 4809 scope.go:117] "RemoveContainer" containerID="8158e56f978214b8752daa3ed7efdcb3d0d4aec4898be5ea663dad82186ccc7e" Dec 06 07:29:34 crc kubenswrapper[4809]: E1206 07:29:34.307499 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8158e56f978214b8752daa3ed7efdcb3d0d4aec4898be5ea663dad82186ccc7e\": container with ID starting with 8158e56f978214b8752daa3ed7efdcb3d0d4aec4898be5ea663dad82186ccc7e not found: ID does not exist" containerID="8158e56f978214b8752daa3ed7efdcb3d0d4aec4898be5ea663dad82186ccc7e" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.307538 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8158e56f978214b8752daa3ed7efdcb3d0d4aec4898be5ea663dad82186ccc7e"} err="failed to get container status \"8158e56f978214b8752daa3ed7efdcb3d0d4aec4898be5ea663dad82186ccc7e\": rpc error: code = NotFound desc = could not find container \"8158e56f978214b8752daa3ed7efdcb3d0d4aec4898be5ea663dad82186ccc7e\": container with ID starting with 8158e56f978214b8752daa3ed7efdcb3d0d4aec4898be5ea663dad82186ccc7e not found: ID does not exist" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.307569 4809 scope.go:117] "RemoveContainer" containerID="27b9db1d148e7ae45c884dc8a98752dc92038f48b37deb5a2be0f826db6ee929" Dec 06 07:29:34 crc kubenswrapper[4809]: E1206 07:29:34.307895 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27b9db1d148e7ae45c884dc8a98752dc92038f48b37deb5a2be0f826db6ee929\": container with ID starting with 27b9db1d148e7ae45c884dc8a98752dc92038f48b37deb5a2be0f826db6ee929 not found: ID does not exist" containerID="27b9db1d148e7ae45c884dc8a98752dc92038f48b37deb5a2be0f826db6ee929" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.308056 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27b9db1d148e7ae45c884dc8a98752dc92038f48b37deb5a2be0f826db6ee929"} err="failed to get container status \"27b9db1d148e7ae45c884dc8a98752dc92038f48b37deb5a2be0f826db6ee929\": rpc error: code = NotFound desc = could not find container \"27b9db1d148e7ae45c884dc8a98752dc92038f48b37deb5a2be0f826db6ee929\": container with ID starting with 27b9db1d148e7ae45c884dc8a98752dc92038f48b37deb5a2be0f826db6ee929 not found: ID does not exist" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.308169 4809 scope.go:117] "RemoveContainer" containerID="a113a401da13312d789f8223882475d9b908dfe12899d2e1fb8f765d1c7d3e67" Dec 06 07:29:34 crc kubenswrapper[4809]: E1206 07:29:34.309541 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a113a401da13312d789f8223882475d9b908dfe12899d2e1fb8f765d1c7d3e67\": container with ID starting with a113a401da13312d789f8223882475d9b908dfe12899d2e1fb8f765d1c7d3e67 not found: ID does not exist" containerID="a113a401da13312d789f8223882475d9b908dfe12899d2e1fb8f765d1c7d3e67" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.309576 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a113a401da13312d789f8223882475d9b908dfe12899d2e1fb8f765d1c7d3e67"} err="failed to get container status \"a113a401da13312d789f8223882475d9b908dfe12899d2e1fb8f765d1c7d3e67\": rpc error: code = NotFound desc = could not find container \"a113a401da13312d789f8223882475d9b908dfe12899d2e1fb8f765d1c7d3e67\": container with ID starting with a113a401da13312d789f8223882475d9b908dfe12899d2e1fb8f765d1c7d3e67 not found: ID does not exist" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.309597 4809 scope.go:117] "RemoveContainer" containerID="c87adb2a97041adac1124cea46a19e5d3e671d8dd6655cffaf2e8c68ccdc7875" Dec 06 07:29:34 crc kubenswrapper[4809]: E1206 07:29:34.309860 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c87adb2a97041adac1124cea46a19e5d3e671d8dd6655cffaf2e8c68ccdc7875\": container with ID starting with c87adb2a97041adac1124cea46a19e5d3e671d8dd6655cffaf2e8c68ccdc7875 not found: ID does not exist" containerID="c87adb2a97041adac1124cea46a19e5d3e671d8dd6655cffaf2e8c68ccdc7875" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.309889 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c87adb2a97041adac1124cea46a19e5d3e671d8dd6655cffaf2e8c68ccdc7875"} err="failed to get container status \"c87adb2a97041adac1124cea46a19e5d3e671d8dd6655cffaf2e8c68ccdc7875\": rpc error: code = NotFound desc = could not find container \"c87adb2a97041adac1124cea46a19e5d3e671d8dd6655cffaf2e8c68ccdc7875\": container with ID starting with c87adb2a97041adac1124cea46a19e5d3e671d8dd6655cffaf2e8c68ccdc7875 not found: ID does not exist" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.309907 4809 scope.go:117] "RemoveContainer" containerID="8158e56f978214b8752daa3ed7efdcb3d0d4aec4898be5ea663dad82186ccc7e" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.310332 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8158e56f978214b8752daa3ed7efdcb3d0d4aec4898be5ea663dad82186ccc7e"} err="failed to get container status \"8158e56f978214b8752daa3ed7efdcb3d0d4aec4898be5ea663dad82186ccc7e\": rpc error: code = NotFound desc = could not find container \"8158e56f978214b8752daa3ed7efdcb3d0d4aec4898be5ea663dad82186ccc7e\": container with ID starting with 8158e56f978214b8752daa3ed7efdcb3d0d4aec4898be5ea663dad82186ccc7e not found: ID does not exist" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.310356 4809 scope.go:117] "RemoveContainer" containerID="27b9db1d148e7ae45c884dc8a98752dc92038f48b37deb5a2be0f826db6ee929" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.310725 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27b9db1d148e7ae45c884dc8a98752dc92038f48b37deb5a2be0f826db6ee929"} err="failed to get container status \"27b9db1d148e7ae45c884dc8a98752dc92038f48b37deb5a2be0f826db6ee929\": rpc error: code = NotFound desc = could not find container \"27b9db1d148e7ae45c884dc8a98752dc92038f48b37deb5a2be0f826db6ee929\": container with ID starting with 27b9db1d148e7ae45c884dc8a98752dc92038f48b37deb5a2be0f826db6ee929 not found: ID does not exist" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.310756 4809 scope.go:117] "RemoveContainer" containerID="a113a401da13312d789f8223882475d9b908dfe12899d2e1fb8f765d1c7d3e67" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.311223 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a113a401da13312d789f8223882475d9b908dfe12899d2e1fb8f765d1c7d3e67"} err="failed to get container status \"a113a401da13312d789f8223882475d9b908dfe12899d2e1fb8f765d1c7d3e67\": rpc error: code = NotFound desc = could not find container \"a113a401da13312d789f8223882475d9b908dfe12899d2e1fb8f765d1c7d3e67\": container with ID starting with a113a401da13312d789f8223882475d9b908dfe12899d2e1fb8f765d1c7d3e67 not found: ID does not exist" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.311298 4809 scope.go:117] "RemoveContainer" containerID="c87adb2a97041adac1124cea46a19e5d3e671d8dd6655cffaf2e8c68ccdc7875" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.311646 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c87adb2a97041adac1124cea46a19e5d3e671d8dd6655cffaf2e8c68ccdc7875"} err="failed to get container status \"c87adb2a97041adac1124cea46a19e5d3e671d8dd6655cffaf2e8c68ccdc7875\": rpc error: code = NotFound desc = could not find container \"c87adb2a97041adac1124cea46a19e5d3e671d8dd6655cffaf2e8c68ccdc7875\": container with ID starting with c87adb2a97041adac1124cea46a19e5d3e671d8dd6655cffaf2e8c68ccdc7875 not found: ID does not exist" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.311674 4809 scope.go:117] "RemoveContainer" containerID="8158e56f978214b8752daa3ed7efdcb3d0d4aec4898be5ea663dad82186ccc7e" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.311974 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8158e56f978214b8752daa3ed7efdcb3d0d4aec4898be5ea663dad82186ccc7e"} err="failed to get container status \"8158e56f978214b8752daa3ed7efdcb3d0d4aec4898be5ea663dad82186ccc7e\": rpc error: code = NotFound desc = could not find container \"8158e56f978214b8752daa3ed7efdcb3d0d4aec4898be5ea663dad82186ccc7e\": container with ID starting with 8158e56f978214b8752daa3ed7efdcb3d0d4aec4898be5ea663dad82186ccc7e not found: ID does not exist" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.312081 4809 scope.go:117] "RemoveContainer" containerID="27b9db1d148e7ae45c884dc8a98752dc92038f48b37deb5a2be0f826db6ee929" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.312392 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27b9db1d148e7ae45c884dc8a98752dc92038f48b37deb5a2be0f826db6ee929"} err="failed to get container status \"27b9db1d148e7ae45c884dc8a98752dc92038f48b37deb5a2be0f826db6ee929\": rpc error: code = NotFound desc = could not find container \"27b9db1d148e7ae45c884dc8a98752dc92038f48b37deb5a2be0f826db6ee929\": container with ID starting with 27b9db1d148e7ae45c884dc8a98752dc92038f48b37deb5a2be0f826db6ee929 not found: ID does not exist" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.312419 4809 scope.go:117] "RemoveContainer" containerID="a113a401da13312d789f8223882475d9b908dfe12899d2e1fb8f765d1c7d3e67" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.312684 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a113a401da13312d789f8223882475d9b908dfe12899d2e1fb8f765d1c7d3e67"} err="failed to get container status \"a113a401da13312d789f8223882475d9b908dfe12899d2e1fb8f765d1c7d3e67\": rpc error: code = NotFound desc = could not find container \"a113a401da13312d789f8223882475d9b908dfe12899d2e1fb8f765d1c7d3e67\": container with ID starting with a113a401da13312d789f8223882475d9b908dfe12899d2e1fb8f765d1c7d3e67 not found: ID does not exist" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.312774 4809 scope.go:117] "RemoveContainer" containerID="c87adb2a97041adac1124cea46a19e5d3e671d8dd6655cffaf2e8c68ccdc7875" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.313168 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c87adb2a97041adac1124cea46a19e5d3e671d8dd6655cffaf2e8c68ccdc7875"} err="failed to get container status \"c87adb2a97041adac1124cea46a19e5d3e671d8dd6655cffaf2e8c68ccdc7875\": rpc error: code = NotFound desc = could not find container \"c87adb2a97041adac1124cea46a19e5d3e671d8dd6655cffaf2e8c68ccdc7875\": container with ID starting with c87adb2a97041adac1124cea46a19e5d3e671d8dd6655cffaf2e8c68ccdc7875 not found: ID does not exist" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.313222 4809 scope.go:117] "RemoveContainer" containerID="8158e56f978214b8752daa3ed7efdcb3d0d4aec4898be5ea663dad82186ccc7e" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.313507 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8158e56f978214b8752daa3ed7efdcb3d0d4aec4898be5ea663dad82186ccc7e"} err="failed to get container status \"8158e56f978214b8752daa3ed7efdcb3d0d4aec4898be5ea663dad82186ccc7e\": rpc error: code = NotFound desc = could not find container \"8158e56f978214b8752daa3ed7efdcb3d0d4aec4898be5ea663dad82186ccc7e\": container with ID starting with 8158e56f978214b8752daa3ed7efdcb3d0d4aec4898be5ea663dad82186ccc7e not found: ID does not exist" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.313540 4809 scope.go:117] "RemoveContainer" containerID="27b9db1d148e7ae45c884dc8a98752dc92038f48b37deb5a2be0f826db6ee929" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.313909 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27b9db1d148e7ae45c884dc8a98752dc92038f48b37deb5a2be0f826db6ee929"} err="failed to get container status \"27b9db1d148e7ae45c884dc8a98752dc92038f48b37deb5a2be0f826db6ee929\": rpc error: code = NotFound desc = could not find container \"27b9db1d148e7ae45c884dc8a98752dc92038f48b37deb5a2be0f826db6ee929\": container with ID starting with 27b9db1d148e7ae45c884dc8a98752dc92038f48b37deb5a2be0f826db6ee929 not found: ID does not exist" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.313955 4809 scope.go:117] "RemoveContainer" containerID="a113a401da13312d789f8223882475d9b908dfe12899d2e1fb8f765d1c7d3e67" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.314248 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a113a401da13312d789f8223882475d9b908dfe12899d2e1fb8f765d1c7d3e67"} err="failed to get container status \"a113a401da13312d789f8223882475d9b908dfe12899d2e1fb8f765d1c7d3e67\": rpc error: code = NotFound desc = could not find container \"a113a401da13312d789f8223882475d9b908dfe12899d2e1fb8f765d1c7d3e67\": container with ID starting with a113a401da13312d789f8223882475d9b908dfe12899d2e1fb8f765d1c7d3e67 not found: ID does not exist" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.314325 4809 scope.go:117] "RemoveContainer" containerID="c87adb2a97041adac1124cea46a19e5d3e671d8dd6655cffaf2e8c68ccdc7875" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.314631 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c87adb2a97041adac1124cea46a19e5d3e671d8dd6655cffaf2e8c68ccdc7875"} err="failed to get container status \"c87adb2a97041adac1124cea46a19e5d3e671d8dd6655cffaf2e8c68ccdc7875\": rpc error: code = NotFound desc = could not find container \"c87adb2a97041adac1124cea46a19e5d3e671d8dd6655cffaf2e8c68ccdc7875\": container with ID starting with c87adb2a97041adac1124cea46a19e5d3e671d8dd6655cffaf2e8c68ccdc7875 not found: ID does not exist" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.341166 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f868ba4b-d293-4711-a358-fcc5f65a43c7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f868ba4b-d293-4711-a358-fcc5f65a43c7" (UID: "f868ba4b-d293-4711-a358-fcc5f65a43c7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.386538 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f868ba4b-d293-4711-a358-fcc5f65a43c7-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "f868ba4b-d293-4711-a358-fcc5f65a43c7" (UID: "f868ba4b-d293-4711-a358-fcc5f65a43c7"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.396742 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hmvmf\" (UniqueName: \"kubernetes.io/projected/f868ba4b-d293-4711-a358-fcc5f65a43c7-kube-api-access-hmvmf\") on node \"crc\" DevicePath \"\"" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.396771 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f868ba4b-d293-4711-a358-fcc5f65a43c7-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.396781 4809 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f868ba4b-d293-4711-a358-fcc5f65a43c7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.396789 4809 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f868ba4b-d293-4711-a358-fcc5f65a43c7-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.396797 4809 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f868ba4b-d293-4711-a358-fcc5f65a43c7-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.424305 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f868ba4b-d293-4711-a358-fcc5f65a43c7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f868ba4b-d293-4711-a358-fcc5f65a43c7" (UID: "f868ba4b-d293-4711-a358-fcc5f65a43c7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.447261 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f868ba4b-d293-4711-a358-fcc5f65a43c7-config-data" (OuterVolumeSpecName: "config-data") pod "f868ba4b-d293-4711-a358-fcc5f65a43c7" (UID: "f868ba4b-d293-4711-a358-fcc5f65a43c7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.495959 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.496251 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.496326 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-npms2" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.497446 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"26a5b8063411f48708112d11ebddb0587d0fe6fc8f6a5c0aa5728fedd5b56f1d"} pod="openshift-machine-config-operator/machine-config-daemon-npms2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.497514 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" containerID="cri-o://26a5b8063411f48708112d11ebddb0587d0fe6fc8f6a5c0aa5728fedd5b56f1d" gracePeriod=600 Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.498509 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f868ba4b-d293-4711-a358-fcc5f65a43c7-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.498530 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f868ba4b-d293-4711-a358-fcc5f65a43c7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.766956 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.782422 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.800878 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:29:34 crc kubenswrapper[4809]: E1206 07:29:34.801505 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f868ba4b-d293-4711-a358-fcc5f65a43c7" containerName="sg-core" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.801528 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f868ba4b-d293-4711-a358-fcc5f65a43c7" containerName="sg-core" Dec 06 07:29:34 crc kubenswrapper[4809]: E1206 07:29:34.801562 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f868ba4b-d293-4711-a358-fcc5f65a43c7" containerName="ceilometer-central-agent" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.801570 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f868ba4b-d293-4711-a358-fcc5f65a43c7" containerName="ceilometer-central-agent" Dec 06 07:29:34 crc kubenswrapper[4809]: E1206 07:29:34.801599 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f868ba4b-d293-4711-a358-fcc5f65a43c7" containerName="ceilometer-notification-agent" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.801605 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f868ba4b-d293-4711-a358-fcc5f65a43c7" containerName="ceilometer-notification-agent" Dec 06 07:29:34 crc kubenswrapper[4809]: E1206 07:29:34.801616 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f868ba4b-d293-4711-a358-fcc5f65a43c7" containerName="proxy-httpd" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.801622 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f868ba4b-d293-4711-a358-fcc5f65a43c7" containerName="proxy-httpd" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.801833 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f868ba4b-d293-4711-a358-fcc5f65a43c7" containerName="proxy-httpd" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.801855 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f868ba4b-d293-4711-a358-fcc5f65a43c7" containerName="ceilometer-notification-agent" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.801868 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f868ba4b-d293-4711-a358-fcc5f65a43c7" containerName="ceilometer-central-agent" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.801885 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f868ba4b-d293-4711-a358-fcc5f65a43c7" containerName="sg-core" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.803918 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.806906 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.807123 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.807373 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.814846 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.906532 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4a36ec8-2831-4f0f-b20c-137fa3d081c8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b4a36ec8-2831-4f0f-b20c-137fa3d081c8\") " pod="openstack/ceilometer-0" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.906739 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4a36ec8-2831-4f0f-b20c-137fa3d081c8-config-data\") pod \"ceilometer-0\" (UID: \"b4a36ec8-2831-4f0f-b20c-137fa3d081c8\") " pod="openstack/ceilometer-0" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.906791 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b4a36ec8-2831-4f0f-b20c-137fa3d081c8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b4a36ec8-2831-4f0f-b20c-137fa3d081c8\") " pod="openstack/ceilometer-0" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.906974 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjslt\" (UniqueName: \"kubernetes.io/projected/b4a36ec8-2831-4f0f-b20c-137fa3d081c8-kube-api-access-sjslt\") pod \"ceilometer-0\" (UID: \"b4a36ec8-2831-4f0f-b20c-137fa3d081c8\") " pod="openstack/ceilometer-0" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.907050 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4a36ec8-2831-4f0f-b20c-137fa3d081c8-run-httpd\") pod \"ceilometer-0\" (UID: \"b4a36ec8-2831-4f0f-b20c-137fa3d081c8\") " pod="openstack/ceilometer-0" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.907129 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4a36ec8-2831-4f0f-b20c-137fa3d081c8-scripts\") pod \"ceilometer-0\" (UID: \"b4a36ec8-2831-4f0f-b20c-137fa3d081c8\") " pod="openstack/ceilometer-0" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.907218 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4a36ec8-2831-4f0f-b20c-137fa3d081c8-log-httpd\") pod \"ceilometer-0\" (UID: \"b4a36ec8-2831-4f0f-b20c-137fa3d081c8\") " pod="openstack/ceilometer-0" Dec 06 07:29:34 crc kubenswrapper[4809]: I1206 07:29:34.907293 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b4a36ec8-2831-4f0f-b20c-137fa3d081c8-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b4a36ec8-2831-4f0f-b20c-137fa3d081c8\") " pod="openstack/ceilometer-0" Dec 06 07:29:35 crc kubenswrapper[4809]: I1206 07:29:35.009665 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4a36ec8-2831-4f0f-b20c-137fa3d081c8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b4a36ec8-2831-4f0f-b20c-137fa3d081c8\") " pod="openstack/ceilometer-0" Dec 06 07:29:35 crc kubenswrapper[4809]: I1206 07:29:35.009771 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4a36ec8-2831-4f0f-b20c-137fa3d081c8-config-data\") pod \"ceilometer-0\" (UID: \"b4a36ec8-2831-4f0f-b20c-137fa3d081c8\") " pod="openstack/ceilometer-0" Dec 06 07:29:35 crc kubenswrapper[4809]: I1206 07:29:35.009797 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b4a36ec8-2831-4f0f-b20c-137fa3d081c8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b4a36ec8-2831-4f0f-b20c-137fa3d081c8\") " pod="openstack/ceilometer-0" Dec 06 07:29:35 crc kubenswrapper[4809]: I1206 07:29:35.009859 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjslt\" (UniqueName: \"kubernetes.io/projected/b4a36ec8-2831-4f0f-b20c-137fa3d081c8-kube-api-access-sjslt\") pod \"ceilometer-0\" (UID: \"b4a36ec8-2831-4f0f-b20c-137fa3d081c8\") " pod="openstack/ceilometer-0" Dec 06 07:29:35 crc kubenswrapper[4809]: I1206 07:29:35.009895 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4a36ec8-2831-4f0f-b20c-137fa3d081c8-run-httpd\") pod \"ceilometer-0\" (UID: \"b4a36ec8-2831-4f0f-b20c-137fa3d081c8\") " pod="openstack/ceilometer-0" Dec 06 07:29:35 crc kubenswrapper[4809]: I1206 07:29:35.009971 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4a36ec8-2831-4f0f-b20c-137fa3d081c8-scripts\") pod \"ceilometer-0\" (UID: \"b4a36ec8-2831-4f0f-b20c-137fa3d081c8\") " pod="openstack/ceilometer-0" Dec 06 07:29:35 crc kubenswrapper[4809]: I1206 07:29:35.010022 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4a36ec8-2831-4f0f-b20c-137fa3d081c8-log-httpd\") pod \"ceilometer-0\" (UID: \"b4a36ec8-2831-4f0f-b20c-137fa3d081c8\") " pod="openstack/ceilometer-0" Dec 06 07:29:35 crc kubenswrapper[4809]: I1206 07:29:35.010075 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b4a36ec8-2831-4f0f-b20c-137fa3d081c8-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b4a36ec8-2831-4f0f-b20c-137fa3d081c8\") " pod="openstack/ceilometer-0" Dec 06 07:29:35 crc kubenswrapper[4809]: I1206 07:29:35.010872 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4a36ec8-2831-4f0f-b20c-137fa3d081c8-run-httpd\") pod \"ceilometer-0\" (UID: \"b4a36ec8-2831-4f0f-b20c-137fa3d081c8\") " pod="openstack/ceilometer-0" Dec 06 07:29:35 crc kubenswrapper[4809]: I1206 07:29:35.010948 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4a36ec8-2831-4f0f-b20c-137fa3d081c8-log-httpd\") pod \"ceilometer-0\" (UID: \"b4a36ec8-2831-4f0f-b20c-137fa3d081c8\") " pod="openstack/ceilometer-0" Dec 06 07:29:35 crc kubenswrapper[4809]: I1206 07:29:35.015187 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4a36ec8-2831-4f0f-b20c-137fa3d081c8-config-data\") pod \"ceilometer-0\" (UID: \"b4a36ec8-2831-4f0f-b20c-137fa3d081c8\") " pod="openstack/ceilometer-0" Dec 06 07:29:35 crc kubenswrapper[4809]: I1206 07:29:35.015667 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4a36ec8-2831-4f0f-b20c-137fa3d081c8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b4a36ec8-2831-4f0f-b20c-137fa3d081c8\") " pod="openstack/ceilometer-0" Dec 06 07:29:35 crc kubenswrapper[4809]: I1206 07:29:35.016084 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b4a36ec8-2831-4f0f-b20c-137fa3d081c8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b4a36ec8-2831-4f0f-b20c-137fa3d081c8\") " pod="openstack/ceilometer-0" Dec 06 07:29:35 crc kubenswrapper[4809]: I1206 07:29:35.016385 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b4a36ec8-2831-4f0f-b20c-137fa3d081c8-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b4a36ec8-2831-4f0f-b20c-137fa3d081c8\") " pod="openstack/ceilometer-0" Dec 06 07:29:35 crc kubenswrapper[4809]: I1206 07:29:35.017405 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4a36ec8-2831-4f0f-b20c-137fa3d081c8-scripts\") pod \"ceilometer-0\" (UID: \"b4a36ec8-2831-4f0f-b20c-137fa3d081c8\") " pod="openstack/ceilometer-0" Dec 06 07:29:35 crc kubenswrapper[4809]: I1206 07:29:35.029463 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjslt\" (UniqueName: \"kubernetes.io/projected/b4a36ec8-2831-4f0f-b20c-137fa3d081c8-kube-api-access-sjslt\") pod \"ceilometer-0\" (UID: \"b4a36ec8-2831-4f0f-b20c-137fa3d081c8\") " pod="openstack/ceilometer-0" Dec 06 07:29:35 crc kubenswrapper[4809]: I1206 07:29:35.131970 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 07:29:35 crc kubenswrapper[4809]: I1206 07:29:35.140957 4809 generic.go:334] "Generic (PLEG): container finished" podID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerID="26a5b8063411f48708112d11ebddb0587d0fe6fc8f6a5c0aa5728fedd5b56f1d" exitCode=0 Dec 06 07:29:35 crc kubenswrapper[4809]: I1206 07:29:35.141044 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" event={"ID":"cbf4cf62-024e-4703-a8b8-9aecda9cd26a","Type":"ContainerDied","Data":"26a5b8063411f48708112d11ebddb0587d0fe6fc8f6a5c0aa5728fedd5b56f1d"} Dec 06 07:29:35 crc kubenswrapper[4809]: I1206 07:29:35.141082 4809 scope.go:117] "RemoveContainer" containerID="8685c71ca1e6bc659b1168658926b473eb3d43e2ebbafcbe1ef610af7804649e" Dec 06 07:29:35 crc kubenswrapper[4809]: E1206 07:29:35.144760 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:29:35 crc kubenswrapper[4809]: I1206 07:29:35.407740 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f868ba4b-d293-4711-a358-fcc5f65a43c7" path="/var/lib/kubelet/pods/f868ba4b-d293-4711-a358-fcc5f65a43c7/volumes" Dec 06 07:29:35 crc kubenswrapper[4809]: I1206 07:29:35.638291 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 07:29:36 crc kubenswrapper[4809]: I1206 07:29:36.159817 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4a36ec8-2831-4f0f-b20c-137fa3d081c8","Type":"ContainerStarted","Data":"30d4679c7b92ed772232e5b44117c431a50b61f3a69b061ad784c25baa4ff03e"} Dec 06 07:29:36 crc kubenswrapper[4809]: I1206 07:29:36.163717 4809 scope.go:117] "RemoveContainer" containerID="26a5b8063411f48708112d11ebddb0587d0fe6fc8f6a5c0aa5728fedd5b56f1d" Dec 06 07:29:36 crc kubenswrapper[4809]: E1206 07:29:36.164167 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:29:37 crc kubenswrapper[4809]: I1206 07:29:37.192975 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4a36ec8-2831-4f0f-b20c-137fa3d081c8","Type":"ContainerStarted","Data":"3213f686080d1ed7a3e4561fe3c1d92a003b1ea7667aeb39de7c8e73167d7f80"} Dec 06 07:29:38 crc kubenswrapper[4809]: I1206 07:29:38.205262 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4a36ec8-2831-4f0f-b20c-137fa3d081c8","Type":"ContainerStarted","Data":"a37b97fc5fd262c9d9697b819cbf3d23471f4f24344220064bdc94ed9eec37c3"} Dec 06 07:29:39 crc kubenswrapper[4809]: I1206 07:29:39.219495 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4a36ec8-2831-4f0f-b20c-137fa3d081c8","Type":"ContainerStarted","Data":"67340c2dd540b237b7bfd9f4ebc21eb8cceb11039380aac2e6900ded78db70a2"} Dec 06 07:29:40 crc kubenswrapper[4809]: I1206 07:29:40.234165 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4a36ec8-2831-4f0f-b20c-137fa3d081c8","Type":"ContainerStarted","Data":"2a21d75df54a5a36a2dc4dd0d33f11c33f9d330fe0a094b1c3e4631bd6e43c21"} Dec 06 07:29:40 crc kubenswrapper[4809]: I1206 07:29:40.234732 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 07:29:44 crc kubenswrapper[4809]: I1206 07:29:44.972705 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=6.9273874079999995 podStartE2EDuration="10.972684397s" podCreationTimestamp="2025-12-06 07:29:34 +0000 UTC" firstStartedPulling="2025-12-06 07:29:35.653382382 +0000 UTC m=+5900.542365334" lastFinishedPulling="2025-12-06 07:29:39.698679381 +0000 UTC m=+5904.587662323" observedRunningTime="2025-12-06 07:29:40.511508638 +0000 UTC m=+5905.400491580" watchObservedRunningTime="2025-12-06 07:29:44.972684397 +0000 UTC m=+5909.861667339" Dec 06 07:29:44 crc kubenswrapper[4809]: I1206 07:29:44.985798 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-bn94p/must-gather-t7bfm"] Dec 06 07:29:44 crc kubenswrapper[4809]: I1206 07:29:44.987820 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bn94p/must-gather-t7bfm" Dec 06 07:29:44 crc kubenswrapper[4809]: I1206 07:29:44.996416 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-bn94p"/"openshift-service-ca.crt" Dec 06 07:29:44 crc kubenswrapper[4809]: I1206 07:29:44.997903 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-bn94p"/"kube-root-ca.crt" Dec 06 07:29:45 crc kubenswrapper[4809]: I1206 07:29:45.009901 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-bn94p"/"default-dockercfg-vjkqs" Dec 06 07:29:45 crc kubenswrapper[4809]: I1206 07:29:45.018504 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-bn94p/must-gather-t7bfm"] Dec 06 07:29:45 crc kubenswrapper[4809]: I1206 07:29:45.122035 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rghj2\" (UniqueName: \"kubernetes.io/projected/3603c2a8-988f-4ecf-ad1b-51b74176a0a1-kube-api-access-rghj2\") pod \"must-gather-t7bfm\" (UID: \"3603c2a8-988f-4ecf-ad1b-51b74176a0a1\") " pod="openshift-must-gather-bn94p/must-gather-t7bfm" Dec 06 07:29:45 crc kubenswrapper[4809]: I1206 07:29:45.122158 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3603c2a8-988f-4ecf-ad1b-51b74176a0a1-must-gather-output\") pod \"must-gather-t7bfm\" (UID: \"3603c2a8-988f-4ecf-ad1b-51b74176a0a1\") " pod="openshift-must-gather-bn94p/must-gather-t7bfm" Dec 06 07:29:45 crc kubenswrapper[4809]: I1206 07:29:45.224493 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rghj2\" (UniqueName: \"kubernetes.io/projected/3603c2a8-988f-4ecf-ad1b-51b74176a0a1-kube-api-access-rghj2\") pod \"must-gather-t7bfm\" (UID: \"3603c2a8-988f-4ecf-ad1b-51b74176a0a1\") " pod="openshift-must-gather-bn94p/must-gather-t7bfm" Dec 06 07:29:45 crc kubenswrapper[4809]: I1206 07:29:45.224595 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3603c2a8-988f-4ecf-ad1b-51b74176a0a1-must-gather-output\") pod \"must-gather-t7bfm\" (UID: \"3603c2a8-988f-4ecf-ad1b-51b74176a0a1\") " pod="openshift-must-gather-bn94p/must-gather-t7bfm" Dec 06 07:29:45 crc kubenswrapper[4809]: I1206 07:29:45.225258 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3603c2a8-988f-4ecf-ad1b-51b74176a0a1-must-gather-output\") pod \"must-gather-t7bfm\" (UID: \"3603c2a8-988f-4ecf-ad1b-51b74176a0a1\") " pod="openshift-must-gather-bn94p/must-gather-t7bfm" Dec 06 07:29:45 crc kubenswrapper[4809]: I1206 07:29:45.255683 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rghj2\" (UniqueName: \"kubernetes.io/projected/3603c2a8-988f-4ecf-ad1b-51b74176a0a1-kube-api-access-rghj2\") pod \"must-gather-t7bfm\" (UID: \"3603c2a8-988f-4ecf-ad1b-51b74176a0a1\") " pod="openshift-must-gather-bn94p/must-gather-t7bfm" Dec 06 07:29:45 crc kubenswrapper[4809]: I1206 07:29:45.308879 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bn94p/must-gather-t7bfm" Dec 06 07:29:46 crc kubenswrapper[4809]: I1206 07:29:46.106080 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-bn94p/must-gather-t7bfm"] Dec 06 07:29:46 crc kubenswrapper[4809]: I1206 07:29:46.313124 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bn94p/must-gather-t7bfm" event={"ID":"3603c2a8-988f-4ecf-ad1b-51b74176a0a1","Type":"ContainerStarted","Data":"fc43c7fe6b4479d7614cfec271f81bcb6f15649c5c75caf678c52ccbeb7f3cfb"} Dec 06 07:29:51 crc kubenswrapper[4809]: I1206 07:29:51.389616 4809 scope.go:117] "RemoveContainer" containerID="26a5b8063411f48708112d11ebddb0587d0fe6fc8f6a5c0aa5728fedd5b56f1d" Dec 06 07:29:51 crc kubenswrapper[4809]: E1206 07:29:51.390684 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:29:52 crc kubenswrapper[4809]: I1206 07:29:52.571291 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bn94p/must-gather-t7bfm" event={"ID":"3603c2a8-988f-4ecf-ad1b-51b74176a0a1","Type":"ContainerStarted","Data":"e9785bbc5d7c66af1d9400512f3e1cf9163eba7b5c7fffd5163e48af26640ad4"} Dec 06 07:29:53 crc kubenswrapper[4809]: I1206 07:29:53.617770 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bn94p/must-gather-t7bfm" event={"ID":"3603c2a8-988f-4ecf-ad1b-51b74176a0a1","Type":"ContainerStarted","Data":"58bf556966fa07e4d904e4d1811fec480bf1a6508c911e97b07a41f00dc004e0"} Dec 06 07:29:53 crc kubenswrapper[4809]: I1206 07:29:53.638176 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-bn94p/must-gather-t7bfm" podStartSLOduration=3.672143074 podStartE2EDuration="9.638155902s" podCreationTimestamp="2025-12-06 07:29:44 +0000 UTC" firstStartedPulling="2025-12-06 07:29:46.09725077 +0000 UTC m=+5910.986233722" lastFinishedPulling="2025-12-06 07:29:52.063263608 +0000 UTC m=+5916.952246550" observedRunningTime="2025-12-06 07:29:53.631552073 +0000 UTC m=+5918.520535045" watchObservedRunningTime="2025-12-06 07:29:53.638155902 +0000 UTC m=+5918.527138834" Dec 06 07:29:58 crc kubenswrapper[4809]: I1206 07:29:58.908835 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-bn94p/crc-debug-dkcmj"] Dec 06 07:29:58 crc kubenswrapper[4809]: I1206 07:29:58.912386 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bn94p/crc-debug-dkcmj" Dec 06 07:29:58 crc kubenswrapper[4809]: I1206 07:29:58.980556 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/581ac32b-a24f-4def-aab5-a2123f5674b3-host\") pod \"crc-debug-dkcmj\" (UID: \"581ac32b-a24f-4def-aab5-a2123f5674b3\") " pod="openshift-must-gather-bn94p/crc-debug-dkcmj" Dec 06 07:29:58 crc kubenswrapper[4809]: I1206 07:29:58.980615 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkwt5\" (UniqueName: \"kubernetes.io/projected/581ac32b-a24f-4def-aab5-a2123f5674b3-kube-api-access-xkwt5\") pod \"crc-debug-dkcmj\" (UID: \"581ac32b-a24f-4def-aab5-a2123f5674b3\") " pod="openshift-must-gather-bn94p/crc-debug-dkcmj" Dec 06 07:29:59 crc kubenswrapper[4809]: I1206 07:29:59.082634 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/581ac32b-a24f-4def-aab5-a2123f5674b3-host\") pod \"crc-debug-dkcmj\" (UID: \"581ac32b-a24f-4def-aab5-a2123f5674b3\") " pod="openshift-must-gather-bn94p/crc-debug-dkcmj" Dec 06 07:29:59 crc kubenswrapper[4809]: I1206 07:29:59.082902 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkwt5\" (UniqueName: \"kubernetes.io/projected/581ac32b-a24f-4def-aab5-a2123f5674b3-kube-api-access-xkwt5\") pod \"crc-debug-dkcmj\" (UID: \"581ac32b-a24f-4def-aab5-a2123f5674b3\") " pod="openshift-must-gather-bn94p/crc-debug-dkcmj" Dec 06 07:29:59 crc kubenswrapper[4809]: I1206 07:29:59.082800 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/581ac32b-a24f-4def-aab5-a2123f5674b3-host\") pod \"crc-debug-dkcmj\" (UID: \"581ac32b-a24f-4def-aab5-a2123f5674b3\") " pod="openshift-must-gather-bn94p/crc-debug-dkcmj" Dec 06 07:29:59 crc kubenswrapper[4809]: I1206 07:29:59.103634 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkwt5\" (UniqueName: \"kubernetes.io/projected/581ac32b-a24f-4def-aab5-a2123f5674b3-kube-api-access-xkwt5\") pod \"crc-debug-dkcmj\" (UID: \"581ac32b-a24f-4def-aab5-a2123f5674b3\") " pod="openshift-must-gather-bn94p/crc-debug-dkcmj" Dec 06 07:29:59 crc kubenswrapper[4809]: I1206 07:29:59.235920 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bn94p/crc-debug-dkcmj" Dec 06 07:29:59 crc kubenswrapper[4809]: I1206 07:29:59.685177 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bn94p/crc-debug-dkcmj" event={"ID":"581ac32b-a24f-4def-aab5-a2123f5674b3","Type":"ContainerStarted","Data":"559bd0d3684e3875b4b4a2a4159b0d2ce7a3f8355b9334821ea5586c13faaa37"} Dec 06 07:30:00 crc kubenswrapper[4809]: I1206 07:30:00.174401 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416770-jbsv4"] Dec 06 07:30:00 crc kubenswrapper[4809]: I1206 07:30:00.176153 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416770-jbsv4" Dec 06 07:30:00 crc kubenswrapper[4809]: I1206 07:30:00.178175 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 07:30:00 crc kubenswrapper[4809]: I1206 07:30:00.178546 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 07:30:00 crc kubenswrapper[4809]: I1206 07:30:00.187143 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416770-jbsv4"] Dec 06 07:30:00 crc kubenswrapper[4809]: I1206 07:30:00.215476 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/034aef25-c1f0-487a-80c3-2ba30229cc13-config-volume\") pod \"collect-profiles-29416770-jbsv4\" (UID: \"034aef25-c1f0-487a-80c3-2ba30229cc13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416770-jbsv4" Dec 06 07:30:00 crc kubenswrapper[4809]: I1206 07:30:00.215713 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwlxx\" (UniqueName: \"kubernetes.io/projected/034aef25-c1f0-487a-80c3-2ba30229cc13-kube-api-access-fwlxx\") pod \"collect-profiles-29416770-jbsv4\" (UID: \"034aef25-c1f0-487a-80c3-2ba30229cc13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416770-jbsv4" Dec 06 07:30:00 crc kubenswrapper[4809]: I1206 07:30:00.215796 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/034aef25-c1f0-487a-80c3-2ba30229cc13-secret-volume\") pod \"collect-profiles-29416770-jbsv4\" (UID: \"034aef25-c1f0-487a-80c3-2ba30229cc13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416770-jbsv4" Dec 06 07:30:00 crc kubenswrapper[4809]: I1206 07:30:00.318054 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/034aef25-c1f0-487a-80c3-2ba30229cc13-config-volume\") pod \"collect-profiles-29416770-jbsv4\" (UID: \"034aef25-c1f0-487a-80c3-2ba30229cc13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416770-jbsv4" Dec 06 07:30:00 crc kubenswrapper[4809]: I1206 07:30:00.318417 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwlxx\" (UniqueName: \"kubernetes.io/projected/034aef25-c1f0-487a-80c3-2ba30229cc13-kube-api-access-fwlxx\") pod \"collect-profiles-29416770-jbsv4\" (UID: \"034aef25-c1f0-487a-80c3-2ba30229cc13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416770-jbsv4" Dec 06 07:30:00 crc kubenswrapper[4809]: I1206 07:30:00.318539 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/034aef25-c1f0-487a-80c3-2ba30229cc13-secret-volume\") pod \"collect-profiles-29416770-jbsv4\" (UID: \"034aef25-c1f0-487a-80c3-2ba30229cc13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416770-jbsv4" Dec 06 07:30:00 crc kubenswrapper[4809]: I1206 07:30:00.319246 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/034aef25-c1f0-487a-80c3-2ba30229cc13-config-volume\") pod \"collect-profiles-29416770-jbsv4\" (UID: \"034aef25-c1f0-487a-80c3-2ba30229cc13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416770-jbsv4" Dec 06 07:30:00 crc kubenswrapper[4809]: I1206 07:30:00.329587 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/034aef25-c1f0-487a-80c3-2ba30229cc13-secret-volume\") pod \"collect-profiles-29416770-jbsv4\" (UID: \"034aef25-c1f0-487a-80c3-2ba30229cc13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416770-jbsv4" Dec 06 07:30:00 crc kubenswrapper[4809]: I1206 07:30:00.338198 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwlxx\" (UniqueName: \"kubernetes.io/projected/034aef25-c1f0-487a-80c3-2ba30229cc13-kube-api-access-fwlxx\") pod \"collect-profiles-29416770-jbsv4\" (UID: \"034aef25-c1f0-487a-80c3-2ba30229cc13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416770-jbsv4" Dec 06 07:30:00 crc kubenswrapper[4809]: I1206 07:30:00.508698 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416770-jbsv4" Dec 06 07:30:01 crc kubenswrapper[4809]: I1206 07:30:01.147408 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416770-jbsv4"] Dec 06 07:30:01 crc kubenswrapper[4809]: W1206 07:30:01.169793 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod034aef25_c1f0_487a_80c3_2ba30229cc13.slice/crio-2ed0a8f25a1817bc881c32051d0d68b38208f71cb3603d18ffb5bd5329ba0bf9 WatchSource:0}: Error finding container 2ed0a8f25a1817bc881c32051d0d68b38208f71cb3603d18ffb5bd5329ba0bf9: Status 404 returned error can't find the container with id 2ed0a8f25a1817bc881c32051d0d68b38208f71cb3603d18ffb5bd5329ba0bf9 Dec 06 07:30:01 crc kubenswrapper[4809]: I1206 07:30:01.722263 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416770-jbsv4" event={"ID":"034aef25-c1f0-487a-80c3-2ba30229cc13","Type":"ContainerStarted","Data":"a0078c7f19da9e27a90437721971da5aed7b8a297844ad169b786711f1ceb0ab"} Dec 06 07:30:01 crc kubenswrapper[4809]: I1206 07:30:01.722648 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416770-jbsv4" event={"ID":"034aef25-c1f0-487a-80c3-2ba30229cc13","Type":"ContainerStarted","Data":"2ed0a8f25a1817bc881c32051d0d68b38208f71cb3603d18ffb5bd5329ba0bf9"} Dec 06 07:30:01 crc kubenswrapper[4809]: I1206 07:30:01.741459 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29416770-jbsv4" podStartSLOduration=1.741435922 podStartE2EDuration="1.741435922s" podCreationTimestamp="2025-12-06 07:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:30:01.736084417 +0000 UTC m=+5926.625067359" watchObservedRunningTime="2025-12-06 07:30:01.741435922 +0000 UTC m=+5926.630418864" Dec 06 07:30:02 crc kubenswrapper[4809]: I1206 07:30:02.389305 4809 scope.go:117] "RemoveContainer" containerID="26a5b8063411f48708112d11ebddb0587d0fe6fc8f6a5c0aa5728fedd5b56f1d" Dec 06 07:30:02 crc kubenswrapper[4809]: E1206 07:30:02.389847 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:30:04 crc kubenswrapper[4809]: I1206 07:30:04.765092 4809 generic.go:334] "Generic (PLEG): container finished" podID="034aef25-c1f0-487a-80c3-2ba30229cc13" containerID="a0078c7f19da9e27a90437721971da5aed7b8a297844ad169b786711f1ceb0ab" exitCode=0 Dec 06 07:30:04 crc kubenswrapper[4809]: I1206 07:30:04.765159 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416770-jbsv4" event={"ID":"034aef25-c1f0-487a-80c3-2ba30229cc13","Type":"ContainerDied","Data":"a0078c7f19da9e27a90437721971da5aed7b8a297844ad169b786711f1ceb0ab"} Dec 06 07:30:05 crc kubenswrapper[4809]: I1206 07:30:05.235556 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 06 07:30:09 crc kubenswrapper[4809]: I1206 07:30:09.278874 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-72dcl"] Dec 06 07:30:09 crc kubenswrapper[4809]: I1206 07:30:09.282594 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-72dcl" Dec 06 07:30:09 crc kubenswrapper[4809]: I1206 07:30:09.298208 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-72dcl"] Dec 06 07:30:09 crc kubenswrapper[4809]: I1206 07:30:09.373949 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8e767ec-8d4b-4a29-8562-115db074c7a7-utilities\") pod \"community-operators-72dcl\" (UID: \"f8e767ec-8d4b-4a29-8562-115db074c7a7\") " pod="openshift-marketplace/community-operators-72dcl" Dec 06 07:30:09 crc kubenswrapper[4809]: I1206 07:30:09.374000 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8e767ec-8d4b-4a29-8562-115db074c7a7-catalog-content\") pod \"community-operators-72dcl\" (UID: \"f8e767ec-8d4b-4a29-8562-115db074c7a7\") " pod="openshift-marketplace/community-operators-72dcl" Dec 06 07:30:09 crc kubenswrapper[4809]: I1206 07:30:09.374155 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5f6vf\" (UniqueName: \"kubernetes.io/projected/f8e767ec-8d4b-4a29-8562-115db074c7a7-kube-api-access-5f6vf\") pod \"community-operators-72dcl\" (UID: \"f8e767ec-8d4b-4a29-8562-115db074c7a7\") " pod="openshift-marketplace/community-operators-72dcl" Dec 06 07:30:09 crc kubenswrapper[4809]: I1206 07:30:09.476005 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8e767ec-8d4b-4a29-8562-115db074c7a7-utilities\") pod \"community-operators-72dcl\" (UID: \"f8e767ec-8d4b-4a29-8562-115db074c7a7\") " pod="openshift-marketplace/community-operators-72dcl" Dec 06 07:30:09 crc kubenswrapper[4809]: I1206 07:30:09.476053 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8e767ec-8d4b-4a29-8562-115db074c7a7-catalog-content\") pod \"community-operators-72dcl\" (UID: \"f8e767ec-8d4b-4a29-8562-115db074c7a7\") " pod="openshift-marketplace/community-operators-72dcl" Dec 06 07:30:09 crc kubenswrapper[4809]: I1206 07:30:09.476165 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5f6vf\" (UniqueName: \"kubernetes.io/projected/f8e767ec-8d4b-4a29-8562-115db074c7a7-kube-api-access-5f6vf\") pod \"community-operators-72dcl\" (UID: \"f8e767ec-8d4b-4a29-8562-115db074c7a7\") " pod="openshift-marketplace/community-operators-72dcl" Dec 06 07:30:09 crc kubenswrapper[4809]: I1206 07:30:09.477110 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8e767ec-8d4b-4a29-8562-115db074c7a7-utilities\") pod \"community-operators-72dcl\" (UID: \"f8e767ec-8d4b-4a29-8562-115db074c7a7\") " pod="openshift-marketplace/community-operators-72dcl" Dec 06 07:30:09 crc kubenswrapper[4809]: I1206 07:30:09.480274 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8e767ec-8d4b-4a29-8562-115db074c7a7-catalog-content\") pod \"community-operators-72dcl\" (UID: \"f8e767ec-8d4b-4a29-8562-115db074c7a7\") " pod="openshift-marketplace/community-operators-72dcl" Dec 06 07:30:09 crc kubenswrapper[4809]: I1206 07:30:09.502161 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5f6vf\" (UniqueName: \"kubernetes.io/projected/f8e767ec-8d4b-4a29-8562-115db074c7a7-kube-api-access-5f6vf\") pod \"community-operators-72dcl\" (UID: \"f8e767ec-8d4b-4a29-8562-115db074c7a7\") " pod="openshift-marketplace/community-operators-72dcl" Dec 06 07:30:09 crc kubenswrapper[4809]: I1206 07:30:09.611989 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-72dcl" Dec 06 07:30:13 crc kubenswrapper[4809]: I1206 07:30:13.394176 4809 scope.go:117] "RemoveContainer" containerID="26a5b8063411f48708112d11ebddb0587d0fe6fc8f6a5c0aa5728fedd5b56f1d" Dec 06 07:30:13 crc kubenswrapper[4809]: E1206 07:30:13.395102 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:30:15 crc kubenswrapper[4809]: E1206 07:30:15.486363 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296" Dec 06 07:30:15 crc kubenswrapper[4809]: E1206 07:30:15.488463 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:container-00,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296,Command:[chroot /host bash -c echo 'TOOLBOX_NAME=toolbox-osp' > /root/.toolboxrc ; rm -rf \"/var/tmp/sos-osp\" && mkdir -p \"/var/tmp/sos-osp\" && sudo podman rm --force toolbox-osp; sudo --preserve-env podman pull --authfile /var/lib/kubelet/config.json registry.redhat.io/rhel9/support-tools && toolbox sos report --batch --all-logs --only-plugins block,cifs,crio,devicemapper,devices,firewall_tables,firewalld,iscsi,lvm2,memory,multipath,nfs,nis,nvme,podman,process,processor,selinux,scsi,udev,logs,crypto --tmp-dir=\"/var/tmp/sos-osp\" && if [[ \"$(ls /var/log/pods/*/{*.log.*,*/*.log.*} 2>/dev/null)\" != '' ]]; then tar --ignore-failed-read --warning=no-file-changed -cJf \"/var/tmp/sos-osp/podlogs.tar.xz\" --transform 's,^,podlogs/,' /var/log/pods/*/{*.log.*,*/*.log.*} || true; fi],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:TMOUT,Value:900,ValueFrom:nil,},EnvVar{Name:HOST,Value:/host,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host,ReadOnly:false,MountPath:/host,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xkwt5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod crc-debug-dkcmj_openshift-must-gather-bn94p(581ac32b-a24f-4def-aab5-a2123f5674b3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 07:30:15 crc kubenswrapper[4809]: E1206 07:30:15.491435 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"container-00\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openshift-must-gather-bn94p/crc-debug-dkcmj" podUID="581ac32b-a24f-4def-aab5-a2123f5674b3" Dec 06 07:30:15 crc kubenswrapper[4809]: I1206 07:30:15.576772 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416770-jbsv4" Dec 06 07:30:15 crc kubenswrapper[4809]: I1206 07:30:15.632441 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/034aef25-c1f0-487a-80c3-2ba30229cc13-config-volume\") pod \"034aef25-c1f0-487a-80c3-2ba30229cc13\" (UID: \"034aef25-c1f0-487a-80c3-2ba30229cc13\") " Dec 06 07:30:15 crc kubenswrapper[4809]: I1206 07:30:15.632705 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwlxx\" (UniqueName: \"kubernetes.io/projected/034aef25-c1f0-487a-80c3-2ba30229cc13-kube-api-access-fwlxx\") pod \"034aef25-c1f0-487a-80c3-2ba30229cc13\" (UID: \"034aef25-c1f0-487a-80c3-2ba30229cc13\") " Dec 06 07:30:15 crc kubenswrapper[4809]: I1206 07:30:15.632763 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/034aef25-c1f0-487a-80c3-2ba30229cc13-secret-volume\") pod \"034aef25-c1f0-487a-80c3-2ba30229cc13\" (UID: \"034aef25-c1f0-487a-80c3-2ba30229cc13\") " Dec 06 07:30:15 crc kubenswrapper[4809]: I1206 07:30:15.633256 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/034aef25-c1f0-487a-80c3-2ba30229cc13-config-volume" (OuterVolumeSpecName: "config-volume") pod "034aef25-c1f0-487a-80c3-2ba30229cc13" (UID: "034aef25-c1f0-487a-80c3-2ba30229cc13"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:30:15 crc kubenswrapper[4809]: I1206 07:30:15.633746 4809 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/034aef25-c1f0-487a-80c3-2ba30229cc13-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 07:30:15 crc kubenswrapper[4809]: I1206 07:30:15.640569 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/034aef25-c1f0-487a-80c3-2ba30229cc13-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "034aef25-c1f0-487a-80c3-2ba30229cc13" (UID: "034aef25-c1f0-487a-80c3-2ba30229cc13"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:30:15 crc kubenswrapper[4809]: I1206 07:30:15.640631 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/034aef25-c1f0-487a-80c3-2ba30229cc13-kube-api-access-fwlxx" (OuterVolumeSpecName: "kube-api-access-fwlxx") pod "034aef25-c1f0-487a-80c3-2ba30229cc13" (UID: "034aef25-c1f0-487a-80c3-2ba30229cc13"). InnerVolumeSpecName "kube-api-access-fwlxx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:30:15 crc kubenswrapper[4809]: I1206 07:30:15.736539 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwlxx\" (UniqueName: \"kubernetes.io/projected/034aef25-c1f0-487a-80c3-2ba30229cc13-kube-api-access-fwlxx\") on node \"crc\" DevicePath \"\"" Dec 06 07:30:15 crc kubenswrapper[4809]: I1206 07:30:15.736565 4809 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/034aef25-c1f0-487a-80c3-2ba30229cc13-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 07:30:15 crc kubenswrapper[4809]: I1206 07:30:15.957718 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416770-jbsv4" event={"ID":"034aef25-c1f0-487a-80c3-2ba30229cc13","Type":"ContainerDied","Data":"2ed0a8f25a1817bc881c32051d0d68b38208f71cb3603d18ffb5bd5329ba0bf9"} Dec 06 07:30:15 crc kubenswrapper[4809]: I1206 07:30:15.957986 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2ed0a8f25a1817bc881c32051d0d68b38208f71cb3603d18ffb5bd5329ba0bf9" Dec 06 07:30:15 crc kubenswrapper[4809]: I1206 07:30:15.958143 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416770-jbsv4" Dec 06 07:30:15 crc kubenswrapper[4809]: E1206 07:30:15.960539 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"container-00\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296\\\"\"" pod="openshift-must-gather-bn94p/crc-debug-dkcmj" podUID="581ac32b-a24f-4def-aab5-a2123f5674b3" Dec 06 07:30:16 crc kubenswrapper[4809]: I1206 07:30:16.111434 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-72dcl"] Dec 06 07:30:16 crc kubenswrapper[4809]: W1206 07:30:16.115927 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf8e767ec_8d4b_4a29_8562_115db074c7a7.slice/crio-cdfa35a9caab4e7d4ee8a85167cce8a1bffd2cdef3690f5352b9f64fc31c9b18 WatchSource:0}: Error finding container cdfa35a9caab4e7d4ee8a85167cce8a1bffd2cdef3690f5352b9f64fc31c9b18: Status 404 returned error can't find the container with id cdfa35a9caab4e7d4ee8a85167cce8a1bffd2cdef3690f5352b9f64fc31c9b18 Dec 06 07:30:16 crc kubenswrapper[4809]: I1206 07:30:16.679073 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416725-kq2pl"] Dec 06 07:30:16 crc kubenswrapper[4809]: I1206 07:30:16.691598 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416725-kq2pl"] Dec 06 07:30:16 crc kubenswrapper[4809]: I1206 07:30:16.971272 4809 generic.go:334] "Generic (PLEG): container finished" podID="f8e767ec-8d4b-4a29-8562-115db074c7a7" containerID="0b97e4ffae7d6277c83fa3bd2373a41ee4058001dc6905aee5d79d1ed1246815" exitCode=0 Dec 06 07:30:16 crc kubenswrapper[4809]: I1206 07:30:16.971334 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-72dcl" event={"ID":"f8e767ec-8d4b-4a29-8562-115db074c7a7","Type":"ContainerDied","Data":"0b97e4ffae7d6277c83fa3bd2373a41ee4058001dc6905aee5d79d1ed1246815"} Dec 06 07:30:16 crc kubenswrapper[4809]: I1206 07:30:16.971365 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-72dcl" event={"ID":"f8e767ec-8d4b-4a29-8562-115db074c7a7","Type":"ContainerStarted","Data":"cdfa35a9caab4e7d4ee8a85167cce8a1bffd2cdef3690f5352b9f64fc31c9b18"} Dec 06 07:30:17 crc kubenswrapper[4809]: I1206 07:30:17.467318 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9bf15846-fb7f-45fc-9146-02092049e864" path="/var/lib/kubelet/pods/9bf15846-fb7f-45fc-9146-02092049e864/volumes" Dec 06 07:30:23 crc kubenswrapper[4809]: I1206 07:30:23.062957 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-72dcl" event={"ID":"f8e767ec-8d4b-4a29-8562-115db074c7a7","Type":"ContainerStarted","Data":"bc3e1fba660cd1a38b4a6710fc99428c30d9fef71badbf770363fee6f0c0ae70"} Dec 06 07:30:24 crc kubenswrapper[4809]: I1206 07:30:24.074673 4809 generic.go:334] "Generic (PLEG): container finished" podID="f8e767ec-8d4b-4a29-8562-115db074c7a7" containerID="bc3e1fba660cd1a38b4a6710fc99428c30d9fef71badbf770363fee6f0c0ae70" exitCode=0 Dec 06 07:30:24 crc kubenswrapper[4809]: I1206 07:30:24.074747 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-72dcl" event={"ID":"f8e767ec-8d4b-4a29-8562-115db074c7a7","Type":"ContainerDied","Data":"bc3e1fba660cd1a38b4a6710fc99428c30d9fef71badbf770363fee6f0c0ae70"} Dec 06 07:30:25 crc kubenswrapper[4809]: I1206 07:30:25.087857 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-72dcl" event={"ID":"f8e767ec-8d4b-4a29-8562-115db074c7a7","Type":"ContainerStarted","Data":"28b0fa63657b2a9cbfcd7ca4107468150dfb6df35e721157ad2197001e137d3a"} Dec 06 07:30:25 crc kubenswrapper[4809]: I1206 07:30:25.109643 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-72dcl" podStartSLOduration=8.604881738 podStartE2EDuration="16.109624052s" podCreationTimestamp="2025-12-06 07:30:09 +0000 UTC" firstStartedPulling="2025-12-06 07:30:16.974053068 +0000 UTC m=+5941.863036010" lastFinishedPulling="2025-12-06 07:30:24.478795382 +0000 UTC m=+5949.367778324" observedRunningTime="2025-12-06 07:30:25.103896007 +0000 UTC m=+5949.992878959" watchObservedRunningTime="2025-12-06 07:30:25.109624052 +0000 UTC m=+5949.998606994" Dec 06 07:30:27 crc kubenswrapper[4809]: I1206 07:30:27.391847 4809 scope.go:117] "RemoveContainer" containerID="26a5b8063411f48708112d11ebddb0587d0fe6fc8f6a5c0aa5728fedd5b56f1d" Dec 06 07:30:27 crc kubenswrapper[4809]: E1206 07:30:27.392628 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:30:29 crc kubenswrapper[4809]: I1206 07:30:29.613621 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-72dcl" Dec 06 07:30:29 crc kubenswrapper[4809]: I1206 07:30:29.613970 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-72dcl" Dec 06 07:30:29 crc kubenswrapper[4809]: I1206 07:30:29.672837 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-72dcl" Dec 06 07:30:30 crc kubenswrapper[4809]: I1206 07:30:30.201830 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-72dcl" Dec 06 07:30:30 crc kubenswrapper[4809]: I1206 07:30:30.452741 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-72dcl"] Dec 06 07:30:30 crc kubenswrapper[4809]: I1206 07:30:30.526616 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4p5rc"] Dec 06 07:30:30 crc kubenswrapper[4809]: I1206 07:30:30.527788 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4p5rc" podUID="6fa58271-edf3-4d3d-8782-83a959a691fe" containerName="registry-server" containerID="cri-o://779b773e8f19051505ec32af881e6187c8ad42f783c8123b2f9f9191193699dc" gracePeriod=2 Dec 06 07:30:31 crc kubenswrapper[4809]: I1206 07:30:31.159720 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bn94p/crc-debug-dkcmj" event={"ID":"581ac32b-a24f-4def-aab5-a2123f5674b3","Type":"ContainerStarted","Data":"f2994613b8199bc578ca18d9e08b31c4952eea95c7c4653c21b9c1e9c9477b98"} Dec 06 07:30:31 crc kubenswrapper[4809]: I1206 07:30:31.179452 4809 generic.go:334] "Generic (PLEG): container finished" podID="6fa58271-edf3-4d3d-8782-83a959a691fe" containerID="779b773e8f19051505ec32af881e6187c8ad42f783c8123b2f9f9191193699dc" exitCode=0 Dec 06 07:30:31 crc kubenswrapper[4809]: I1206 07:30:31.180919 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4p5rc" event={"ID":"6fa58271-edf3-4d3d-8782-83a959a691fe","Type":"ContainerDied","Data":"779b773e8f19051505ec32af881e6187c8ad42f783c8123b2f9f9191193699dc"} Dec 06 07:30:31 crc kubenswrapper[4809]: I1206 07:30:31.190352 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-bn94p/crc-debug-dkcmj" podStartSLOduration=1.945010002 podStartE2EDuration="33.19033619s" podCreationTimestamp="2025-12-06 07:29:58 +0000 UTC" firstStartedPulling="2025-12-06 07:29:59.280654547 +0000 UTC m=+5924.169637489" lastFinishedPulling="2025-12-06 07:30:30.525980735 +0000 UTC m=+5955.414963677" observedRunningTime="2025-12-06 07:30:31.18922535 +0000 UTC m=+5956.078208292" watchObservedRunningTime="2025-12-06 07:30:31.19033619 +0000 UTC m=+5956.079319132" Dec 06 07:30:31 crc kubenswrapper[4809]: I1206 07:30:31.347712 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4p5rc" Dec 06 07:30:31 crc kubenswrapper[4809]: I1206 07:30:31.388838 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4qlt\" (UniqueName: \"kubernetes.io/projected/6fa58271-edf3-4d3d-8782-83a959a691fe-kube-api-access-m4qlt\") pod \"6fa58271-edf3-4d3d-8782-83a959a691fe\" (UID: \"6fa58271-edf3-4d3d-8782-83a959a691fe\") " Dec 06 07:30:31 crc kubenswrapper[4809]: I1206 07:30:31.389278 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fa58271-edf3-4d3d-8782-83a959a691fe-catalog-content\") pod \"6fa58271-edf3-4d3d-8782-83a959a691fe\" (UID: \"6fa58271-edf3-4d3d-8782-83a959a691fe\") " Dec 06 07:30:31 crc kubenswrapper[4809]: I1206 07:30:31.389369 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fa58271-edf3-4d3d-8782-83a959a691fe-utilities\") pod \"6fa58271-edf3-4d3d-8782-83a959a691fe\" (UID: \"6fa58271-edf3-4d3d-8782-83a959a691fe\") " Dec 06 07:30:31 crc kubenswrapper[4809]: I1206 07:30:31.396728 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6fa58271-edf3-4d3d-8782-83a959a691fe-utilities" (OuterVolumeSpecName: "utilities") pod "6fa58271-edf3-4d3d-8782-83a959a691fe" (UID: "6fa58271-edf3-4d3d-8782-83a959a691fe"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:30:31 crc kubenswrapper[4809]: I1206 07:30:31.413469 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fa58271-edf3-4d3d-8782-83a959a691fe-kube-api-access-m4qlt" (OuterVolumeSpecName: "kube-api-access-m4qlt") pod "6fa58271-edf3-4d3d-8782-83a959a691fe" (UID: "6fa58271-edf3-4d3d-8782-83a959a691fe"). InnerVolumeSpecName "kube-api-access-m4qlt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:30:31 crc kubenswrapper[4809]: I1206 07:30:31.492120 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4qlt\" (UniqueName: \"kubernetes.io/projected/6fa58271-edf3-4d3d-8782-83a959a691fe-kube-api-access-m4qlt\") on node \"crc\" DevicePath \"\"" Dec 06 07:30:31 crc kubenswrapper[4809]: I1206 07:30:31.492146 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fa58271-edf3-4d3d-8782-83a959a691fe-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 07:30:31 crc kubenswrapper[4809]: I1206 07:30:31.544152 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6fa58271-edf3-4d3d-8782-83a959a691fe-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6fa58271-edf3-4d3d-8782-83a959a691fe" (UID: "6fa58271-edf3-4d3d-8782-83a959a691fe"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:30:31 crc kubenswrapper[4809]: I1206 07:30:31.595035 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fa58271-edf3-4d3d-8782-83a959a691fe-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 07:30:32 crc kubenswrapper[4809]: I1206 07:30:32.192917 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4p5rc" Dec 06 07:30:32 crc kubenswrapper[4809]: I1206 07:30:32.194054 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4p5rc" event={"ID":"6fa58271-edf3-4d3d-8782-83a959a691fe","Type":"ContainerDied","Data":"c7521299070a87cd972f9e2a72cc51c65cf3bb93f7a26f49c7051daddcdfca45"} Dec 06 07:30:32 crc kubenswrapper[4809]: I1206 07:30:32.194091 4809 scope.go:117] "RemoveContainer" containerID="779b773e8f19051505ec32af881e6187c8ad42f783c8123b2f9f9191193699dc" Dec 06 07:30:32 crc kubenswrapper[4809]: I1206 07:30:32.233818 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4p5rc"] Dec 06 07:30:32 crc kubenswrapper[4809]: I1206 07:30:32.248650 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4p5rc"] Dec 06 07:30:32 crc kubenswrapper[4809]: I1206 07:30:32.250668 4809 scope.go:117] "RemoveContainer" containerID="73e1ce115ef05e486697731b5849964166674d52c24b8da30ad9101d42eef954" Dec 06 07:30:32 crc kubenswrapper[4809]: I1206 07:30:32.285688 4809 scope.go:117] "RemoveContainer" containerID="d0928cbb94dfa71eb6848cc8cf6441a184a1dbec2a82e18c94293ad8554754e0" Dec 06 07:30:33 crc kubenswrapper[4809]: I1206 07:30:33.441084 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6fa58271-edf3-4d3d-8782-83a959a691fe" path="/var/lib/kubelet/pods/6fa58271-edf3-4d3d-8782-83a959a691fe/volumes" Dec 06 07:30:36 crc kubenswrapper[4809]: I1206 07:30:36.247697 4809 generic.go:334] "Generic (PLEG): container finished" podID="b4f96141-d70e-4715-9e30-516c9e783cc9" containerID="cac60d4c52fd5a110675ebf867d8dabeb845a5056ba0753e52a1a1f592f0790d" exitCode=0 Dec 06 07:30:36 crc kubenswrapper[4809]: I1206 07:30:36.247777 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-556bbc5579-8wblw" event={"ID":"b4f96141-d70e-4715-9e30-516c9e783cc9","Type":"ContainerDied","Data":"cac60d4c52fd5a110675ebf867d8dabeb845a5056ba0753e52a1a1f592f0790d"} Dec 06 07:30:37 crc kubenswrapper[4809]: I1206 07:30:37.260760 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-556bbc5579-8wblw" event={"ID":"b4f96141-d70e-4715-9e30-516c9e783cc9","Type":"ContainerStarted","Data":"56876babc9f9d93636e382a69b7ae3b9dbbb0a1b6c8ca6be1b9e93631575301b"} Dec 06 07:30:41 crc kubenswrapper[4809]: I1206 07:30:41.224686 4809 scope.go:117] "RemoveContainer" containerID="154f2d114d6b997794c52348b060d97bce9656c84765b90e42eee6e301a8d492" Dec 06 07:30:41 crc kubenswrapper[4809]: I1206 07:30:41.394560 4809 scope.go:117] "RemoveContainer" containerID="26a5b8063411f48708112d11ebddb0587d0fe6fc8f6a5c0aa5728fedd5b56f1d" Dec 06 07:30:41 crc kubenswrapper[4809]: E1206 07:30:41.395013 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:30:44 crc kubenswrapper[4809]: I1206 07:30:44.422609 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/metrics-server-556bbc5579-8wblw" Dec 06 07:30:44 crc kubenswrapper[4809]: I1206 07:30:44.424497 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/metrics-server-556bbc5579-8wblw" Dec 06 07:30:53 crc kubenswrapper[4809]: I1206 07:30:53.335680 4809 patch_prober.go:28] interesting pod/logging-loki-gateway-65498c4f8f-48rjq container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.53:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:30:53 crc kubenswrapper[4809]: I1206 07:30:53.336212 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-65498c4f8f-48rjq" podUID="72a3ed95-3e3d-4faf-88e4-ad1731902910" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.53:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:30:54 crc kubenswrapper[4809]: I1206 07:30:54.033471 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.435748343s: [/var/lib/containers/storage/overlay/36697314b3e9da74a9047d3bf0b33b766f40aa6592cbe6e8789a36d52b30fe94/diff /var/log/pods/openshift-must-gather-bn94p_must-gather-t7bfm_3603c2a8-988f-4ecf-ad1b-51b74176a0a1/copy/0.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:30:54 crc kubenswrapper[4809]: I1206 07:30:54.033528 4809 trace.go:236] Trace[1956249286]: "Calculate volume metrics of available-featuregates for pod openshift-config-operator/openshift-config-operator-7777fb866f-qqv6x" (06-Dec-2025 07:30:51.899) (total time: 2133ms): Dec 06 07:30:54 crc kubenswrapper[4809]: Trace[1956249286]: [2.13386154s] [2.13386154s] END Dec 06 07:30:54 crc kubenswrapper[4809]: E1206 07:30:54.037393 4809 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="2.65s" Dec 06 07:30:54 crc kubenswrapper[4809]: I1206 07:30:54.058654 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.747468707s: [/var/lib/containers/storage/overlay/182d020d8b2fd5700ff02688ff5b9c351165e994a1e24e3564d81ffe371f0cff/diff /var/log/pods/openshift-must-gather-bn94p_must-gather-t7bfm_3603c2a8-988f-4ecf-ad1b-51b74176a0a1/gather/0.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:30:55 crc kubenswrapper[4809]: I1206 07:30:55.399282 4809 scope.go:117] "RemoveContainer" containerID="26a5b8063411f48708112d11ebddb0587d0fe6fc8f6a5c0aa5728fedd5b56f1d" Dec 06 07:30:55 crc kubenswrapper[4809]: E1206 07:30:55.401540 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:31:05 crc kubenswrapper[4809]: I1206 07:31:05.958482 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 2.709721844s: [/var/lib/containers/storage/overlay/592e597f4f1dc4ed366ad8d9bf72c4bb167de09d4e09e74f7b1a1e8bcc8ede24/diff /var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-27fdn_e6e0283a-55e9-4613-b74d-81fbb840690e/manager/0.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:31:05 crc kubenswrapper[4809]: I1206 07:31:05.959148 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 2.710408363s: [/var/lib/containers/storage/overlay/01cd3cad70f55a8458da9780217b5192c28bc6785a20f995ccceb70e8f583174/diff /var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-7trbt_45f87079-e852-4b47-a240-6f2bcf247509/manager/0.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:31:05 crc kubenswrapper[4809]: I1206 07:31:05.959287 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 2.710532395s: [/var/lib/containers/storage/overlay/31abc487b3c221391fd74ba3739c66e2b50821af1eb38f209f8a0984cf5b5dcd/diff /var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-vqvvr_803d54f1-fbac-4ec5-b486-efae1981b46d/manager/0.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:31:06 crc kubenswrapper[4809]: I1206 07:31:06.070284 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 2.821561677s: [/var/lib/containers/storage/overlay/0f7c249c8a6be967dd2f80384a7f6cffbb8064abc1705d0201e77401ec2dd0ea/diff /var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-gzv6k_ced7591a-f80d-46a0-b8ea-024c2e37ae9f/manager/0.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:31:06 crc kubenswrapper[4809]: I1206 07:31:06.070459 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 2.8217117s: [/var/lib/containers/storage/overlay/a7efdd7f8c0be946aa4c794faa4afb7809516fbd5bcc1605730fd4aa1b29ba1f/diff /var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-sxcr7_c37c165f-5cd1-496f-b092-41efdfcd7eca/manager/0.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:31:06 crc kubenswrapper[4809]: I1206 07:31:06.070502 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 2.820649692s: [/var/lib/containers/storage/overlay/8ad7f66e3a19b1297b6a65ccc08bd85ca736fbefd8e5f918a3a5075f46ba49b0/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:31:06 crc kubenswrapper[4809]: I1206 07:31:06.073451 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.848246231s: [/var/lib/containers/storage/overlay/0838bd50ae081f813e26a2458df0b7654b1bb9b2aaea91fcfdf7ff71657a7e26/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:31:06 crc kubenswrapper[4809]: I1206 07:31:06.082269 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 2.586592375s: [/var/lib/containers/storage/overlay/7abf4f10e79568aeea94cb2a5f7a3b6901e85534b72e27654dc1cd7c7c5a3111/diff /var/log/pods/openstack-operators_telemetry-operator-controller-manager-b4644b845-zxwvr_24aed9fd-d792-4bce-82a8-3ea4d70094b4/manager/0.log]; will not log again for this container unless duration exceeds 3s Dec 06 07:31:06 crc kubenswrapper[4809]: I1206 07:31:06.097758 4809 patch_prober.go:28] interesting pod/logging-loki-gateway-65498c4f8f-48rjq container/gateway namespace/openshift-logging: Readiness probe status=failure output="" start-of-body= Dec 06 07:31:06 crc kubenswrapper[4809]: I1206 07:31:06.109285 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 2.832713497s: [/var/lib/containers/storage/overlay/9a118a0daad237f17305d791695e07aa76e40f69087f49fa94a9c7c4c7ae7f28/diff /var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-bhbcl_fd7dd0a5-5e39-4812-b93f-a27c39284d33/manager/0.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:31:06 crc kubenswrapper[4809]: I1206 07:31:06.115850 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/metrics-server-556bbc5579-8wblw" Dec 06 07:31:06 crc kubenswrapper[4809]: I1206 07:31:06.117396 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 2.582290189s: [/var/lib/containers/storage/overlay/5181ed4775950b0dd2042abba893e704caea0c4da288ad7cdd1678004c62a41e/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:31:06 crc kubenswrapper[4809]: I1206 07:31:06.119663 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 2.62078554s: [/var/lib/containers/storage/overlay/cb9fbf785fa7fc39b3f073e742f76731d6c85932f7d38693def427fe32cf04fe/diff /var/log/pods/openstack_aodh-0_7e16d0f3-9bd2-4887-84dd-3b8fb35d6451/aodh-api/0.log]; will not log again for this container unless duration exceeds 3s Dec 06 07:31:06 crc kubenswrapper[4809]: I1206 07:31:06.124250 4809 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.55:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:31:06 crc kubenswrapper[4809]: I1206 07:31:06.124301 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/metrics-server-556bbc5579-8wblw" Dec 06 07:31:06 crc kubenswrapper[4809]: I1206 07:31:06.124297 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="0f3f468b-668a-4b15-af5b-c519c4ef03d7" containerName="loki-ingester" probeResult="failure" output="Get \"https://10.217.0.55:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:31:06 crc kubenswrapper[4809]: I1206 07:31:06.168627 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 2.314755599s: [/var/lib/containers/storage/overlay/59ddffa261ff3e9a0ac83c44edc629000f8e30a37ad91be2c20fb812437d3680/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:31:06 crc kubenswrapper[4809]: I1206 07:31:06.176673 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 2.034955248s: [/var/lib/containers/storage/overlay/3147df11bdb38b4c0b894282f458916d9cfef0095ee0e0fa78ddde1aa7d89271/diff /var/log/pods/openstack_nova-cell1-novncproxy-0_663c54d8-faa9-466a-9126-678a92bc92f9/nova-cell1-novncproxy-novncproxy/0.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:31:06 crc kubenswrapper[4809]: I1206 07:31:06.249459 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8xxdj"] Dec 06 07:31:06 crc kubenswrapper[4809]: E1206 07:31:06.250169 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fa58271-edf3-4d3d-8782-83a959a691fe" containerName="extract-utilities" Dec 06 07:31:06 crc kubenswrapper[4809]: I1206 07:31:06.250186 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fa58271-edf3-4d3d-8782-83a959a691fe" containerName="extract-utilities" Dec 06 07:31:06 crc kubenswrapper[4809]: E1206 07:31:06.250199 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fa58271-edf3-4d3d-8782-83a959a691fe" containerName="registry-server" Dec 06 07:31:06 crc kubenswrapper[4809]: I1206 07:31:06.250207 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fa58271-edf3-4d3d-8782-83a959a691fe" containerName="registry-server" Dec 06 07:31:06 crc kubenswrapper[4809]: E1206 07:31:06.250223 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fa58271-edf3-4d3d-8782-83a959a691fe" containerName="extract-content" Dec 06 07:31:06 crc kubenswrapper[4809]: I1206 07:31:06.250231 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fa58271-edf3-4d3d-8782-83a959a691fe" containerName="extract-content" Dec 06 07:31:06 crc kubenswrapper[4809]: E1206 07:31:06.250258 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="034aef25-c1f0-487a-80c3-2ba30229cc13" containerName="collect-profiles" Dec 06 07:31:06 crc kubenswrapper[4809]: I1206 07:31:06.250266 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="034aef25-c1f0-487a-80c3-2ba30229cc13" containerName="collect-profiles" Dec 06 07:31:06 crc kubenswrapper[4809]: I1206 07:31:06.258337 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="034aef25-c1f0-487a-80c3-2ba30229cc13" containerName="collect-profiles" Dec 06 07:31:06 crc kubenswrapper[4809]: I1206 07:31:06.258387 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fa58271-edf3-4d3d-8782-83a959a691fe" containerName="registry-server" Dec 06 07:31:06 crc kubenswrapper[4809]: I1206 07:31:06.260596 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8xxdj" Dec 06 07:31:06 crc kubenswrapper[4809]: I1206 07:31:06.306619 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8xxdj"] Dec 06 07:31:06 crc kubenswrapper[4809]: I1206 07:31:06.330578 4809 trace.go:236] Trace[2076337083]: "Calculate volume metrics of storage for pod minio-dev/minio" (06-Dec-2025 07:31:01.348) (total time: 4981ms): Dec 06 07:31:06 crc kubenswrapper[4809]: Trace[2076337083]: [4.981769818s] [4.981769818s] END Dec 06 07:31:06 crc kubenswrapper[4809]: I1206 07:31:06.362623 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98ca3772-27e9-4eb9-8864-d9fc08c9fb42-catalog-content\") pod \"redhat-operators-8xxdj\" (UID: \"98ca3772-27e9-4eb9-8864-d9fc08c9fb42\") " pod="openshift-marketplace/redhat-operators-8xxdj" Dec 06 07:31:06 crc kubenswrapper[4809]: I1206 07:31:06.362755 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfrg7\" (UniqueName: \"kubernetes.io/projected/98ca3772-27e9-4eb9-8864-d9fc08c9fb42-kube-api-access-vfrg7\") pod \"redhat-operators-8xxdj\" (UID: \"98ca3772-27e9-4eb9-8864-d9fc08c9fb42\") " pod="openshift-marketplace/redhat-operators-8xxdj" Dec 06 07:31:06 crc kubenswrapper[4809]: I1206 07:31:06.362855 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98ca3772-27e9-4eb9-8864-d9fc08c9fb42-utilities\") pod \"redhat-operators-8xxdj\" (UID: \"98ca3772-27e9-4eb9-8864-d9fc08c9fb42\") " pod="openshift-marketplace/redhat-operators-8xxdj" Dec 06 07:31:06 crc kubenswrapper[4809]: I1206 07:31:06.388351 4809 scope.go:117] "RemoveContainer" containerID="26a5b8063411f48708112d11ebddb0587d0fe6fc8f6a5c0aa5728fedd5b56f1d" Dec 06 07:31:06 crc kubenswrapper[4809]: E1206 07:31:06.388768 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:31:06 crc kubenswrapper[4809]: I1206 07:31:06.465776 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98ca3772-27e9-4eb9-8864-d9fc08c9fb42-utilities\") pod \"redhat-operators-8xxdj\" (UID: \"98ca3772-27e9-4eb9-8864-d9fc08c9fb42\") " pod="openshift-marketplace/redhat-operators-8xxdj" Dec 06 07:31:06 crc kubenswrapper[4809]: I1206 07:31:06.466180 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98ca3772-27e9-4eb9-8864-d9fc08c9fb42-catalog-content\") pod \"redhat-operators-8xxdj\" (UID: \"98ca3772-27e9-4eb9-8864-d9fc08c9fb42\") " pod="openshift-marketplace/redhat-operators-8xxdj" Dec 06 07:31:06 crc kubenswrapper[4809]: I1206 07:31:06.466329 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfrg7\" (UniqueName: \"kubernetes.io/projected/98ca3772-27e9-4eb9-8864-d9fc08c9fb42-kube-api-access-vfrg7\") pod \"redhat-operators-8xxdj\" (UID: \"98ca3772-27e9-4eb9-8864-d9fc08c9fb42\") " pod="openshift-marketplace/redhat-operators-8xxdj" Dec 06 07:31:06 crc kubenswrapper[4809]: I1206 07:31:06.466708 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98ca3772-27e9-4eb9-8864-d9fc08c9fb42-utilities\") pod \"redhat-operators-8xxdj\" (UID: \"98ca3772-27e9-4eb9-8864-d9fc08c9fb42\") " pod="openshift-marketplace/redhat-operators-8xxdj" Dec 06 07:31:06 crc kubenswrapper[4809]: I1206 07:31:06.466971 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98ca3772-27e9-4eb9-8864-d9fc08c9fb42-catalog-content\") pod \"redhat-operators-8xxdj\" (UID: \"98ca3772-27e9-4eb9-8864-d9fc08c9fb42\") " pod="openshift-marketplace/redhat-operators-8xxdj" Dec 06 07:31:06 crc kubenswrapper[4809]: I1206 07:31:06.500297 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfrg7\" (UniqueName: \"kubernetes.io/projected/98ca3772-27e9-4eb9-8864-d9fc08c9fb42-kube-api-access-vfrg7\") pod \"redhat-operators-8xxdj\" (UID: \"98ca3772-27e9-4eb9-8864-d9fc08c9fb42\") " pod="openshift-marketplace/redhat-operators-8xxdj" Dec 06 07:31:06 crc kubenswrapper[4809]: I1206 07:31:06.650179 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8xxdj" Dec 06 07:31:07 crc kubenswrapper[4809]: I1206 07:31:07.468963 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8xxdj"] Dec 06 07:31:08 crc kubenswrapper[4809]: I1206 07:31:08.192506 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8xxdj" event={"ID":"98ca3772-27e9-4eb9-8864-d9fc08c9fb42","Type":"ContainerStarted","Data":"edce605df71505081f732fbc56c8b58d3dbeebec79655b93c88b4b5e4c47719e"} Dec 06 07:31:09 crc kubenswrapper[4809]: I1206 07:31:09.206244 4809 generic.go:334] "Generic (PLEG): container finished" podID="98ca3772-27e9-4eb9-8864-d9fc08c9fb42" containerID="826dedb50f6220b735a389dbd7b9c873d1da47bc0d6912b5906bbcda9cbdfd13" exitCode=0 Dec 06 07:31:09 crc kubenswrapper[4809]: I1206 07:31:09.207442 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8xxdj" event={"ID":"98ca3772-27e9-4eb9-8864-d9fc08c9fb42","Type":"ContainerDied","Data":"826dedb50f6220b735a389dbd7b9c873d1da47bc0d6912b5906bbcda9cbdfd13"} Dec 06 07:31:11 crc kubenswrapper[4809]: I1206 07:31:11.955209 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/speaker-5t9z6" podUID="21411d89-4a48-4556-8ad2-193f1c8be0db" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:31:13 crc kubenswrapper[4809]: I1206 07:31:13.395094 4809 patch_prober.go:28] interesting pod/logging-loki-gateway-65498c4f8f-9s8qq container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.54:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:31:13 crc kubenswrapper[4809]: I1206 07:31:13.395177 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq" podUID="ef8f3e7e-57ea-42b8-a777-1778e5ed975b" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.54:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:31:13 crc kubenswrapper[4809]: I1206 07:31:13.550262 4809 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-r2lg6 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.63:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:31:13 crc kubenswrapper[4809]: I1206 07:31:13.550571 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-r2lg6" podUID="286c8de4-0ea9-43ad-bb4f-970319e0f4d3" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.63:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:31:14 crc kubenswrapper[4809]: I1206 07:31:14.860144 4809 patch_prober.go:28] interesting pod/observability-operator-d8bb48f5d-bfc8s container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.22:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:31:14 crc kubenswrapper[4809]: I1206 07:31:14.860214 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-d8bb48f5d-bfc8s" podUID="6d4f55df-5dc2-4e72-b7f8-cee9229da476" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.22:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:31:19 crc kubenswrapper[4809]: I1206 07:31:19.390006 4809 scope.go:117] "RemoveContainer" containerID="26a5b8063411f48708112d11ebddb0587d0fe6fc8f6a5c0aa5728fedd5b56f1d" Dec 06 07:31:19 crc kubenswrapper[4809]: E1206 07:31:19.392771 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:31:20 crc kubenswrapper[4809]: I1206 07:31:20.345716 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8xxdj" event={"ID":"98ca3772-27e9-4eb9-8864-d9fc08c9fb42","Type":"ContainerStarted","Data":"b7af08a5c2120118aa31ddfa7ec6c9f86d7557a335eee3153b7f728bf2967d07"} Dec 06 07:31:29 crc kubenswrapper[4809]: I1206 07:31:29.482322 4809 generic.go:334] "Generic (PLEG): container finished" podID="98ca3772-27e9-4eb9-8864-d9fc08c9fb42" containerID="b7af08a5c2120118aa31ddfa7ec6c9f86d7557a335eee3153b7f728bf2967d07" exitCode=0 Dec 06 07:31:29 crc kubenswrapper[4809]: I1206 07:31:29.482439 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8xxdj" event={"ID":"98ca3772-27e9-4eb9-8864-d9fc08c9fb42","Type":"ContainerDied","Data":"b7af08a5c2120118aa31ddfa7ec6c9f86d7557a335eee3153b7f728bf2967d07"} Dec 06 07:31:31 crc kubenswrapper[4809]: I1206 07:31:31.520648 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8xxdj" event={"ID":"98ca3772-27e9-4eb9-8864-d9fc08c9fb42","Type":"ContainerStarted","Data":"882a83d706b50e319f33db4d48863e22255e724901d50821a648c0b404d98f96"} Dec 06 07:31:31 crc kubenswrapper[4809]: I1206 07:31:31.553844 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8xxdj" podStartSLOduration=6.474571604 podStartE2EDuration="26.553804676s" podCreationTimestamp="2025-12-06 07:31:05 +0000 UTC" firstStartedPulling="2025-12-06 07:31:10.221239452 +0000 UTC m=+5995.110222404" lastFinishedPulling="2025-12-06 07:31:30.300472534 +0000 UTC m=+6015.189455476" observedRunningTime="2025-12-06 07:31:31.5513634 +0000 UTC m=+6016.440346352" watchObservedRunningTime="2025-12-06 07:31:31.553804676 +0000 UTC m=+6016.442787618" Dec 06 07:31:33 crc kubenswrapper[4809]: I1206 07:31:33.389539 4809 scope.go:117] "RemoveContainer" containerID="26a5b8063411f48708112d11ebddb0587d0fe6fc8f6a5c0aa5728fedd5b56f1d" Dec 06 07:31:33 crc kubenswrapper[4809]: E1206 07:31:33.390106 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:31:36 crc kubenswrapper[4809]: I1206 07:31:36.651103 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8xxdj" Dec 06 07:31:36 crc kubenswrapper[4809]: I1206 07:31:36.652788 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8xxdj" Dec 06 07:31:37 crc kubenswrapper[4809]: I1206 07:31:37.723469 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8xxdj" podUID="98ca3772-27e9-4eb9-8864-d9fc08c9fb42" containerName="registry-server" probeResult="failure" output=< Dec 06 07:31:37 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Dec 06 07:31:37 crc kubenswrapper[4809]: > Dec 06 07:31:46 crc kubenswrapper[4809]: I1206 07:31:46.389095 4809 scope.go:117] "RemoveContainer" containerID="26a5b8063411f48708112d11ebddb0587d0fe6fc8f6a5c0aa5728fedd5b56f1d" Dec 06 07:31:46 crc kubenswrapper[4809]: E1206 07:31:46.389900 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:31:47 crc kubenswrapper[4809]: I1206 07:31:47.712770 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8xxdj" podUID="98ca3772-27e9-4eb9-8864-d9fc08c9fb42" containerName="registry-server" probeResult="failure" output=< Dec 06 07:31:47 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Dec 06 07:31:47 crc kubenswrapper[4809]: > Dec 06 07:31:56 crc kubenswrapper[4809]: I1206 07:31:56.710141 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8xxdj" Dec 06 07:31:58 crc kubenswrapper[4809]: I1206 07:31:56.788098 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8xxdj" Dec 06 07:31:58 crc kubenswrapper[4809]: I1206 07:31:56.952529 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8xxdj"] Dec 06 07:31:58 crc kubenswrapper[4809]: I1206 07:31:57.827844 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8xxdj" podUID="98ca3772-27e9-4eb9-8864-d9fc08c9fb42" containerName="registry-server" containerID="cri-o://882a83d706b50e319f33db4d48863e22255e724901d50821a648c0b404d98f96" gracePeriod=2 Dec 06 07:31:58 crc kubenswrapper[4809]: I1206 07:31:57.955536 4809 patch_prober.go:28] interesting pod/oauth-openshift-55889b984c-nxw8t container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.56:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:31:58 crc kubenswrapper[4809]: I1206 07:31:57.955608 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" podUID="404214e0-30aa-44ff-b2d0-fb3c127068f9" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.56:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:32:00 crc kubenswrapper[4809]: I1206 07:32:00.871537 4809 generic.go:334] "Generic (PLEG): container finished" podID="98ca3772-27e9-4eb9-8864-d9fc08c9fb42" containerID="882a83d706b50e319f33db4d48863e22255e724901d50821a648c0b404d98f96" exitCode=0 Dec 06 07:32:00 crc kubenswrapper[4809]: I1206 07:32:00.871619 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8xxdj" event={"ID":"98ca3772-27e9-4eb9-8864-d9fc08c9fb42","Type":"ContainerDied","Data":"882a83d706b50e319f33db4d48863e22255e724901d50821a648c0b404d98f96"} Dec 06 07:32:01 crc kubenswrapper[4809]: I1206 07:32:01.391582 4809 scope.go:117] "RemoveContainer" containerID="26a5b8063411f48708112d11ebddb0587d0fe6fc8f6a5c0aa5728fedd5b56f1d" Dec 06 07:32:01 crc kubenswrapper[4809]: E1206 07:32:01.391845 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:32:01 crc kubenswrapper[4809]: I1206 07:32:01.730436 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8xxdj" Dec 06 07:32:01 crc kubenswrapper[4809]: I1206 07:32:01.743743 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfrg7\" (UniqueName: \"kubernetes.io/projected/98ca3772-27e9-4eb9-8864-d9fc08c9fb42-kube-api-access-vfrg7\") pod \"98ca3772-27e9-4eb9-8864-d9fc08c9fb42\" (UID: \"98ca3772-27e9-4eb9-8864-d9fc08c9fb42\") " Dec 06 07:32:01 crc kubenswrapper[4809]: I1206 07:32:01.745761 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98ca3772-27e9-4eb9-8864-d9fc08c9fb42-utilities\") pod \"98ca3772-27e9-4eb9-8864-d9fc08c9fb42\" (UID: \"98ca3772-27e9-4eb9-8864-d9fc08c9fb42\") " Dec 06 07:32:01 crc kubenswrapper[4809]: I1206 07:32:01.748677 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98ca3772-27e9-4eb9-8864-d9fc08c9fb42-catalog-content\") pod \"98ca3772-27e9-4eb9-8864-d9fc08c9fb42\" (UID: \"98ca3772-27e9-4eb9-8864-d9fc08c9fb42\") " Dec 06 07:32:01 crc kubenswrapper[4809]: I1206 07:32:01.748554 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98ca3772-27e9-4eb9-8864-d9fc08c9fb42-utilities" (OuterVolumeSpecName: "utilities") pod "98ca3772-27e9-4eb9-8864-d9fc08c9fb42" (UID: "98ca3772-27e9-4eb9-8864-d9fc08c9fb42"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:32:01 crc kubenswrapper[4809]: I1206 07:32:01.749883 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98ca3772-27e9-4eb9-8864-d9fc08c9fb42-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 07:32:01 crc kubenswrapper[4809]: I1206 07:32:01.752044 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98ca3772-27e9-4eb9-8864-d9fc08c9fb42-kube-api-access-vfrg7" (OuterVolumeSpecName: "kube-api-access-vfrg7") pod "98ca3772-27e9-4eb9-8864-d9fc08c9fb42" (UID: "98ca3772-27e9-4eb9-8864-d9fc08c9fb42"). InnerVolumeSpecName "kube-api-access-vfrg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:32:01 crc kubenswrapper[4809]: I1206 07:32:01.852000 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfrg7\" (UniqueName: \"kubernetes.io/projected/98ca3772-27e9-4eb9-8864-d9fc08c9fb42-kube-api-access-vfrg7\") on node \"crc\" DevicePath \"\"" Dec 06 07:32:01 crc kubenswrapper[4809]: I1206 07:32:01.862068 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98ca3772-27e9-4eb9-8864-d9fc08c9fb42-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "98ca3772-27e9-4eb9-8864-d9fc08c9fb42" (UID: "98ca3772-27e9-4eb9-8864-d9fc08c9fb42"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:32:01 crc kubenswrapper[4809]: I1206 07:32:01.885684 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8xxdj" event={"ID":"98ca3772-27e9-4eb9-8864-d9fc08c9fb42","Type":"ContainerDied","Data":"edce605df71505081f732fbc56c8b58d3dbeebec79655b93c88b4b5e4c47719e"} Dec 06 07:32:01 crc kubenswrapper[4809]: I1206 07:32:01.885753 4809 scope.go:117] "RemoveContainer" containerID="882a83d706b50e319f33db4d48863e22255e724901d50821a648c0b404d98f96" Dec 06 07:32:01 crc kubenswrapper[4809]: I1206 07:32:01.885919 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8xxdj" Dec 06 07:32:01 crc kubenswrapper[4809]: I1206 07:32:01.929134 4809 scope.go:117] "RemoveContainer" containerID="b7af08a5c2120118aa31ddfa7ec6c9f86d7557a335eee3153b7f728bf2967d07" Dec 06 07:32:01 crc kubenswrapper[4809]: I1206 07:32:01.947002 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8xxdj"] Dec 06 07:32:01 crc kubenswrapper[4809]: I1206 07:32:01.955380 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98ca3772-27e9-4eb9-8864-d9fc08c9fb42-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 07:32:01 crc kubenswrapper[4809]: I1206 07:32:01.955424 4809 scope.go:117] "RemoveContainer" containerID="826dedb50f6220b735a389dbd7b9c873d1da47bc0d6912b5906bbcda9cbdfd13" Dec 06 07:32:01 crc kubenswrapper[4809]: I1206 07:32:01.958910 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8xxdj"] Dec 06 07:32:03 crc kubenswrapper[4809]: I1206 07:32:03.407144 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98ca3772-27e9-4eb9-8864-d9fc08c9fb42" path="/var/lib/kubelet/pods/98ca3772-27e9-4eb9-8864-d9fc08c9fb42/volumes" Dec 06 07:32:09 crc kubenswrapper[4809]: I1206 07:32:09.976310 4809 generic.go:334] "Generic (PLEG): container finished" podID="581ac32b-a24f-4def-aab5-a2123f5674b3" containerID="f2994613b8199bc578ca18d9e08b31c4952eea95c7c4653c21b9c1e9c9477b98" exitCode=0 Dec 06 07:32:09 crc kubenswrapper[4809]: I1206 07:32:09.976415 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bn94p/crc-debug-dkcmj" event={"ID":"581ac32b-a24f-4def-aab5-a2123f5674b3","Type":"ContainerDied","Data":"f2994613b8199bc578ca18d9e08b31c4952eea95c7c4653c21b9c1e9c9477b98"} Dec 06 07:32:11 crc kubenswrapper[4809]: I1206 07:32:11.122344 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bn94p/crc-debug-dkcmj" Dec 06 07:32:11 crc kubenswrapper[4809]: I1206 07:32:11.164992 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xkwt5\" (UniqueName: \"kubernetes.io/projected/581ac32b-a24f-4def-aab5-a2123f5674b3-kube-api-access-xkwt5\") pod \"581ac32b-a24f-4def-aab5-a2123f5674b3\" (UID: \"581ac32b-a24f-4def-aab5-a2123f5674b3\") " Dec 06 07:32:11 crc kubenswrapper[4809]: I1206 07:32:11.165132 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/581ac32b-a24f-4def-aab5-a2123f5674b3-host\") pod \"581ac32b-a24f-4def-aab5-a2123f5674b3\" (UID: \"581ac32b-a24f-4def-aab5-a2123f5674b3\") " Dec 06 07:32:11 crc kubenswrapper[4809]: I1206 07:32:11.166199 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/581ac32b-a24f-4def-aab5-a2123f5674b3-host" (OuterVolumeSpecName: "host") pod "581ac32b-a24f-4def-aab5-a2123f5674b3" (UID: "581ac32b-a24f-4def-aab5-a2123f5674b3"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 07:32:11 crc kubenswrapper[4809]: I1206 07:32:11.168152 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-bn94p/crc-debug-dkcmj"] Dec 06 07:32:11 crc kubenswrapper[4809]: I1206 07:32:11.173802 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/581ac32b-a24f-4def-aab5-a2123f5674b3-kube-api-access-xkwt5" (OuterVolumeSpecName: "kube-api-access-xkwt5") pod "581ac32b-a24f-4def-aab5-a2123f5674b3" (UID: "581ac32b-a24f-4def-aab5-a2123f5674b3"). InnerVolumeSpecName "kube-api-access-xkwt5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:32:11 crc kubenswrapper[4809]: I1206 07:32:11.183824 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-bn94p/crc-debug-dkcmj"] Dec 06 07:32:11 crc kubenswrapper[4809]: I1206 07:32:11.268019 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xkwt5\" (UniqueName: \"kubernetes.io/projected/581ac32b-a24f-4def-aab5-a2123f5674b3-kube-api-access-xkwt5\") on node \"crc\" DevicePath \"\"" Dec 06 07:32:11 crc kubenswrapper[4809]: I1206 07:32:11.268054 4809 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/581ac32b-a24f-4def-aab5-a2123f5674b3-host\") on node \"crc\" DevicePath \"\"" Dec 06 07:32:11 crc kubenswrapper[4809]: I1206 07:32:11.403849 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="581ac32b-a24f-4def-aab5-a2123f5674b3" path="/var/lib/kubelet/pods/581ac32b-a24f-4def-aab5-a2123f5674b3/volumes" Dec 06 07:32:12 crc kubenswrapper[4809]: I1206 07:32:12.146239 4809 scope.go:117] "RemoveContainer" containerID="f2994613b8199bc578ca18d9e08b31c4952eea95c7c4653c21b9c1e9c9477b98" Dec 06 07:32:12 crc kubenswrapper[4809]: I1206 07:32:12.146278 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bn94p/crc-debug-dkcmj" Dec 06 07:32:12 crc kubenswrapper[4809]: I1206 07:32:12.390092 4809 scope.go:117] "RemoveContainer" containerID="26a5b8063411f48708112d11ebddb0587d0fe6fc8f6a5c0aa5728fedd5b56f1d" Dec 06 07:32:12 crc kubenswrapper[4809]: E1206 07:32:12.391007 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:32:12 crc kubenswrapper[4809]: I1206 07:32:12.720302 4809 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-qqv6x container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.30:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:32:12 crc kubenswrapper[4809]: I1206 07:32:12.720355 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qqv6x" podUID="8a225b5c-4ad5-4f08-a0cb-ae38a8fa83d4" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.30:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:32:14 crc kubenswrapper[4809]: I1206 07:32:14.335974 4809 patch_prober.go:28] interesting pod/logging-loki-gateway-65498c4f8f-48rjq container/opa namespace/openshift-logging: Liveness probe status=failure output="Get \"https://10.217.0.53:8083/live\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:32:14 crc kubenswrapper[4809]: I1206 07:32:14.336317 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-logging/logging-loki-gateway-65498c4f8f-48rjq" podUID="72a3ed95-3e3d-4faf-88e4-ad1731902910" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.53:8083/live\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:32:14 crc kubenswrapper[4809]: I1206 07:32:14.563862 4809 patch_prober.go:28] interesting pod/logging-loki-compactor-0 container/loki-compactor namespace/openshift-logging: Liveness probe status=failure output="Get \"https://10.217.0.57:3101/loki/api/v1/status/buildinfo\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:32:14 crc kubenswrapper[4809]: I1206 07:32:14.563961 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-logging/logging-loki-compactor-0" podUID="2365b183-7913-4ce1-94bc-b30705d653b2" containerName="loki-compactor" probeResult="failure" output="Get \"https://10.217.0.57:3101/loki/api/v1/status/buildinfo\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:32:15 crc kubenswrapper[4809]: I1206 07:32:15.300557 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-bn94p/crc-debug-wkkzk"] Dec 06 07:32:15 crc kubenswrapper[4809]: E1206 07:32:15.301387 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98ca3772-27e9-4eb9-8864-d9fc08c9fb42" containerName="registry-server" Dec 06 07:32:15 crc kubenswrapper[4809]: I1206 07:32:15.301404 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="98ca3772-27e9-4eb9-8864-d9fc08c9fb42" containerName="registry-server" Dec 06 07:32:15 crc kubenswrapper[4809]: E1206 07:32:15.301425 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98ca3772-27e9-4eb9-8864-d9fc08c9fb42" containerName="extract-content" Dec 06 07:32:15 crc kubenswrapper[4809]: I1206 07:32:15.301431 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="98ca3772-27e9-4eb9-8864-d9fc08c9fb42" containerName="extract-content" Dec 06 07:32:15 crc kubenswrapper[4809]: E1206 07:32:15.301449 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98ca3772-27e9-4eb9-8864-d9fc08c9fb42" containerName="extract-utilities" Dec 06 07:32:15 crc kubenswrapper[4809]: I1206 07:32:15.301456 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="98ca3772-27e9-4eb9-8864-d9fc08c9fb42" containerName="extract-utilities" Dec 06 07:32:15 crc kubenswrapper[4809]: E1206 07:32:15.301502 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="581ac32b-a24f-4def-aab5-a2123f5674b3" containerName="container-00" Dec 06 07:32:15 crc kubenswrapper[4809]: I1206 07:32:15.301510 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="581ac32b-a24f-4def-aab5-a2123f5674b3" containerName="container-00" Dec 06 07:32:15 crc kubenswrapper[4809]: I1206 07:32:15.301728 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="581ac32b-a24f-4def-aab5-a2123f5674b3" containerName="container-00" Dec 06 07:32:15 crc kubenswrapper[4809]: I1206 07:32:15.301751 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="98ca3772-27e9-4eb9-8864-d9fc08c9fb42" containerName="registry-server" Dec 06 07:32:15 crc kubenswrapper[4809]: I1206 07:32:15.302667 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bn94p/crc-debug-wkkzk" Dec 06 07:32:15 crc kubenswrapper[4809]: I1206 07:32:15.372497 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98vbh\" (UniqueName: \"kubernetes.io/projected/44eb3ca8-7510-4c4f-a0cc-56916afb903d-kube-api-access-98vbh\") pod \"crc-debug-wkkzk\" (UID: \"44eb3ca8-7510-4c4f-a0cc-56916afb903d\") " pod="openshift-must-gather-bn94p/crc-debug-wkkzk" Dec 06 07:32:15 crc kubenswrapper[4809]: I1206 07:32:15.373287 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/44eb3ca8-7510-4c4f-a0cc-56916afb903d-host\") pod \"crc-debug-wkkzk\" (UID: \"44eb3ca8-7510-4c4f-a0cc-56916afb903d\") " pod="openshift-must-gather-bn94p/crc-debug-wkkzk" Dec 06 07:32:15 crc kubenswrapper[4809]: I1206 07:32:15.475297 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98vbh\" (UniqueName: \"kubernetes.io/projected/44eb3ca8-7510-4c4f-a0cc-56916afb903d-kube-api-access-98vbh\") pod \"crc-debug-wkkzk\" (UID: \"44eb3ca8-7510-4c4f-a0cc-56916afb903d\") " pod="openshift-must-gather-bn94p/crc-debug-wkkzk" Dec 06 07:32:15 crc kubenswrapper[4809]: I1206 07:32:15.475681 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/44eb3ca8-7510-4c4f-a0cc-56916afb903d-host\") pod \"crc-debug-wkkzk\" (UID: \"44eb3ca8-7510-4c4f-a0cc-56916afb903d\") " pod="openshift-must-gather-bn94p/crc-debug-wkkzk" Dec 06 07:32:15 crc kubenswrapper[4809]: I1206 07:32:15.475894 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/44eb3ca8-7510-4c4f-a0cc-56916afb903d-host\") pod \"crc-debug-wkkzk\" (UID: \"44eb3ca8-7510-4c4f-a0cc-56916afb903d\") " pod="openshift-must-gather-bn94p/crc-debug-wkkzk" Dec 06 07:32:15 crc kubenswrapper[4809]: I1206 07:32:15.507953 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98vbh\" (UniqueName: \"kubernetes.io/projected/44eb3ca8-7510-4c4f-a0cc-56916afb903d-kube-api-access-98vbh\") pod \"crc-debug-wkkzk\" (UID: \"44eb3ca8-7510-4c4f-a0cc-56916afb903d\") " pod="openshift-must-gather-bn94p/crc-debug-wkkzk" Dec 06 07:32:15 crc kubenswrapper[4809]: I1206 07:32:15.632642 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bn94p/crc-debug-wkkzk" Dec 06 07:32:16 crc kubenswrapper[4809]: I1206 07:32:16.225173 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bn94p/crc-debug-wkkzk" event={"ID":"44eb3ca8-7510-4c4f-a0cc-56916afb903d","Type":"ContainerStarted","Data":"2a45806bc0aba25efade70d8027bf2bb2b7cd5cf89bf18b947be3b2c5226910a"} Dec 06 07:32:17 crc kubenswrapper[4809]: I1206 07:32:17.313225 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bn94p/crc-debug-wkkzk" event={"ID":"44eb3ca8-7510-4c4f-a0cc-56916afb903d","Type":"ContainerStarted","Data":"02aabf36b6a93c6a064da2230e09faaf3ccd6ce4facde411766574f66588e4ea"} Dec 06 07:32:17 crc kubenswrapper[4809]: I1206 07:32:17.332524 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-bn94p/crc-debug-wkkzk" podStartSLOduration=2.332499458 podStartE2EDuration="2.332499458s" podCreationTimestamp="2025-12-06 07:32:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:32:17.329967519 +0000 UTC m=+6062.218950491" watchObservedRunningTime="2025-12-06 07:32:17.332499458 +0000 UTC m=+6062.221482440" Dec 06 07:32:18 crc kubenswrapper[4809]: I1206 07:32:18.331321 4809 generic.go:334] "Generic (PLEG): container finished" podID="44eb3ca8-7510-4c4f-a0cc-56916afb903d" containerID="02aabf36b6a93c6a064da2230e09faaf3ccd6ce4facde411766574f66588e4ea" exitCode=0 Dec 06 07:32:18 crc kubenswrapper[4809]: I1206 07:32:18.331381 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bn94p/crc-debug-wkkzk" event={"ID":"44eb3ca8-7510-4c4f-a0cc-56916afb903d","Type":"ContainerDied","Data":"02aabf36b6a93c6a064da2230e09faaf3ccd6ce4facde411766574f66588e4ea"} Dec 06 07:32:19 crc kubenswrapper[4809]: I1206 07:32:19.482923 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bn94p/crc-debug-wkkzk" Dec 06 07:32:19 crc kubenswrapper[4809]: I1206 07:32:19.517459 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-bn94p/crc-debug-wkkzk"] Dec 06 07:32:19 crc kubenswrapper[4809]: I1206 07:32:19.528553 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-bn94p/crc-debug-wkkzk"] Dec 06 07:32:19 crc kubenswrapper[4809]: I1206 07:32:19.599058 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/44eb3ca8-7510-4c4f-a0cc-56916afb903d-host\") pod \"44eb3ca8-7510-4c4f-a0cc-56916afb903d\" (UID: \"44eb3ca8-7510-4c4f-a0cc-56916afb903d\") " Dec 06 07:32:19 crc kubenswrapper[4809]: I1206 07:32:19.599224 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/44eb3ca8-7510-4c4f-a0cc-56916afb903d-host" (OuterVolumeSpecName: "host") pod "44eb3ca8-7510-4c4f-a0cc-56916afb903d" (UID: "44eb3ca8-7510-4c4f-a0cc-56916afb903d"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 07:32:19 crc kubenswrapper[4809]: I1206 07:32:19.599989 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98vbh\" (UniqueName: \"kubernetes.io/projected/44eb3ca8-7510-4c4f-a0cc-56916afb903d-kube-api-access-98vbh\") pod \"44eb3ca8-7510-4c4f-a0cc-56916afb903d\" (UID: \"44eb3ca8-7510-4c4f-a0cc-56916afb903d\") " Dec 06 07:32:19 crc kubenswrapper[4809]: I1206 07:32:19.600849 4809 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/44eb3ca8-7510-4c4f-a0cc-56916afb903d-host\") on node \"crc\" DevicePath \"\"" Dec 06 07:32:19 crc kubenswrapper[4809]: I1206 07:32:19.616353 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44eb3ca8-7510-4c4f-a0cc-56916afb903d-kube-api-access-98vbh" (OuterVolumeSpecName: "kube-api-access-98vbh") pod "44eb3ca8-7510-4c4f-a0cc-56916afb903d" (UID: "44eb3ca8-7510-4c4f-a0cc-56916afb903d"). InnerVolumeSpecName "kube-api-access-98vbh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:32:19 crc kubenswrapper[4809]: I1206 07:32:19.712497 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98vbh\" (UniqueName: \"kubernetes.io/projected/44eb3ca8-7510-4c4f-a0cc-56916afb903d-kube-api-access-98vbh\") on node \"crc\" DevicePath \"\"" Dec 06 07:32:20 crc kubenswrapper[4809]: I1206 07:32:20.353656 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a45806bc0aba25efade70d8027bf2bb2b7cd5cf89bf18b947be3b2c5226910a" Dec 06 07:32:20 crc kubenswrapper[4809]: I1206 07:32:20.353706 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bn94p/crc-debug-wkkzk" Dec 06 07:32:32 crc kubenswrapper[4809]: I1206 07:32:20.683626 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-bn94p/crc-debug-ssjw9"] Dec 06 07:32:32 crc kubenswrapper[4809]: E1206 07:32:20.684128 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44eb3ca8-7510-4c4f-a0cc-56916afb903d" containerName="container-00" Dec 06 07:32:32 crc kubenswrapper[4809]: I1206 07:32:20.684140 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="44eb3ca8-7510-4c4f-a0cc-56916afb903d" containerName="container-00" Dec 06 07:32:32 crc kubenswrapper[4809]: I1206 07:32:20.684354 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="44eb3ca8-7510-4c4f-a0cc-56916afb903d" containerName="container-00" Dec 06 07:32:32 crc kubenswrapper[4809]: I1206 07:32:20.685110 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bn94p/crc-debug-ssjw9" Dec 06 07:32:32 crc kubenswrapper[4809]: I1206 07:32:20.841883 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ae8ced66-494a-46e2-b3a7-441b48607347-host\") pod \"crc-debug-ssjw9\" (UID: \"ae8ced66-494a-46e2-b3a7-441b48607347\") " pod="openshift-must-gather-bn94p/crc-debug-ssjw9" Dec 06 07:32:32 crc kubenswrapper[4809]: I1206 07:32:20.842007 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzjlg\" (UniqueName: \"kubernetes.io/projected/ae8ced66-494a-46e2-b3a7-441b48607347-kube-api-access-bzjlg\") pod \"crc-debug-ssjw9\" (UID: \"ae8ced66-494a-46e2-b3a7-441b48607347\") " pod="openshift-must-gather-bn94p/crc-debug-ssjw9" Dec 06 07:32:32 crc kubenswrapper[4809]: I1206 07:32:20.940153 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/frr-k8s-z95cs" podUID="a3b6197b-05d4-49b2-9a67-56b675046a1d" containerName="controller" probeResult="failure" output="Get \"http://127.0.0.1:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:32:32 crc kubenswrapper[4809]: I1206 07:32:20.944671 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ae8ced66-494a-46e2-b3a7-441b48607347-host\") pod \"crc-debug-ssjw9\" (UID: \"ae8ced66-494a-46e2-b3a7-441b48607347\") " pod="openshift-must-gather-bn94p/crc-debug-ssjw9" Dec 06 07:32:32 crc kubenswrapper[4809]: I1206 07:32:20.944791 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzjlg\" (UniqueName: \"kubernetes.io/projected/ae8ced66-494a-46e2-b3a7-441b48607347-kube-api-access-bzjlg\") pod \"crc-debug-ssjw9\" (UID: \"ae8ced66-494a-46e2-b3a7-441b48607347\") " pod="openshift-must-gather-bn94p/crc-debug-ssjw9" Dec 06 07:32:32 crc kubenswrapper[4809]: I1206 07:32:20.944820 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ae8ced66-494a-46e2-b3a7-441b48607347-host\") pod \"crc-debug-ssjw9\" (UID: \"ae8ced66-494a-46e2-b3a7-441b48607347\") " pod="openshift-must-gather-bn94p/crc-debug-ssjw9" Dec 06 07:32:32 crc kubenswrapper[4809]: I1206 07:32:20.964611 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzjlg\" (UniqueName: \"kubernetes.io/projected/ae8ced66-494a-46e2-b3a7-441b48607347-kube-api-access-bzjlg\") pod \"crc-debug-ssjw9\" (UID: \"ae8ced66-494a-46e2-b3a7-441b48607347\") " pod="openshift-must-gather-bn94p/crc-debug-ssjw9" Dec 06 07:32:32 crc kubenswrapper[4809]: I1206 07:32:21.012642 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bn94p/crc-debug-ssjw9" Dec 06 07:32:32 crc kubenswrapper[4809]: I1206 07:32:21.368303 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bn94p/crc-debug-ssjw9" event={"ID":"ae8ced66-494a-46e2-b3a7-441b48607347","Type":"ContainerStarted","Data":"8c5a8c131a2d044213c019d1bb37dba8b8997c03ecb8dcc6abb6db9cf417c01f"} Dec 06 07:32:32 crc kubenswrapper[4809]: I1206 07:32:21.407293 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44eb3ca8-7510-4c4f-a0cc-56916afb903d" path="/var/lib/kubelet/pods/44eb3ca8-7510-4c4f-a0cc-56916afb903d/volumes" Dec 06 07:32:33 crc kubenswrapper[4809]: I1206 07:32:33.052893 4809 scope.go:117] "RemoveContainer" containerID="26a5b8063411f48708112d11ebddb0587d0fe6fc8f6a5c0aa5728fedd5b56f1d" Dec 06 07:32:33 crc kubenswrapper[4809]: E1206 07:32:33.057631 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:32:37 crc kubenswrapper[4809]: I1206 07:32:37.935825 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-9gfg2" podUID="ee834860-9609-482d-95db-82fa1a1a6941" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.104:8081/readyz\": dial tcp 10.217.0.104:8081: i/o timeout (Client.Timeout exceeded while awaiting headers)" Dec 06 07:32:39 crc kubenswrapper[4809]: I1206 07:32:39.083014 4809 generic.go:334] "Generic (PLEG): container finished" podID="ae8ced66-494a-46e2-b3a7-441b48607347" containerID="c28c31075407ae9894aa2f2d33bb7851dcb503bdab12b4d209488f0e07870ad6" exitCode=0 Dec 06 07:32:39 crc kubenswrapper[4809]: I1206 07:32:39.083126 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bn94p/crc-debug-ssjw9" event={"ID":"ae8ced66-494a-46e2-b3a7-441b48607347","Type":"ContainerDied","Data":"c28c31075407ae9894aa2f2d33bb7851dcb503bdab12b4d209488f0e07870ad6"} Dec 06 07:32:39 crc kubenswrapper[4809]: I1206 07:32:39.122435 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-bn94p/crc-debug-ssjw9"] Dec 06 07:32:39 crc kubenswrapper[4809]: I1206 07:32:39.133622 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-bn94p/crc-debug-ssjw9"] Dec 06 07:32:40 crc kubenswrapper[4809]: I1206 07:32:40.947369 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bn94p/crc-debug-ssjw9" Dec 06 07:32:41 crc kubenswrapper[4809]: I1206 07:32:41.022401 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bzjlg\" (UniqueName: \"kubernetes.io/projected/ae8ced66-494a-46e2-b3a7-441b48607347-kube-api-access-bzjlg\") pod \"ae8ced66-494a-46e2-b3a7-441b48607347\" (UID: \"ae8ced66-494a-46e2-b3a7-441b48607347\") " Dec 06 07:32:41 crc kubenswrapper[4809]: I1206 07:32:41.022525 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ae8ced66-494a-46e2-b3a7-441b48607347-host\") pod \"ae8ced66-494a-46e2-b3a7-441b48607347\" (UID: \"ae8ced66-494a-46e2-b3a7-441b48607347\") " Dec 06 07:32:41 crc kubenswrapper[4809]: I1206 07:32:41.022568 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ae8ced66-494a-46e2-b3a7-441b48607347-host" (OuterVolumeSpecName: "host") pod "ae8ced66-494a-46e2-b3a7-441b48607347" (UID: "ae8ced66-494a-46e2-b3a7-441b48607347"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 07:32:41 crc kubenswrapper[4809]: I1206 07:32:41.023363 4809 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ae8ced66-494a-46e2-b3a7-441b48607347-host\") on node \"crc\" DevicePath \"\"" Dec 06 07:32:41 crc kubenswrapper[4809]: I1206 07:32:41.027186 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae8ced66-494a-46e2-b3a7-441b48607347-kube-api-access-bzjlg" (OuterVolumeSpecName: "kube-api-access-bzjlg") pod "ae8ced66-494a-46e2-b3a7-441b48607347" (UID: "ae8ced66-494a-46e2-b3a7-441b48607347"). InnerVolumeSpecName "kube-api-access-bzjlg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:32:41 crc kubenswrapper[4809]: I1206 07:32:41.107347 4809 scope.go:117] "RemoveContainer" containerID="c28c31075407ae9894aa2f2d33bb7851dcb503bdab12b4d209488f0e07870ad6" Dec 06 07:32:41 crc kubenswrapper[4809]: I1206 07:32:41.107375 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bn94p/crc-debug-ssjw9" Dec 06 07:32:41 crc kubenswrapper[4809]: I1206 07:32:41.124350 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bzjlg\" (UniqueName: \"kubernetes.io/projected/ae8ced66-494a-46e2-b3a7-441b48607347-kube-api-access-bzjlg\") on node \"crc\" DevicePath \"\"" Dec 06 07:32:41 crc kubenswrapper[4809]: I1206 07:32:41.403151 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae8ced66-494a-46e2-b3a7-441b48607347" path="/var/lib/kubelet/pods/ae8ced66-494a-46e2-b3a7-441b48607347/volumes" Dec 06 07:32:47 crc kubenswrapper[4809]: I1206 07:32:47.391743 4809 scope.go:117] "RemoveContainer" containerID="26a5b8063411f48708112d11ebddb0587d0fe6fc8f6a5c0aa5728fedd5b56f1d" Dec 06 07:32:47 crc kubenswrapper[4809]: E1206 07:32:47.392686 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:32:50 crc kubenswrapper[4809]: I1206 07:32:50.405189 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-ttn9g" podUID="b483382f-c2e3-40bf-ab4b-b58c2f3f9132" containerName="frr-k8s-webhook-server" probeResult="failure" output="Get \"http://10.217.0.94:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:32:50 crc kubenswrapper[4809]: I1206 07:32:50.405236 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-ttn9g" podUID="b483382f-c2e3-40bf-ab4b-b58c2f3f9132" containerName="frr-k8s-webhook-server" probeResult="failure" output="Get \"http://10.217.0.94:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:33:00 crc kubenswrapper[4809]: I1206 07:33:00.389415 4809 scope.go:117] "RemoveContainer" containerID="26a5b8063411f48708112d11ebddb0587d0fe6fc8f6a5c0aa5728fedd5b56f1d" Dec 06 07:33:00 crc kubenswrapper[4809]: E1206 07:33:00.392633 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:33:10 crc kubenswrapper[4809]: I1206 07:33:10.993748 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_7e16d0f3-9bd2-4887-84dd-3b8fb35d6451/aodh-api/0.log" Dec 06 07:33:11 crc kubenswrapper[4809]: I1206 07:33:11.575094 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_7e16d0f3-9bd2-4887-84dd-3b8fb35d6451/aodh-evaluator/0.log" Dec 06 07:33:11 crc kubenswrapper[4809]: I1206 07:33:11.587452 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_7e16d0f3-9bd2-4887-84dd-3b8fb35d6451/aodh-listener/0.log" Dec 06 07:33:11 crc kubenswrapper[4809]: I1206 07:33:11.592190 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_7e16d0f3-9bd2-4887-84dd-3b8fb35d6451/aodh-notifier/0.log" Dec 06 07:33:11 crc kubenswrapper[4809]: I1206 07:33:11.788328 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5bb7f9f986-k7477_7df2cd07-a4c8-4934-8755-7fa7a4c4b3fa/barbican-api/0.log" Dec 06 07:33:11 crc kubenswrapper[4809]: I1206 07:33:11.791830 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5bb7f9f986-k7477_7df2cd07-a4c8-4934-8755-7fa7a4c4b3fa/barbican-api-log/0.log" Dec 06 07:33:11 crc kubenswrapper[4809]: I1206 07:33:11.898541 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5cdb89b766-6mc5c_1d925ed8-47dc-4bf2-a476-2526d257d5d0/barbican-keystone-listener/0.log" Dec 06 07:33:12 crc kubenswrapper[4809]: I1206 07:33:12.097392 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5cdb89b766-6mc5c_1d925ed8-47dc-4bf2-a476-2526d257d5d0/barbican-keystone-listener-log/0.log" Dec 06 07:33:12 crc kubenswrapper[4809]: I1206 07:33:12.122748 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-646ffc98d7-5l7d6_79161677-35e4-41b3-8fb4-f2bd4beebe60/barbican-worker/0.log" Dec 06 07:33:12 crc kubenswrapper[4809]: I1206 07:33:12.178409 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-646ffc98d7-5l7d6_79161677-35e4-41b3-8fb4-f2bd4beebe60/barbican-worker-log/0.log" Dec 06 07:33:12 crc kubenswrapper[4809]: I1206 07:33:12.420494 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_b4a36ec8-2831-4f0f-b20c-137fa3d081c8/ceilometer-central-agent/0.log" Dec 06 07:33:12 crc kubenswrapper[4809]: I1206 07:33:12.513402 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-4c8nk_13f6bc16-e535-42af-953e-b6b3827b35d1/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 07:33:12 crc kubenswrapper[4809]: I1206 07:33:12.829540 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_b4a36ec8-2831-4f0f-b20c-137fa3d081c8/proxy-httpd/0.log" Dec 06 07:33:12 crc kubenswrapper[4809]: I1206 07:33:12.833060 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_b4a36ec8-2831-4f0f-b20c-137fa3d081c8/sg-core/0.log" Dec 06 07:33:12 crc kubenswrapper[4809]: I1206 07:33:12.864352 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_b4a36ec8-2831-4f0f-b20c-137fa3d081c8/ceilometer-notification-agent/0.log" Dec 06 07:33:13 crc kubenswrapper[4809]: I1206 07:33:13.112349 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_8a274a8e-452b-4087-92b1-c9bb713df5b5/cinder-api-log/0.log" Dec 06 07:33:13 crc kubenswrapper[4809]: I1206 07:33:13.207428 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_47aec31f-059f-4f3c-baef-4f9406e8cd4c/cinder-scheduler/1.log" Dec 06 07:33:13 crc kubenswrapper[4809]: I1206 07:33:13.230737 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_8a274a8e-452b-4087-92b1-c9bb713df5b5/cinder-api/0.log" Dec 06 07:33:13 crc kubenswrapper[4809]: I1206 07:33:13.385862 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_47aec31f-059f-4f3c-baef-4f9406e8cd4c/cinder-scheduler/0.log" Dec 06 07:33:13 crc kubenswrapper[4809]: I1206 07:33:13.466236 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_47aec31f-059f-4f3c-baef-4f9406e8cd4c/probe/0.log" Dec 06 07:33:13 crc kubenswrapper[4809]: I1206 07:33:13.484676 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-f4bfz_e2ad775c-a1f4-4770-918b-4b3ab597634d/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 07:33:13 crc kubenswrapper[4809]: I1206 07:33:13.782886 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f6df4f56c-2frbl_2558b055-d8d9-4309-acb2-addf23e33dcc/init/0.log" Dec 06 07:33:13 crc kubenswrapper[4809]: I1206 07:33:13.852118 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-6zmrk_c4012316-1bc7-4ca2-bdb8-d541646d6a9c/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 07:33:14 crc kubenswrapper[4809]: I1206 07:33:14.102723 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-xphmc_cb21c6c6-eaf8-440b-bccb-7a5a6b3b90ad/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 07:33:14 crc kubenswrapper[4809]: I1206 07:33:14.146487 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f6df4f56c-2frbl_2558b055-d8d9-4309-acb2-addf23e33dcc/init/0.log" Dec 06 07:33:14 crc kubenswrapper[4809]: I1206 07:33:14.285500 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f6df4f56c-2frbl_2558b055-d8d9-4309-acb2-addf23e33dcc/dnsmasq-dns/0.log" Dec 06 07:33:14 crc kubenswrapper[4809]: I1206 07:33:14.461228 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_807d7e0c-6c12-442b-8618-305e15340e02/glance-log/0.log" Dec 06 07:33:14 crc kubenswrapper[4809]: I1206 07:33:14.479221 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_807d7e0c-6c12-442b-8618-305e15340e02/glance-httpd/0.log" Dec 06 07:33:14 crc kubenswrapper[4809]: I1206 07:33:14.846435 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_0fdb3b6c-da8d-4493-b931-a55561dfb2ae/glance-log/0.log" Dec 06 07:33:14 crc kubenswrapper[4809]: I1206 07:33:14.876574 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_0fdb3b6c-da8d-4493-b931-a55561dfb2ae/glance-httpd/0.log" Dec 06 07:33:15 crc kubenswrapper[4809]: I1206 07:33:15.397558 4809 scope.go:117] "RemoveContainer" containerID="26a5b8063411f48708112d11ebddb0587d0fe6fc8f6a5c0aa5728fedd5b56f1d" Dec 06 07:33:15 crc kubenswrapper[4809]: E1206 07:33:15.397875 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:33:15 crc kubenswrapper[4809]: I1206 07:33:15.484938 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-7fcd5d8964-kw6sz_a71b0d8f-059d-4801-a2e4-6dc8ec44c78a/heat-engine/0.log" Dec 06 07:33:15 crc kubenswrapper[4809]: I1206 07:33:15.609975 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-7849fbcbcb-msd4w_dccd4eeb-4ba8-4200-8366-28d4ead2e476/heat-api/0.log" Dec 06 07:33:16 crc kubenswrapper[4809]: I1206 07:33:16.732137 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29416741-fwlb2_a360f0fd-86ee-4fab-bf60-aed62c57d106/keystone-cron/0.log" Dec 06 07:33:16 crc kubenswrapper[4809]: I1206 07:33:16.752100 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-p5xhb_c2a5ffc7-6399-4cca-b2b1-d73542617259/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 07:33:16 crc kubenswrapper[4809]: I1206 07:33:16.768175 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-p4jjg_c492d1c2-3ee8-491a-a237-98691bfacec6/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 07:33:16 crc kubenswrapper[4809]: I1206 07:33:16.780085 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="7c596e14-b96f-4e5f-a8ac-3dc66fa0276d" containerName="galera" probeResult="failure" output="command timed out" Dec 06 07:33:16 crc kubenswrapper[4809]: I1206 07:33:16.787089 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-galera-0" podUID="7c596e14-b96f-4e5f-a8ac-3dc66fa0276d" containerName="galera" probeResult="failure" output="command timed out" Dec 06 07:33:16 crc kubenswrapper[4809]: I1206 07:33:16.806306 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-794dfc48dc-pkjpx_b09a69de-0d5f-44f7-a3c9-2e6902da17b8/heat-cfnapi/0.log" Dec 06 07:33:16 crc kubenswrapper[4809]: I1206 07:33:16.963982 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_36f27192-451f-4ce2-afda-c0e66a6028b2/kube-state-metrics/1.log" Dec 06 07:33:17 crc kubenswrapper[4809]: I1206 07:33:17.058130 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_36f27192-451f-4ce2-afda-c0e66a6028b2/kube-state-metrics/0.log" Dec 06 07:33:17 crc kubenswrapper[4809]: I1206 07:33:17.288040 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-lmgn9_c4b57a32-8e04-4fbe-864d-387273af423f/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 07:33:17 crc kubenswrapper[4809]: I1206 07:33:17.359029 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_logging-edpm-deployment-openstack-edpm-ipam-crjgz_756d1854-be5a-421f-b620-ef10b83e7a61/logging-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 07:33:17 crc kubenswrapper[4809]: I1206 07:33:17.705835 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mysqld-exporter-0_f4b85ab2-e407-4779-bcc8-4ea2d35aee72/mysqld-exporter/0.log" Dec 06 07:33:18 crc kubenswrapper[4809]: I1206 07:33:18.098718 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-988458f5c-5zk66_6796bbd5-687e-4925-9ca1-700187c93ab8/neutron-httpd/0.log" Dec 06 07:33:18 crc kubenswrapper[4809]: I1206 07:33:18.253091 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-6bb4ccfdc4-9jqw5_b6e7307d-ec6c-4a4f-a253-b5e3f0123c0a/keystone-api/0.log" Dec 06 07:33:18 crc kubenswrapper[4809]: I1206 07:33:18.303650 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-988458f5c-5zk66_6796bbd5-687e-4925-9ca1-700187c93ab8/neutron-api/0.log" Dec 06 07:33:18 crc kubenswrapper[4809]: I1206 07:33:18.507319 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-mthc5_eb2b0f4f-7e1b-4e7d-a376-50a53f24879e/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 07:33:19 crc kubenswrapper[4809]: I1206 07:33:19.006650 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_fc561fa6-2f1c-41b6-8b04-b8aecf53d25a/nova-cell0-conductor-conductor/0.log" Dec 06 07:33:19 crc kubenswrapper[4809]: I1206 07:33:19.291049 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_61c6cc41-4c7d-422e-987c-e05e77ae463d/nova-api-log/0.log" Dec 06 07:33:19 crc kubenswrapper[4809]: I1206 07:33:19.452122 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_ae53b50b-d35c-4035-ab90-0c0575a95200/nova-cell1-conductor-conductor/0.log" Dec 06 07:33:19 crc kubenswrapper[4809]: I1206 07:33:19.662624 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_61c6cc41-4c7d-422e-987c-e05e77ae463d/nova-api-api/0.log" Dec 06 07:33:19 crc kubenswrapper[4809]: I1206 07:33:19.744846 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_663c54d8-faa9-466a-9126-678a92bc92f9/nova-cell1-novncproxy-novncproxy/0.log" Dec 06 07:33:19 crc kubenswrapper[4809]: I1206 07:33:19.785368 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-njt6f_ec2bf0dd-c16c-4912-a487-01711c8a0d84/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 07:33:20 crc kubenswrapper[4809]: I1206 07:33:20.025229 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_e0909590-844d-41ec-9443-5caacd682230/nova-metadata-log/0.log" Dec 06 07:33:20 crc kubenswrapper[4809]: I1206 07:33:20.748051 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_43f53993-991a-450e-a663-acd9f070d9b6/nova-scheduler-scheduler/0.log" Dec 06 07:33:20 crc kubenswrapper[4809]: I1206 07:33:20.764330 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e7056938-6e36-4631-adfb-01f116eba651/mysql-bootstrap/0.log" Dec 06 07:33:20 crc kubenswrapper[4809]: I1206 07:33:20.931371 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e7056938-6e36-4631-adfb-01f116eba651/mysql-bootstrap/0.log" Dec 06 07:33:21 crc kubenswrapper[4809]: I1206 07:33:21.021412 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e7056938-6e36-4631-adfb-01f116eba651/galera/0.log" Dec 06 07:33:21 crc kubenswrapper[4809]: I1206 07:33:21.042881 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e7056938-6e36-4631-adfb-01f116eba651/galera/1.log" Dec 06 07:33:21 crc kubenswrapper[4809]: I1206 07:33:21.294676 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_7c596e14-b96f-4e5f-a8ac-3dc66fa0276d/mysql-bootstrap/0.log" Dec 06 07:33:21 crc kubenswrapper[4809]: I1206 07:33:21.591849 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_7c596e14-b96f-4e5f-a8ac-3dc66fa0276d/mysql-bootstrap/0.log" Dec 06 07:33:21 crc kubenswrapper[4809]: I1206 07:33:21.625838 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_7c596e14-b96f-4e5f-a8ac-3dc66fa0276d/galera/1.log" Dec 06 07:33:21 crc kubenswrapper[4809]: I1206 07:33:21.641895 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_7c596e14-b96f-4e5f-a8ac-3dc66fa0276d/galera/0.log" Dec 06 07:33:21 crc kubenswrapper[4809]: I1206 07:33:21.947877 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_6c6578f3-bf7b-4892-abc9-5f4f145c8d36/openstackclient/0.log" Dec 06 07:33:22 crc kubenswrapper[4809]: I1206 07:33:22.121707 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-9rd86_9077c720-7ad1-4f5c-a525-e28acd3dccb8/openstack-network-exporter/0.log" Dec 06 07:33:22 crc kubenswrapper[4809]: I1206 07:33:22.265708 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jgqjw_92ce9234-d708-449e-9874-97ac8a4b994c/ovsdb-server-init/0.log" Dec 06 07:33:22 crc kubenswrapper[4809]: I1206 07:33:22.400548 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jgqjw_92ce9234-d708-449e-9874-97ac8a4b994c/ovsdb-server-init/0.log" Dec 06 07:33:22 crc kubenswrapper[4809]: I1206 07:33:22.478871 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jgqjw_92ce9234-d708-449e-9874-97ac8a4b994c/ovs-vswitchd/0.log" Dec 06 07:33:22 crc kubenswrapper[4809]: I1206 07:33:22.571328 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jgqjw_92ce9234-d708-449e-9874-97ac8a4b994c/ovsdb-server/0.log" Dec 06 07:33:22 crc kubenswrapper[4809]: I1206 07:33:22.752041 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-xhlzm_6b64391f-74d7-4b69-a898-f7375057c6d5/ovn-controller/0.log" Dec 06 07:33:22 crc kubenswrapper[4809]: I1206 07:33:22.933117 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_e0909590-844d-41ec-9443-5caacd682230/nova-metadata-metadata/0.log" Dec 06 07:33:22 crc kubenswrapper[4809]: I1206 07:33:22.993260 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-xm4wm_1fd335d6-64e4-4bb6-8617-4a08cfc65951/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 07:33:23 crc kubenswrapper[4809]: I1206 07:33:23.143097 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_1360c733-da74-4d37-8842-75050a356334/openstack-network-exporter/0.log" Dec 06 07:33:23 crc kubenswrapper[4809]: I1206 07:33:23.171590 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_1360c733-da74-4d37-8842-75050a356334/ovn-northd/0.log" Dec 06 07:33:23 crc kubenswrapper[4809]: I1206 07:33:23.370598 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_932f2f72-a070-425a-b065-64e293954cc5/openstack-network-exporter/0.log" Dec 06 07:33:23 crc kubenswrapper[4809]: I1206 07:33:23.459103 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_932f2f72-a070-425a-b065-64e293954cc5/ovsdbserver-nb/0.log" Dec 06 07:33:23 crc kubenswrapper[4809]: I1206 07:33:23.571726 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_1ffa5f54-0c40-45a6-9ebb-a5eca04810ee/openstack-network-exporter/0.log" Dec 06 07:33:23 crc kubenswrapper[4809]: I1206 07:33:23.706075 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_1ffa5f54-0c40-45a6-9ebb-a5eca04810ee/ovsdbserver-sb/0.log" Dec 06 07:33:24 crc kubenswrapper[4809]: I1206 07:33:24.012132 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5fc7d5fdc6-llxdt_b8ba7416-80be-4115-b960-58d147b56e9d/placement-log/0.log" Dec 06 07:33:24 crc kubenswrapper[4809]: I1206 07:33:24.329469 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5fc7d5fdc6-llxdt_b8ba7416-80be-4115-b960-58d147b56e9d/placement-api/0.log" Dec 06 07:33:24 crc kubenswrapper[4809]: I1206 07:33:24.375876 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_ca8e11eb-2da1-4706-b589-650bcb817d4c/init-config-reloader/0.log" Dec 06 07:33:24 crc kubenswrapper[4809]: I1206 07:33:24.581059 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_ca8e11eb-2da1-4706-b589-650bcb817d4c/init-config-reloader/0.log" Dec 06 07:33:24 crc kubenswrapper[4809]: I1206 07:33:24.626566 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_ca8e11eb-2da1-4706-b589-650bcb817d4c/config-reloader/0.log" Dec 06 07:33:24 crc kubenswrapper[4809]: I1206 07:33:24.635162 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_ca8e11eb-2da1-4706-b589-650bcb817d4c/thanos-sidecar/0.log" Dec 06 07:33:24 crc kubenswrapper[4809]: I1206 07:33:24.641852 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_ca8e11eb-2da1-4706-b589-650bcb817d4c/prometheus/0.log" Dec 06 07:33:24 crc kubenswrapper[4809]: I1206 07:33:24.847337 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_20646b41-62ab-42e7-8b11-538605db0506/setup-container/0.log" Dec 06 07:33:25 crc kubenswrapper[4809]: I1206 07:33:25.167397 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_20646b41-62ab-42e7-8b11-538605db0506/setup-container/0.log" Dec 06 07:33:25 crc kubenswrapper[4809]: I1206 07:33:25.224865 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_2cc53bf7-6e54-427a-9b63-88d694609c75/setup-container/0.log" Dec 06 07:33:25 crc kubenswrapper[4809]: I1206 07:33:25.253850 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_20646b41-62ab-42e7-8b11-538605db0506/rabbitmq/0.log" Dec 06 07:33:25 crc kubenswrapper[4809]: I1206 07:33:25.469587 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_2cc53bf7-6e54-427a-9b63-88d694609c75/setup-container/0.log" Dec 06 07:33:25 crc kubenswrapper[4809]: I1206 07:33:25.523662 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_2cc53bf7-6e54-427a-9b63-88d694609c75/rabbitmq/0.log" Dec 06 07:33:25 crc kubenswrapper[4809]: I1206 07:33:25.527136 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-7tb8w_e254b2e6-7bfa-4244-b1ce-d336ae05fa90/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 07:33:25 crc kubenswrapper[4809]: I1206 07:33:25.803876 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-5rgp6_362c4175-4f34-4343-9c19-f06e53132f6b/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 07:33:25 crc kubenswrapper[4809]: I1206 07:33:25.937059 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-s5cs4_55c8fbdf-1083-4893-9e7c-76f98c35d656/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 07:33:26 crc kubenswrapper[4809]: I1206 07:33:26.097599 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-mcdd9_896d8df1-4885-4ffa-a368-87b680cd909c/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 07:33:26 crc kubenswrapper[4809]: I1206 07:33:26.333076 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-4tq4x_5e4aa7e6-47ff-4f5f-abd9-4c4a193f2280/ssh-known-hosts-edpm-deployment/0.log" Dec 06 07:33:26 crc kubenswrapper[4809]: I1206 07:33:26.600760 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-58bb4f66d7-bdk65_d496af20-e3cf-4ce6-86f5-df841a9ba42c/proxy-server/0.log" Dec 06 07:33:26 crc kubenswrapper[4809]: I1206 07:33:26.610683 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-wgmjm_5f4f2323-f99a-4d85-8e0f-a0eb968de27b/swift-ring-rebalance/0.log" Dec 06 07:33:26 crc kubenswrapper[4809]: I1206 07:33:26.697446 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-58bb4f66d7-bdk65_d496af20-e3cf-4ce6-86f5-df841a9ba42c/proxy-httpd/0.log" Dec 06 07:33:26 crc kubenswrapper[4809]: I1206 07:33:26.848628 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_cfd13e5d-ca28-4061-8ca5-2c4566332bf0/account-auditor/0.log" Dec 06 07:33:26 crc kubenswrapper[4809]: I1206 07:33:26.950103 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_cfd13e5d-ca28-4061-8ca5-2c4566332bf0/account-reaper/0.log" Dec 06 07:33:27 crc kubenswrapper[4809]: I1206 07:33:27.099219 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_cfd13e5d-ca28-4061-8ca5-2c4566332bf0/account-replicator/0.log" Dec 06 07:33:27 crc kubenswrapper[4809]: I1206 07:33:27.125854 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_cfd13e5d-ca28-4061-8ca5-2c4566332bf0/account-server/0.log" Dec 06 07:33:27 crc kubenswrapper[4809]: I1206 07:33:27.183059 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_cfd13e5d-ca28-4061-8ca5-2c4566332bf0/container-auditor/0.log" Dec 06 07:33:27 crc kubenswrapper[4809]: I1206 07:33:27.262886 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_cfd13e5d-ca28-4061-8ca5-2c4566332bf0/container-replicator/0.log" Dec 06 07:33:27 crc kubenswrapper[4809]: I1206 07:33:27.379892 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_cfd13e5d-ca28-4061-8ca5-2c4566332bf0/container-server/0.log" Dec 06 07:33:27 crc kubenswrapper[4809]: I1206 07:33:27.402493 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_cfd13e5d-ca28-4061-8ca5-2c4566332bf0/container-updater/0.log" Dec 06 07:33:27 crc kubenswrapper[4809]: I1206 07:33:27.484708 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_cfd13e5d-ca28-4061-8ca5-2c4566332bf0/object-auditor/0.log" Dec 06 07:33:27 crc kubenswrapper[4809]: I1206 07:33:27.538502 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_cfd13e5d-ca28-4061-8ca5-2c4566332bf0/object-expirer/0.log" Dec 06 07:33:27 crc kubenswrapper[4809]: I1206 07:33:27.662231 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_cfd13e5d-ca28-4061-8ca5-2c4566332bf0/object-replicator/0.log" Dec 06 07:33:27 crc kubenswrapper[4809]: I1206 07:33:27.732555 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_cfd13e5d-ca28-4061-8ca5-2c4566332bf0/object-server/0.log" Dec 06 07:33:27 crc kubenswrapper[4809]: I1206 07:33:27.983170 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_cfd13e5d-ca28-4061-8ca5-2c4566332bf0/rsync/0.log" Dec 06 07:33:28 crc kubenswrapper[4809]: I1206 07:33:28.253606 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_cfd13e5d-ca28-4061-8ca5-2c4566332bf0/object-updater/0.log" Dec 06 07:33:28 crc kubenswrapper[4809]: I1206 07:33:28.453592 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_cfd13e5d-ca28-4061-8ca5-2c4566332bf0/swift-recon-cron/0.log" Dec 06 07:33:28 crc kubenswrapper[4809]: I1206 07:33:28.639047 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-2mswz_44cad7cb-41c6-453b-895d-7c560e19f84f/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 07:33:28 crc kubenswrapper[4809]: I1206 07:33:28.774426 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-power-monitoring-edpm-deployment-openstack-edpm-b66sw_0a499909-59b6-42ec-a9f7-75218a100ead/telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 07:33:29 crc kubenswrapper[4809]: I1206 07:33:29.077592 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_25b967d8-604b-4b77-9414-9f3922a56bdc/test-operator-logs-container/0.log" Dec 06 07:33:29 crc kubenswrapper[4809]: I1206 07:33:29.265268 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-jzdth_d24e0d36-0864-4ecc-9e52-f4998a338776/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 07:33:29 crc kubenswrapper[4809]: I1206 07:33:29.372482 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_973f75f0-3e24-4648-b57b-864065deb9b1/tempest-tests-tempest-tests-runner/0.log" Dec 06 07:33:30 crc kubenswrapper[4809]: I1206 07:33:30.389434 4809 scope.go:117] "RemoveContainer" containerID="26a5b8063411f48708112d11ebddb0587d0fe6fc8f6a5c0aa5728fedd5b56f1d" Dec 06 07:33:30 crc kubenswrapper[4809]: E1206 07:33:30.389759 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:33:31 crc kubenswrapper[4809]: I1206 07:33:31.223317 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_9d6c40a5-e2ff-4983-b9d8-fc483703870c/memcached/0.log" Dec 06 07:33:42 crc kubenswrapper[4809]: I1206 07:33:42.389544 4809 scope.go:117] "RemoveContainer" containerID="26a5b8063411f48708112d11ebddb0587d0fe6fc8f6a5c0aa5728fedd5b56f1d" Dec 06 07:33:42 crc kubenswrapper[4809]: E1206 07:33:42.390357 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:33:44 crc kubenswrapper[4809]: I1206 07:33:44.077351 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xk96f"] Dec 06 07:33:44 crc kubenswrapper[4809]: E1206 07:33:44.078273 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae8ced66-494a-46e2-b3a7-441b48607347" containerName="container-00" Dec 06 07:33:44 crc kubenswrapper[4809]: I1206 07:33:44.078288 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae8ced66-494a-46e2-b3a7-441b48607347" containerName="container-00" Dec 06 07:33:44 crc kubenswrapper[4809]: I1206 07:33:44.078590 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae8ced66-494a-46e2-b3a7-441b48607347" containerName="container-00" Dec 06 07:33:44 crc kubenswrapper[4809]: I1206 07:33:44.080524 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xk96f" Dec 06 07:33:44 crc kubenswrapper[4809]: I1206 07:33:44.091791 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xk96f"] Dec 06 07:33:44 crc kubenswrapper[4809]: I1206 07:33:44.188021 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f3c416e-5913-4f48-8192-c64e009042b6-catalog-content\") pod \"redhat-marketplace-xk96f\" (UID: \"7f3c416e-5913-4f48-8192-c64e009042b6\") " pod="openshift-marketplace/redhat-marketplace-xk96f" Dec 06 07:33:44 crc kubenswrapper[4809]: I1206 07:33:44.189170 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcngk\" (UniqueName: \"kubernetes.io/projected/7f3c416e-5913-4f48-8192-c64e009042b6-kube-api-access-rcngk\") pod \"redhat-marketplace-xk96f\" (UID: \"7f3c416e-5913-4f48-8192-c64e009042b6\") " pod="openshift-marketplace/redhat-marketplace-xk96f" Dec 06 07:33:44 crc kubenswrapper[4809]: I1206 07:33:44.189333 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f3c416e-5913-4f48-8192-c64e009042b6-utilities\") pod \"redhat-marketplace-xk96f\" (UID: \"7f3c416e-5913-4f48-8192-c64e009042b6\") " pod="openshift-marketplace/redhat-marketplace-xk96f" Dec 06 07:33:44 crc kubenswrapper[4809]: I1206 07:33:44.291808 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f3c416e-5913-4f48-8192-c64e009042b6-catalog-content\") pod \"redhat-marketplace-xk96f\" (UID: \"7f3c416e-5913-4f48-8192-c64e009042b6\") " pod="openshift-marketplace/redhat-marketplace-xk96f" Dec 06 07:33:44 crc kubenswrapper[4809]: I1206 07:33:44.291986 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcngk\" (UniqueName: \"kubernetes.io/projected/7f3c416e-5913-4f48-8192-c64e009042b6-kube-api-access-rcngk\") pod \"redhat-marketplace-xk96f\" (UID: \"7f3c416e-5913-4f48-8192-c64e009042b6\") " pod="openshift-marketplace/redhat-marketplace-xk96f" Dec 06 07:33:44 crc kubenswrapper[4809]: I1206 07:33:44.292041 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f3c416e-5913-4f48-8192-c64e009042b6-utilities\") pod \"redhat-marketplace-xk96f\" (UID: \"7f3c416e-5913-4f48-8192-c64e009042b6\") " pod="openshift-marketplace/redhat-marketplace-xk96f" Dec 06 07:33:44 crc kubenswrapper[4809]: I1206 07:33:44.292591 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f3c416e-5913-4f48-8192-c64e009042b6-utilities\") pod \"redhat-marketplace-xk96f\" (UID: \"7f3c416e-5913-4f48-8192-c64e009042b6\") " pod="openshift-marketplace/redhat-marketplace-xk96f" Dec 06 07:33:44 crc kubenswrapper[4809]: I1206 07:33:44.292638 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f3c416e-5913-4f48-8192-c64e009042b6-catalog-content\") pod \"redhat-marketplace-xk96f\" (UID: \"7f3c416e-5913-4f48-8192-c64e009042b6\") " pod="openshift-marketplace/redhat-marketplace-xk96f" Dec 06 07:33:44 crc kubenswrapper[4809]: I1206 07:33:44.352698 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcngk\" (UniqueName: \"kubernetes.io/projected/7f3c416e-5913-4f48-8192-c64e009042b6-kube-api-access-rcngk\") pod \"redhat-marketplace-xk96f\" (UID: \"7f3c416e-5913-4f48-8192-c64e009042b6\") " pod="openshift-marketplace/redhat-marketplace-xk96f" Dec 06 07:33:44 crc kubenswrapper[4809]: I1206 07:33:44.429919 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xk96f" Dec 06 07:33:44 crc kubenswrapper[4809]: I1206 07:33:44.992598 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xk96f"] Dec 06 07:33:45 crc kubenswrapper[4809]: I1206 07:33:45.198682 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xk96f" event={"ID":"7f3c416e-5913-4f48-8192-c64e009042b6","Type":"ContainerStarted","Data":"a8b5269d4625a6ac18b3fd5130f6a63f993430239ce220270e04a70f0816163d"} Dec 06 07:33:46 crc kubenswrapper[4809]: I1206 07:33:46.210663 4809 generic.go:334] "Generic (PLEG): container finished" podID="7f3c416e-5913-4f48-8192-c64e009042b6" containerID="fb08a9dbbd0cf1a1d9f78ca8b919319e9bfe4a3fd43c44943c3ad21d4ed4adb2" exitCode=0 Dec 06 07:33:46 crc kubenswrapper[4809]: I1206 07:33:46.210759 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xk96f" event={"ID":"7f3c416e-5913-4f48-8192-c64e009042b6","Type":"ContainerDied","Data":"fb08a9dbbd0cf1a1d9f78ca8b919319e9bfe4a3fd43c44943c3ad21d4ed4adb2"} Dec 06 07:33:46 crc kubenswrapper[4809]: I1206 07:33:46.213330 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 07:33:52 crc kubenswrapper[4809]: I1206 07:33:52.580478 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xk96f" event={"ID":"7f3c416e-5913-4f48-8192-c64e009042b6","Type":"ContainerStarted","Data":"087c40b2a2dbbd20795aef5b9b60c600f2c889112d20884cb05df2b6d26da700"} Dec 06 07:33:52 crc kubenswrapper[4809]: E1206 07:33:52.876204 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7f3c416e_5913_4f48_8192_c64e009042b6.slice/crio-conmon-087c40b2a2dbbd20795aef5b9b60c600f2c889112d20884cb05df2b6d26da700.scope\": RecentStats: unable to find data in memory cache]" Dec 06 07:33:53 crc kubenswrapper[4809]: I1206 07:33:53.592834 4809 generic.go:334] "Generic (PLEG): container finished" podID="7f3c416e-5913-4f48-8192-c64e009042b6" containerID="087c40b2a2dbbd20795aef5b9b60c600f2c889112d20884cb05df2b6d26da700" exitCode=0 Dec 06 07:33:53 crc kubenswrapper[4809]: I1206 07:33:53.592890 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xk96f" event={"ID":"7f3c416e-5913-4f48-8192-c64e009042b6","Type":"ContainerDied","Data":"087c40b2a2dbbd20795aef5b9b60c600f2c889112d20884cb05df2b6d26da700"} Dec 06 07:33:55 crc kubenswrapper[4809]: I1206 07:33:55.401485 4809 scope.go:117] "RemoveContainer" containerID="26a5b8063411f48708112d11ebddb0587d0fe6fc8f6a5c0aa5728fedd5b56f1d" Dec 06 07:33:55 crc kubenswrapper[4809]: E1206 07:33:55.402548 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:33:55 crc kubenswrapper[4809]: I1206 07:33:55.619479 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xk96f" event={"ID":"7f3c416e-5913-4f48-8192-c64e009042b6","Type":"ContainerStarted","Data":"85379d0b3c78afca1bfeedce3b1c1b833a1af574ae479521a01e5e000d329cfd"} Dec 06 07:33:55 crc kubenswrapper[4809]: I1206 07:33:55.647086 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xk96f" podStartSLOduration=3.894714242 podStartE2EDuration="11.647064029s" podCreationTimestamp="2025-12-06 07:33:44 +0000 UTC" firstStartedPulling="2025-12-06 07:33:46.212946601 +0000 UTC m=+6151.101929543" lastFinishedPulling="2025-12-06 07:33:53.965296388 +0000 UTC m=+6158.854279330" observedRunningTime="2025-12-06 07:33:55.637407588 +0000 UTC m=+6160.526390530" watchObservedRunningTime="2025-12-06 07:33:55.647064029 +0000 UTC m=+6160.536046971" Dec 06 07:34:00 crc kubenswrapper[4809]: I1206 07:34:00.759065 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2a8c7bb1b949e5c76441d571adf78af7c2049b664ecf2ad024c54dd9ba67kbg_99d85dd8-4137-4876-9062-3c23ee0bb730/util/0.log" Dec 06 07:34:01 crc kubenswrapper[4809]: I1206 07:34:01.023511 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2a8c7bb1b949e5c76441d571adf78af7c2049b664ecf2ad024c54dd9ba67kbg_99d85dd8-4137-4876-9062-3c23ee0bb730/pull/0.log" Dec 06 07:34:01 crc kubenswrapper[4809]: I1206 07:34:01.038152 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2a8c7bb1b949e5c76441d571adf78af7c2049b664ecf2ad024c54dd9ba67kbg_99d85dd8-4137-4876-9062-3c23ee0bb730/util/0.log" Dec 06 07:34:01 crc kubenswrapper[4809]: I1206 07:34:01.041476 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2a8c7bb1b949e5c76441d571adf78af7c2049b664ecf2ad024c54dd9ba67kbg_99d85dd8-4137-4876-9062-3c23ee0bb730/pull/0.log" Dec 06 07:34:01 crc kubenswrapper[4809]: I1206 07:34:01.213445 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2a8c7bb1b949e5c76441d571adf78af7c2049b664ecf2ad024c54dd9ba67kbg_99d85dd8-4137-4876-9062-3c23ee0bb730/pull/0.log" Dec 06 07:34:01 crc kubenswrapper[4809]: I1206 07:34:01.240887 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2a8c7bb1b949e5c76441d571adf78af7c2049b664ecf2ad024c54dd9ba67kbg_99d85dd8-4137-4876-9062-3c23ee0bb730/util/0.log" Dec 06 07:34:01 crc kubenswrapper[4809]: I1206 07:34:01.269569 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2a8c7bb1b949e5c76441d571adf78af7c2049b664ecf2ad024c54dd9ba67kbg_99d85dd8-4137-4876-9062-3c23ee0bb730/extract/0.log" Dec 06 07:34:01 crc kubenswrapper[4809]: I1206 07:34:01.485614 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-l8clc_44c524db-76cf-49c3-8c6e-1ad181d553ae/kube-rbac-proxy/0.log" Dec 06 07:34:01 crc kubenswrapper[4809]: I1206 07:34:01.555561 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-l8clc_44c524db-76cf-49c3-8c6e-1ad181d553ae/manager/0.log" Dec 06 07:34:01 crc kubenswrapper[4809]: I1206 07:34:01.596734 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-j4c62_7233cfab-96cb-4459-a507-3031cac37720/kube-rbac-proxy/0.log" Dec 06 07:34:01 crc kubenswrapper[4809]: I1206 07:34:01.767210 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-j4c62_7233cfab-96cb-4459-a507-3031cac37720/manager/0.log" Dec 06 07:34:01 crc kubenswrapper[4809]: I1206 07:34:01.797085 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-h69gw_8be493fb-5378-467d-9fde-d5154cda3207/kube-rbac-proxy/0.log" Dec 06 07:34:01 crc kubenswrapper[4809]: I1206 07:34:01.822105 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-h69gw_8be493fb-5378-467d-9fde-d5154cda3207/manager/0.log" Dec 06 07:34:02 crc kubenswrapper[4809]: I1206 07:34:02.095184 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-szvjh_204981d9-37b3-4584-b4bc-57df854143cc/manager/0.log" Dec 06 07:34:02 crc kubenswrapper[4809]: I1206 07:34:02.097022 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-szvjh_204981d9-37b3-4584-b4bc-57df854143cc/kube-rbac-proxy/0.log" Dec 06 07:34:02 crc kubenswrapper[4809]: I1206 07:34:02.309265 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-4s24n_35b44606-4fdd-4662-9550-adc5ed2478b3/kube-rbac-proxy/0.log" Dec 06 07:34:02 crc kubenswrapper[4809]: I1206 07:34:02.404067 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-9gfg2_ee834860-9609-482d-95db-82fa1a1a6941/kube-rbac-proxy/0.log" Dec 06 07:34:02 crc kubenswrapper[4809]: I1206 07:34:02.465922 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-4s24n_35b44606-4fdd-4662-9550-adc5ed2478b3/manager/0.log" Dec 06 07:34:02 crc kubenswrapper[4809]: I1206 07:34:02.549093 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-9gfg2_ee834860-9609-482d-95db-82fa1a1a6941/manager/0.log" Dec 06 07:34:02 crc kubenswrapper[4809]: I1206 07:34:02.726915 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-4b9l8_29c6d2cb-27fc-4b79-969a-462be41663e3/kube-rbac-proxy/0.log" Dec 06 07:34:02 crc kubenswrapper[4809]: I1206 07:34:02.788378 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-4b9l8_29c6d2cb-27fc-4b79-969a-462be41663e3/manager/1.log" Dec 06 07:34:03 crc kubenswrapper[4809]: I1206 07:34:03.053495 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-7trbt_45f87079-e852-4b47-a240-6f2bcf247509/kube-rbac-proxy/0.log" Dec 06 07:34:03 crc kubenswrapper[4809]: I1206 07:34:03.095521 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-7trbt_45f87079-e852-4b47-a240-6f2bcf247509/manager/0.log" Dec 06 07:34:03 crc kubenswrapper[4809]: I1206 07:34:03.162110 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-4b9l8_29c6d2cb-27fc-4b79-969a-462be41663e3/manager/0.log" Dec 06 07:34:03 crc kubenswrapper[4809]: I1206 07:34:03.386972 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-gzv6k_ced7591a-f80d-46a0-b8ea-024c2e37ae9f/kube-rbac-proxy/0.log" Dec 06 07:34:03 crc kubenswrapper[4809]: I1206 07:34:03.882622 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-vh8d4_93397035-a1a7-4c92-bdeb-5d0091dea181/kube-rbac-proxy/0.log" Dec 06 07:34:03 crc kubenswrapper[4809]: I1206 07:34:03.916262 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-gzv6k_ced7591a-f80d-46a0-b8ea-024c2e37ae9f/manager/0.log" Dec 06 07:34:04 crc kubenswrapper[4809]: I1206 07:34:04.173870 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-vh8d4_93397035-a1a7-4c92-bdeb-5d0091dea181/manager/0.log" Dec 06 07:34:04 crc kubenswrapper[4809]: I1206 07:34:04.217585 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-sxcr7_c37c165f-5cd1-496f-b092-41efdfcd7eca/kube-rbac-proxy/0.log" Dec 06 07:34:04 crc kubenswrapper[4809]: I1206 07:34:04.300058 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-sxcr7_c37c165f-5cd1-496f-b092-41efdfcd7eca/manager/0.log" Dec 06 07:34:04 crc kubenswrapper[4809]: I1206 07:34:04.431233 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xk96f" Dec 06 07:34:04 crc kubenswrapper[4809]: I1206 07:34:04.431286 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xk96f" Dec 06 07:34:04 crc kubenswrapper[4809]: I1206 07:34:04.494639 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-kcp8k_f736a8a4-ff7e-4179-951f-66405b1fd933/kube-rbac-proxy/0.log" Dec 06 07:34:04 crc kubenswrapper[4809]: I1206 07:34:04.499634 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xk96f" Dec 06 07:34:04 crc kubenswrapper[4809]: I1206 07:34:04.637869 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-kcp8k_f736a8a4-ff7e-4179-951f-66405b1fd933/manager/0.log" Dec 06 07:34:04 crc kubenswrapper[4809]: I1206 07:34:04.800724 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-27fdn_e6e0283a-55e9-4613-b74d-81fbb840690e/kube-rbac-proxy/0.log" Dec 06 07:34:04 crc kubenswrapper[4809]: I1206 07:34:04.836741 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xk96f" Dec 06 07:34:04 crc kubenswrapper[4809]: I1206 07:34:04.961868 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-27fdn_e6e0283a-55e9-4613-b74d-81fbb840690e/manager/0.log" Dec 06 07:34:04 crc kubenswrapper[4809]: I1206 07:34:04.975074 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xk96f"] Dec 06 07:34:05 crc kubenswrapper[4809]: I1206 07:34:05.080885 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-dfs6c_2ac245f3-8f64-4b6f-a51f-09e4b5811412/kube-rbac-proxy/0.log" Dec 06 07:34:05 crc kubenswrapper[4809]: I1206 07:34:05.148607 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-dfs6c_2ac245f3-8f64-4b6f-a51f-09e4b5811412/manager/0.log" Dec 06 07:34:05 crc kubenswrapper[4809]: I1206 07:34:05.338165 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd462lbf_6fbac224-e82d-482e-aeea-c83a9f978775/kube-rbac-proxy/0.log" Dec 06 07:34:05 crc kubenswrapper[4809]: I1206 07:34:05.516409 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd462lbf_6fbac224-e82d-482e-aeea-c83a9f978775/manager/0.log" Dec 06 07:34:05 crc kubenswrapper[4809]: I1206 07:34:05.947730 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6598f77bc9-xjfmg_2d9f04dc-3365-4b3e-b57c-604e657c6483/operator/0.log" Dec 06 07:34:05 crc kubenswrapper[4809]: I1206 07:34:05.972042 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-hv5l4_45b1df34-ffd2-42ca-85ec-885efb8ec7e5/registry-server/0.log" Dec 06 07:34:06 crc kubenswrapper[4809]: I1206 07:34:06.235047 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-ghwtq_b379307b-aafd-42da-b761-38a556a0d143/kube-rbac-proxy/0.log" Dec 06 07:34:06 crc kubenswrapper[4809]: I1206 07:34:06.375093 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-ghwtq_b379307b-aafd-42da-b761-38a556a0d143/manager/0.log" Dec 06 07:34:06 crc kubenswrapper[4809]: I1206 07:34:06.452361 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-zz2mm_b55361fa-8a7c-46ff-92f6-9dd72dbc0536/kube-rbac-proxy/0.log" Dec 06 07:34:06 crc kubenswrapper[4809]: I1206 07:34:06.547643 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-zz2mm_b55361fa-8a7c-46ff-92f6-9dd72dbc0536/manager/0.log" Dec 06 07:34:06 crc kubenswrapper[4809]: I1206 07:34:06.605123 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-848xh_0acb7c70-8d76-4bef-9d66-cd1213802a85/operator/0.log" Dec 06 07:34:06 crc kubenswrapper[4809]: I1206 07:34:06.783264 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xk96f" podUID="7f3c416e-5913-4f48-8192-c64e009042b6" containerName="registry-server" containerID="cri-o://85379d0b3c78afca1bfeedce3b1c1b833a1af574ae479521a01e5e000d329cfd" gracePeriod=2 Dec 06 07:34:06 crc kubenswrapper[4809]: I1206 07:34:06.809393 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-vqvvr_803d54f1-fbac-4ec5-b486-efae1981b46d/kube-rbac-proxy/0.log" Dec 06 07:34:06 crc kubenswrapper[4809]: I1206 07:34:06.841679 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-vqvvr_803d54f1-fbac-4ec5-b486-efae1981b46d/manager/0.log" Dec 06 07:34:06 crc kubenswrapper[4809]: I1206 07:34:06.989182 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-b4644b845-zxwvr_24aed9fd-d792-4bce-82a8-3ea4d70094b4/kube-rbac-proxy/0.log" Dec 06 07:34:07 crc kubenswrapper[4809]: I1206 07:34:07.267054 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-lqkvl_b376f569-0116-41f4-8a57-b5e5769e6a10/manager/1.log" Dec 06 07:34:07 crc kubenswrapper[4809]: I1206 07:34:07.308800 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-lqkvl_b376f569-0116-41f4-8a57-b5e5769e6a10/kube-rbac-proxy/0.log" Dec 06 07:34:07 crc kubenswrapper[4809]: I1206 07:34:07.448635 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xk96f" Dec 06 07:34:07 crc kubenswrapper[4809]: I1206 07:34:07.498556 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-lqkvl_b376f569-0116-41f4-8a57-b5e5769e6a10/manager/0.log" Dec 06 07:34:07 crc kubenswrapper[4809]: I1206 07:34:07.545240 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5c4b598d95-mbx56_4dd349b6-b792-451e-9cab-7c52a5768a3d/manager/0.log" Dec 06 07:34:07 crc kubenswrapper[4809]: I1206 07:34:07.563007 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rcngk\" (UniqueName: \"kubernetes.io/projected/7f3c416e-5913-4f48-8192-c64e009042b6-kube-api-access-rcngk\") pod \"7f3c416e-5913-4f48-8192-c64e009042b6\" (UID: \"7f3c416e-5913-4f48-8192-c64e009042b6\") " Dec 06 07:34:07 crc kubenswrapper[4809]: I1206 07:34:07.563205 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f3c416e-5913-4f48-8192-c64e009042b6-utilities\") pod \"7f3c416e-5913-4f48-8192-c64e009042b6\" (UID: \"7f3c416e-5913-4f48-8192-c64e009042b6\") " Dec 06 07:34:07 crc kubenswrapper[4809]: I1206 07:34:07.563449 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f3c416e-5913-4f48-8192-c64e009042b6-catalog-content\") pod \"7f3c416e-5913-4f48-8192-c64e009042b6\" (UID: \"7f3c416e-5913-4f48-8192-c64e009042b6\") " Dec 06 07:34:07 crc kubenswrapper[4809]: I1206 07:34:07.564425 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f3c416e-5913-4f48-8192-c64e009042b6-utilities" (OuterVolumeSpecName: "utilities") pod "7f3c416e-5913-4f48-8192-c64e009042b6" (UID: "7f3c416e-5913-4f48-8192-c64e009042b6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:34:07 crc kubenswrapper[4809]: I1206 07:34:07.569398 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f3c416e-5913-4f48-8192-c64e009042b6-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 07:34:07 crc kubenswrapper[4809]: I1206 07:34:07.572099 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f3c416e-5913-4f48-8192-c64e009042b6-kube-api-access-rcngk" (OuterVolumeSpecName: "kube-api-access-rcngk") pod "7f3c416e-5913-4f48-8192-c64e009042b6" (UID: "7f3c416e-5913-4f48-8192-c64e009042b6"). InnerVolumeSpecName "kube-api-access-rcngk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:34:07 crc kubenswrapper[4809]: I1206 07:34:07.626528 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f3c416e-5913-4f48-8192-c64e009042b6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7f3c416e-5913-4f48-8192-c64e009042b6" (UID: "7f3c416e-5913-4f48-8192-c64e009042b6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:34:07 crc kubenswrapper[4809]: I1206 07:34:07.637124 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-b4644b845-zxwvr_24aed9fd-d792-4bce-82a8-3ea4d70094b4/manager/0.log" Dec 06 07:34:07 crc kubenswrapper[4809]: I1206 07:34:07.671057 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f3c416e-5913-4f48-8192-c64e009042b6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 07:34:07 crc kubenswrapper[4809]: I1206 07:34:07.671109 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rcngk\" (UniqueName: \"kubernetes.io/projected/7f3c416e-5913-4f48-8192-c64e009042b6-kube-api-access-rcngk\") on node \"crc\" DevicePath \"\"" Dec 06 07:34:07 crc kubenswrapper[4809]: I1206 07:34:07.683607 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-bhbcl_fd7dd0a5-5e39-4812-b93f-a27c39284d33/kube-rbac-proxy/0.log" Dec 06 07:34:07 crc kubenswrapper[4809]: I1206 07:34:07.775151 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-bhbcl_fd7dd0a5-5e39-4812-b93f-a27c39284d33/manager/0.log" Dec 06 07:34:07 crc kubenswrapper[4809]: I1206 07:34:07.801196 4809 generic.go:334] "Generic (PLEG): container finished" podID="7f3c416e-5913-4f48-8192-c64e009042b6" containerID="85379d0b3c78afca1bfeedce3b1c1b833a1af574ae479521a01e5e000d329cfd" exitCode=0 Dec 06 07:34:07 crc kubenswrapper[4809]: I1206 07:34:07.801251 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xk96f" event={"ID":"7f3c416e-5913-4f48-8192-c64e009042b6","Type":"ContainerDied","Data":"85379d0b3c78afca1bfeedce3b1c1b833a1af574ae479521a01e5e000d329cfd"} Dec 06 07:34:07 crc kubenswrapper[4809]: I1206 07:34:07.801279 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xk96f" event={"ID":"7f3c416e-5913-4f48-8192-c64e009042b6","Type":"ContainerDied","Data":"a8b5269d4625a6ac18b3fd5130f6a63f993430239ce220270e04a70f0816163d"} Dec 06 07:34:07 crc kubenswrapper[4809]: I1206 07:34:07.801296 4809 scope.go:117] "RemoveContainer" containerID="85379d0b3c78afca1bfeedce3b1c1b833a1af574ae479521a01e5e000d329cfd" Dec 06 07:34:07 crc kubenswrapper[4809]: I1206 07:34:07.801381 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xk96f" Dec 06 07:34:07 crc kubenswrapper[4809]: I1206 07:34:07.834629 4809 scope.go:117] "RemoveContainer" containerID="087c40b2a2dbbd20795aef5b9b60c600f2c889112d20884cb05df2b6d26da700" Dec 06 07:34:07 crc kubenswrapper[4809]: I1206 07:34:07.862313 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xk96f"] Dec 06 07:34:07 crc kubenswrapper[4809]: I1206 07:34:07.876296 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xk96f"] Dec 06 07:34:07 crc kubenswrapper[4809]: I1206 07:34:07.883442 4809 scope.go:117] "RemoveContainer" containerID="fb08a9dbbd0cf1a1d9f78ca8b919319e9bfe4a3fd43c44943c3ad21d4ed4adb2" Dec 06 07:34:07 crc kubenswrapper[4809]: I1206 07:34:07.951310 4809 scope.go:117] "RemoveContainer" containerID="85379d0b3c78afca1bfeedce3b1c1b833a1af574ae479521a01e5e000d329cfd" Dec 06 07:34:07 crc kubenswrapper[4809]: E1206 07:34:07.953593 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85379d0b3c78afca1bfeedce3b1c1b833a1af574ae479521a01e5e000d329cfd\": container with ID starting with 85379d0b3c78afca1bfeedce3b1c1b833a1af574ae479521a01e5e000d329cfd not found: ID does not exist" containerID="85379d0b3c78afca1bfeedce3b1c1b833a1af574ae479521a01e5e000d329cfd" Dec 06 07:34:07 crc kubenswrapper[4809]: I1206 07:34:07.953681 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85379d0b3c78afca1bfeedce3b1c1b833a1af574ae479521a01e5e000d329cfd"} err="failed to get container status \"85379d0b3c78afca1bfeedce3b1c1b833a1af574ae479521a01e5e000d329cfd\": rpc error: code = NotFound desc = could not find container \"85379d0b3c78afca1bfeedce3b1c1b833a1af574ae479521a01e5e000d329cfd\": container with ID starting with 85379d0b3c78afca1bfeedce3b1c1b833a1af574ae479521a01e5e000d329cfd not found: ID does not exist" Dec 06 07:34:07 crc kubenswrapper[4809]: I1206 07:34:07.953724 4809 scope.go:117] "RemoveContainer" containerID="087c40b2a2dbbd20795aef5b9b60c600f2c889112d20884cb05df2b6d26da700" Dec 06 07:34:07 crc kubenswrapper[4809]: E1206 07:34:07.954235 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"087c40b2a2dbbd20795aef5b9b60c600f2c889112d20884cb05df2b6d26da700\": container with ID starting with 087c40b2a2dbbd20795aef5b9b60c600f2c889112d20884cb05df2b6d26da700 not found: ID does not exist" containerID="087c40b2a2dbbd20795aef5b9b60c600f2c889112d20884cb05df2b6d26da700" Dec 06 07:34:07 crc kubenswrapper[4809]: I1206 07:34:07.954272 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"087c40b2a2dbbd20795aef5b9b60c600f2c889112d20884cb05df2b6d26da700"} err="failed to get container status \"087c40b2a2dbbd20795aef5b9b60c600f2c889112d20884cb05df2b6d26da700\": rpc error: code = NotFound desc = could not find container \"087c40b2a2dbbd20795aef5b9b60c600f2c889112d20884cb05df2b6d26da700\": container with ID starting with 087c40b2a2dbbd20795aef5b9b60c600f2c889112d20884cb05df2b6d26da700 not found: ID does not exist" Dec 06 07:34:07 crc kubenswrapper[4809]: I1206 07:34:07.954305 4809 scope.go:117] "RemoveContainer" containerID="fb08a9dbbd0cf1a1d9f78ca8b919319e9bfe4a3fd43c44943c3ad21d4ed4adb2" Dec 06 07:34:07 crc kubenswrapper[4809]: E1206 07:34:07.954785 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb08a9dbbd0cf1a1d9f78ca8b919319e9bfe4a3fd43c44943c3ad21d4ed4adb2\": container with ID starting with fb08a9dbbd0cf1a1d9f78ca8b919319e9bfe4a3fd43c44943c3ad21d4ed4adb2 not found: ID does not exist" containerID="fb08a9dbbd0cf1a1d9f78ca8b919319e9bfe4a3fd43c44943c3ad21d4ed4adb2" Dec 06 07:34:07 crc kubenswrapper[4809]: I1206 07:34:07.954891 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb08a9dbbd0cf1a1d9f78ca8b919319e9bfe4a3fd43c44943c3ad21d4ed4adb2"} err="failed to get container status \"fb08a9dbbd0cf1a1d9f78ca8b919319e9bfe4a3fd43c44943c3ad21d4ed4adb2\": rpc error: code = NotFound desc = could not find container \"fb08a9dbbd0cf1a1d9f78ca8b919319e9bfe4a3fd43c44943c3ad21d4ed4adb2\": container with ID starting with fb08a9dbbd0cf1a1d9f78ca8b919319e9bfe4a3fd43c44943c3ad21d4ed4adb2 not found: ID does not exist" Dec 06 07:34:08 crc kubenswrapper[4809]: I1206 07:34:08.393952 4809 scope.go:117] "RemoveContainer" containerID="26a5b8063411f48708112d11ebddb0587d0fe6fc8f6a5c0aa5728fedd5b56f1d" Dec 06 07:34:08 crc kubenswrapper[4809]: E1206 07:34:08.394618 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:34:09 crc kubenswrapper[4809]: I1206 07:34:09.402774 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f3c416e-5913-4f48-8192-c64e009042b6" path="/var/lib/kubelet/pods/7f3c416e-5913-4f48-8192-c64e009042b6/volumes" Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.581566 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.592003136s: [/var/lib/containers/storage/overlay/9f081b4cdafd62bfd4ae0ecf48993e6cffcac661fc262ae670b81c270d0068a8/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.583851 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.592417586s: [/var/lib/containers/storage/overlay/dc5f5e365032626a70021d5d8d02c79ce9620283233c36b989c76f73900bd82e/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.583945 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.592515629s: [/var/lib/containers/storage/overlay/e7da9dd8b91d5a076e582127bd5c23cdf89c1e77c81734f75cb9a9b1293d5c2d/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.584004 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.592591231s: [/var/lib/containers/storage/overlay/fcf27b9281c2c346ef327449e0724b512ecba2dd6f09b7776d6e5460a5da1dfd/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.584064 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.592640423s: [/var/lib/containers/storage/overlay/b0f6b255f02e7ef8463772a1890e4e979d30b0bed5cdb05f810d0cbd396c8240/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.584118 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.592688294s: [/var/lib/containers/storage/overlay/ac149503a1b223fa5dcc1d6d35855f38b3c21ba485a8ffafc5143a0f86f8fac9/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.584175 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.592744066s: [/var/lib/containers/storage/overlay/11c1694f27389f3f8c4b48f80fd83d7290995a381914259af128249fe3da0c21/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.584248 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.592811007s: [/var/lib/containers/storage/overlay/76ff29959b93dbfa98c18e8ede2ea549f09ec4317bd30eeb72f091fa9ce45f8a/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.586382 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.571609275s: [/var/lib/containers/storage/overlay/1a11852c0b594aab35edae6649bab2a132428dc5d876a37351dea8b7dde2661e/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.586455 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.571681287s: [/var/lib/containers/storage/overlay/7520c3dbd50f39ae79fff39586c58ac02c666b53bca5eab16bf46906c72ea531/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.586503 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.571672377s: [/var/lib/containers/storage/overlay/339b81b35bfd3820a5e79fc5971239d42db89dde6bbfa5f7f87ebaab6eaf89b5/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.586546 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.571714318s: [/var/lib/containers/storage/overlay/e65ab38f5caa4db2ce2ff40811f1e5631f1c7843db7ce69b2a6b7948da87c364/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.586597 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.571752238s: [/var/lib/containers/storage/overlay/d1a1d19351cf5540cadc98dc691df48732bc700acb6a31f082c7cf9adb84d8a3/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.586642 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.57179549s: [/var/lib/containers/storage/overlay/d491dcd76e04d7006574ae08d91e9e368dd03cb9ca78c80e52631080ed693027/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.586724 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.571876312s: [/var/lib/containers/storage/overlay/03dd75bdccd3f8aca08419c6f4390d4c9b3596fcab7a603d752514fb563c1424/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.586817 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.571967054s: [/var/lib/containers/storage/overlay/bc2ca114ea829350f93d42235847c4d603cd24edb4dd62d2a0e968fa067fe4d2/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.586877 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.572024836s: [/var/lib/containers/storage/overlay/6275ce95dee95ac65d8719d7f8696d2bbfcb7e18c48f8ae2c5e51a150307188f/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.586927 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.572074527s: [/var/lib/containers/storage/overlay/c2e0cf181806a2b7dc784acfefca7285b7bfeca1a2280942cd5f7e214d9f0145/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.587002 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.572119129s: [/var/lib/containers/storage/overlay/c434454caab6eb85a8dd914e4612672070984c65da3d2baa42ff8814ffd263db/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.587048 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.572161819s: [/var/lib/containers/storage/overlay/cd4c1aa9edb73fc4583c74e9845d514f8a7dbca590f261974d97a236687607cc/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.588135 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.598603334s: [/var/lib/containers/storage/overlay/91a577da20ddf648ff99c205ee261aafa3cb5943d61d4e20ecf8649ec54a864d/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.572349 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.583325161s: [/var/lib/containers/storage/overlay/5aefe49886952e2912bb25a715bd92341eb10245f6f85b5f8d3f9df950289454/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.591262 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.602684265s: [/var/lib/containers/storage/overlay/f76c7289b34932eeae359412fd2a5f006f45bba928d8f85d33f0309c38f2164e/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.591323 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.576426765s: [/var/lib/containers/storage/overlay/36fe610433214ba0f70c3a2ca0b4aa68ed85d240dba52243048bbea2f78e2692/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.591355 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.576454606s: [/var/lib/containers/storage/overlay/2c36f5b96940045809d2b743a8da73428b584674daf26621c7431d364c77db10/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.591386 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.576479336s: [/var/lib/containers/storage/overlay/bd2b194b86c4a9cab6ce7b026b75709caa5ef653d438edd74f0a4c75c4a3dd00/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.591418 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.576509337s: [/var/lib/containers/storage/overlay/49e611c85ddc26a0e4ae0299c0b204223de31491ba11a3c9745cb581227d13c9/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.591453 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.572305363s: [/var/lib/containers/storage/overlay/a3118b8a010a0954e51e52e2924e1c7b1c79174822af3af206c05a65b2dbc03e/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.591485 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.572282793s: [/var/lib/containers/storage/overlay/ee56cf651013c0eec70b7140b7876d6d321bf4c1ab6b3ceaf7859a55c710fbc8/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.591516 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.572313314s: [/var/lib/containers/storage/overlay/c38e32d0b22089bd5fe99b47bdcca9d3c4955812be946558b7f3d266e1f45551/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.591547 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.572342415s: [/var/lib/containers/storage/overlay/0c7284ef0d9bd917d249b4099879ab0e3c20f45925fd82b22f98547b0bb37780/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.591578 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.572372226s: [/var/lib/containers/storage/overlay/ddbd23d45f7d368a79b0fba8f6717e4e8b73e1760c50536fa0de30a59ac59399/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.591614 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.572404166s: [/var/lib/containers/storage/overlay/d8baea43bf896fdb9f2137bacd2601e098bd23dfeb3563314fa83ecbd17c1c8e/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.591653 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.572437246s: [/var/lib/containers/storage/overlay/6599b263cc0f71bf42bf06fd0a60ae19c0a0ae72483e7c309ac0160730093973/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.591691 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.572474137s: [/var/lib/containers/storage/overlay/08ff7f0784024879990e62c96bb2ab7868f32624fbe54bf739319ce5f4760d57/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.591961 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.572736416s: [/var/lib/containers/storage/overlay/2a4df73577d6a5319466e831ae0d0ac71d4a83a7378d4e930b9dc90984f232be/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.592007 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.572777817s: [/var/lib/containers/storage/overlay/c6c61623f0caaac5874b3401400fd9d3e5e33a984212dbb44e673a71e09cb55e/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.592059 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.572801977s: [/var/lib/containers/storage/overlay/7d191e945aeb3ddb997af0f733128b2f454a1fe758428f0422480cd796e8e2f0/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.592093 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.568452359s: [/var/lib/containers/storage/overlay/712acb314ec446470161b332c915bc4cfa1dad6af443e2f4745638955dbd7b49/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.592126 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.568455099s: [/var/lib/containers/storage/overlay/7ef12fb628d26068785151e14f99e58183087bdfdb46e16636430009dd50dcba/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.592156 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.56848052s: [/var/lib/containers/storage/overlay/b210948f6e9c35efc62ec946793afb481dff58485df5ba587b2228acdde7cb93/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.592185 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.568505041s: [/var/lib/containers/storage/overlay/d529f690604398d871ccd34fb267bda17fe9bf577222668d561a1e7c46299676/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.592570 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.568879851s: [/var/lib/containers/storage/overlay/0ac05f55d54cebf92f556c8bef1073f1159045bcfdb2a65fe9fe065449d58c3a/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.592605 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.568913052s: [/var/lib/containers/storage/overlay/a7c83f69d6d73b8bb355a3cdf89555e9d443a5a82cde771015cfdfb629a3371d/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.592635 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.557614026s: [/var/lib/containers/storage/overlay/39aed6856b96b28c4babc4846b5f4ae21d4cd7f1f4e17f64ac0985712e252f7d/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.592666 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.557637036s: [/var/lib/containers/storage/overlay/5889a39d2ba3c1c40b408e79f71bf1068169d591a46061786404691e0da0e3cd/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.592697 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.557665517s: [/var/lib/containers/storage/overlay/87971d8ff1d8eba3cd354f12e242584d5a499a7a4b497fef8f25f17d0f67fba7/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.592727 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.557652097s: [/var/lib/containers/storage/overlay/6488caa59886e87c50397d4a4db04e4e900453319ed9141198c481577860670c/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.592757 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.557680888s: [/var/lib/containers/storage/overlay/1cf427d88c92f17784265e00abb900731c4d243c092946508ec08d7c0db96366/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.592787 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.557703558s: [/var/lib/containers/storage/overlay/0ff4c0878af9063d05fd714133b6e41da5c6a0d773916dfd1288f0d1c55a3365/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.592885 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.557799081s: [/var/lib/containers/storage/overlay/fed3bad6ca66d1e44d716fe203aa1272239da2cc0b19495fab925b91d6868f7c/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.592922 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.555681124s: [/var/lib/containers/storage/overlay/4fdc1ca14f688d9a5165678a07b9bb2f9f3d96dd04aa3923f255e6789966d60c/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.592969 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.550684908s: [/var/lib/containers/storage/overlay/17dceee5a4519081a02a2e6785a3337796c52886e857c4dcb5cec00e1e4c77e5/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.593005 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.550712549s: [/var/lib/containers/storage/overlay/e6d146802fa7cd7f1a40977204afe3e661ffcffb27a5c29d091f4c433f9bc3ce/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.592990 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-vqvvr" podUID="803d54f1-fbac-4ec5-b486-efae1981b46d" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.115:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.593034 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.549660351s: [/var/lib/containers/storage/overlay/6648a02da002fd611752facba23edccde57ef3d4d5ee71f48f20132b22a1e62d/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.594455 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.579552979s: [/var/lib/containers/storage/overlay/9b95fac6bb05ca0306c5927366dbbf875d91038e9a04b64e384c30f1cb22c154/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.595343 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.607888764s: [/var/lib/containers/storage/overlay/5e1db95de6a6971c80d10ba01c37cc1c0a1b6fa5e133ae85188bb3f9354b3cb6/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.596600 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.60802138s: [/var/lib/containers/storage/overlay/97675ed24bac7b5458c9e2bdc7b3999c763451e564ad9784151adf2e1954ad51/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.596968 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.597941516s: [/var/lib/containers/storage/overlay/b4f519fa52ccbb2ae3e18259da8987bc875b758c625e52161f1b241e9b073d25/diff /var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-848xh_0acb7c70-8d76-4bef-9d66-cd1213802a85/operator/0.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.597017 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.577794392s: [/var/lib/containers/storage/overlay/44de04e06bc653642202bdee2aa41f59f56b5b209afc4802799a7a621a1e22b5/diff /var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-7vvw8_886a998b-a62e-412c-ae50-fd1ba3a4a098/cert-manager-webhook/0.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.597041 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.573353451s: [/var/lib/containers/storage/overlay/564f4bfd442dd61d6ffd76187485cbd2a0ed8fc69a1d1eb642b83b4e7c6ffe80/diff /var/log/pods/openstack_swift-storage-0_cfd13e5d-ca28-4061-8ca5-2c4566332bf0/account-reaper/0.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.597067 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.495674032s: [/var/lib/containers/storage/overlay/998ca804479d2e83b94503a85eb5582f00f9606e3e341e9729237e844dedac6e/diff /var/log/pods/openshift-monitoring_prometheus-operator-db54df47d-lbjmx_a2c650ac-1257-460a-98c0-9a7db19710e1/prometheus-operator/0.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.597090 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.487331777s: [/var/lib/containers/storage/overlay/bf4ac666b2babba0f0612c0959f3d8e01c1cb903430ee25a83b1128b144fb1ae/diff /var/log/pods/openstack_ovsdbserver-nb-0_932f2f72-a070-425a-b065-64e293954cc5/openstack-network-exporter/0.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.597116 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.487349787s: [/var/lib/containers/storage/overlay/0433e2bb08037ae472238733144e1c2f537c3d8191b8000bd384f4e26283e8f7/diff /var/log/pods/openstack_ovsdbserver-sb-0_1ffa5f54-0c40-45a6-9ebb-a5eca04810ee/openstack-network-exporter/0.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.597120 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.374277191s: [/var/lib/containers/storage/overlay/b49d828c773db81b03246701811c6cc055477ed7a1a6522360bc6b5027fdd741/diff /var/log/pods/openshift-monitoring_prometheus-operator-db54df47d-lbjmx_a2c650ac-1257-460a-98c0-9a7db19710e1/kube-rbac-proxy/0.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.597210 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.607656459s: [/var/lib/containers/storage/overlay/47033ab68e9dcd550901cabb1ea42aa3c328fa1e881356f243b45ae7094d18f3/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.597597 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.610159857s: [/var/lib/containers/storage/overlay/fe16ae70bac191eb7fc8e707dac16eb0f2225544e78bec910345b163b17abd59/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.598631 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.607249118s: [/var/lib/containers/storage/overlay/47741195ba1ba47f828c3f91565001e9be28236bb7b5c7c18d8ea93306a8e8fc/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.599112 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.584212116s: [/var/lib/containers/storage/overlay/7129981cb4c03326a2d75f0f57b523f36c7f144c0a0b44a572aaf5e50e5c4c91/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.599730 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.612303224s: [/var/lib/containers/storage/overlay/664dbdf8b93af1129ddf988be1978456de1157046a9bc27479467e237501b090/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.605120 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.616038126s: [/var/lib/containers/storage/overlay/326b9c04998c0bb0190d753ea64829f8289a984af847653e6f0e5a8bb88f523d/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.605179 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.613775993s: [/var/lib/containers/storage/overlay/90bca8b0f7b64fe5e1704ede36c8a2ba3c4aaba08648511089913c6de46e3112/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.607684 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.618142712s: [/var/lib/containers/storage/overlay/e66e66d57c25bac91ea77e9a5ffca1e6d41e10f2b414f3f205c73369430e6f1c/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.609568 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.62215427s: [/var/lib/containers/storage/overlay/c9b76933cac9329cf3c24a92b7e41b5450e2394d859285501579986567a71ec8/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.619605 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.628238006s: [/var/lib/containers/storage/overlay/217c969f4e7404611c41d66c9149087e4eefbcf81da9165506333b89f3248404/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.620881 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.9558721s: [/var/lib/containers/storage/overlay/78cda72836a15d2f632a8a6aab2b2b51066581d91767a9cb16521ddc1af86c46/diff /var/log/pods/openshift-image-registry_image-registry-66df7c8f76-dgspm_26c71fe2-c0e4-405d-a6c6-2ce933013cba/registry/0.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.620960 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.633518318s: [/var/lib/containers/storage/overlay/abcb5e56ae834b495f98f68fce95c02d73159f237c430a6f60a09286ca733059/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.620997 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.632418079s: [/var/lib/containers/storage/overlay/bede9b9079c0bcbb72b319241fddc5a53a3dfcb8be4bdf103a7932639c2b1778/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.621366 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.63283262s: [/var/lib/containers/storage/overlay/2781134dd834e8bfa0d5ee4a6cf38dd26248770d49311f161418dcc5dd06040c/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.623106 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/octavia-operator-controller-manager-998648c74-dfs6c" podUID="2ac245f3-8f64-4b6f-a51f-09e4b5811412" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.112:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.624283 4809 trace.go:236] Trace[1986898609]: "iptables ChainExists" (06-Dec-2025 07:34:15.394) (total time: 2229ms): Dec 06 07:34:17 crc kubenswrapper[4809]: Trace[1986898609]: [2.22969282s] [2.22969282s] END Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.626870 4809 trace.go:236] Trace[277978949]: "iptables ChainExists" (06-Dec-2025 07:34:15.394) (total time: 2232ms): Dec 06 07:34:17 crc kubenswrapper[4809]: Trace[277978949]: [2.232830375s] [2.232830375s] END Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.628157 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.636803757s: [/var/lib/containers/storage/overlay/0361f6142046640aea03ab1f9f67f711a672c2958f0693e4aa1bd87fa799d391/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:17 crc kubenswrapper[4809]: I1206 07:34:17.629469 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.639932981s: [/var/lib/containers/storage/overlay/dc7b42b2a2832c3eb21d34f6da74ab89d0ec2c5960f572f77067a3e668ec2d27/diff ]; will not log again for this container unless duration exceeds 2s Dec 06 07:34:21 crc kubenswrapper[4809]: I1206 07:34:21.388850 4809 scope.go:117] "RemoveContainer" containerID="26a5b8063411f48708112d11ebddb0587d0fe6fc8f6a5c0aa5728fedd5b56f1d" Dec 06 07:34:21 crc kubenswrapper[4809]: E1206 07:34:21.389803 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:34:30 crc kubenswrapper[4809]: I1206 07:34:30.936909 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-dpbfw_de0bb137-3c1d-4632-a7ef-6922f027a3c1/control-plane-machine-set-operator/0.log" Dec 06 07:34:31 crc kubenswrapper[4809]: I1206 07:34:31.147190 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-852bd_6b48dffc-c51b-48e2-b22b-e809176c8830/kube-rbac-proxy/0.log" Dec 06 07:34:31 crc kubenswrapper[4809]: I1206 07:34:31.170977 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-852bd_6b48dffc-c51b-48e2-b22b-e809176c8830/machine-api-operator/0.log" Dec 06 07:34:34 crc kubenswrapper[4809]: I1206 07:34:34.390735 4809 scope.go:117] "RemoveContainer" containerID="26a5b8063411f48708112d11ebddb0587d0fe6fc8f6a5c0aa5728fedd5b56f1d" Dec 06 07:34:34 crc kubenswrapper[4809]: E1206 07:34:34.394235 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:34:36 crc kubenswrapper[4809]: I1206 07:34:36.168225 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/telemetry-operator-controller-manager-b4644b845-zxwvr" podUID="24aed9fd-d792-4bce-82a8-3ea4d70094b4" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.117:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:34:36 crc kubenswrapper[4809]: I1206 07:34:36.778652 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-galera-0" podUID="7c596e14-b96f-4e5f-a8ac-3dc66fa0276d" containerName="galera" probeResult="failure" output="command timed out" Dec 06 07:34:36 crc kubenswrapper[4809]: I1206 07:34:36.779820 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="7c596e14-b96f-4e5f-a8ac-3dc66fa0276d" containerName="galera" probeResult="failure" output="command timed out" Dec 06 07:34:41 crc kubenswrapper[4809]: I1206 07:34:41.600192 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd462lbf" podUID="6fbac224-e82d-482e-aeea-c83a9f978775" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.113:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:34:43 crc kubenswrapper[4809]: I1206 07:34:41.954205 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/speaker-5t9z6" podUID="21411d89-4a48-4556-8ad2-193f1c8be0db" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:34:46 crc kubenswrapper[4809]: I1206 07:34:46.388711 4809 scope.go:117] "RemoveContainer" containerID="26a5b8063411f48708112d11ebddb0587d0fe6fc8f6a5c0aa5728fedd5b56f1d" Dec 06 07:34:47 crc kubenswrapper[4809]: I1206 07:34:47.006360 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" event={"ID":"cbf4cf62-024e-4703-a8b8-9aecda9cd26a","Type":"ContainerStarted","Data":"2b99cfee1a9995b9796e60fb6b1a787ce8771b97d7af6a38edfaa8224097d30d"} Dec 06 07:34:50 crc kubenswrapper[4809]: I1206 07:34:50.096976 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-wgc72_8bc6fec5-e0e7-416f-b7ac-0a671cac3c0b/cert-manager-controller/0.log" Dec 06 07:34:50 crc kubenswrapper[4809]: I1206 07:34:50.316106 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-s5sjr_79eae299-f38e-4441-954b-833293821894/cert-manager-cainjector/0.log" Dec 06 07:34:50 crc kubenswrapper[4809]: I1206 07:34:50.378422 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-7vvw8_886a998b-a62e-412c-ae50-fd1ba3a4a098/cert-manager-webhook/0.log" Dec 06 07:35:03 crc kubenswrapper[4809]: I1206 07:35:03.078694 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-7dpj4_87deabac-9840-43a3-945b-cb4d054e34ba/nmstate-console-plugin/0.log" Dec 06 07:35:03 crc kubenswrapper[4809]: I1206 07:35:03.302462 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-pvgrd_d39362b1-ae5e-4852-be43-a335be68151b/kube-rbac-proxy/0.log" Dec 06 07:35:03 crc kubenswrapper[4809]: I1206 07:35:03.303248 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-jbb6s_57d09bf3-cccc-4745-834d-57eaad3f96a1/nmstate-handler/0.log" Dec 06 07:35:03 crc kubenswrapper[4809]: I1206 07:35:03.406313 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-pvgrd_d39362b1-ae5e-4852-be43-a335be68151b/nmstate-metrics/0.log" Dec 06 07:35:03 crc kubenswrapper[4809]: I1206 07:35:03.602280 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-t82p5_b86e935d-b651-435e-8218-f39878ec04b9/nmstate-operator/0.log" Dec 06 07:35:03 crc kubenswrapper[4809]: I1206 07:35:03.650661 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-pw6wx_50b406f3-bc91-492f-89ea-8eeae9ae132e/nmstate-webhook/0.log" Dec 06 07:35:15 crc kubenswrapper[4809]: I1206 07:35:15.714390 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-697657f76c-5stmg_19fc8500-fddb-4650-839a-82166e98a567/kube-rbac-proxy/0.log" Dec 06 07:35:15 crc kubenswrapper[4809]: I1206 07:35:15.766340 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-697657f76c-5stmg_19fc8500-fddb-4650-839a-82166e98a567/manager/1.log" Dec 06 07:35:15 crc kubenswrapper[4809]: I1206 07:35:15.912137 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-697657f76c-5stmg_19fc8500-fddb-4650-839a-82166e98a567/manager/0.log" Dec 06 07:35:18 crc kubenswrapper[4809]: I1206 07:35:18.721095 4809 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-qqv6x container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.30:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:35:18 crc kubenswrapper[4809]: I1206 07:35:18.722809 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qqv6x" podUID="8a225b5c-4ad5-4f08-a0cb-ae38a8fa83d4" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.30:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:35:20 crc kubenswrapper[4809]: I1206 07:35:20.879169 4809 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.354291351s: [/var/lib/containers/storage/overlay/416aea4e13593f624275782af3cd2f6dc5640263d2aad5b3d738efd0e73ebd0b/diff /var/log/pods/openshift-ingress_router-default-5444994796-n6z57_e5356f1d-5afd-47bd-89af-72abfa8f901f/router/1.log]; will not log again for this container unless duration exceeds 2s Dec 06 07:35:30 crc kubenswrapper[4809]: I1206 07:35:30.693418 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_cluster-logging-operator-ff9846bd-lcbr4_167bcfaa-27d5-4604-92f4-502351d91eba/cluster-logging-operator/0.log" Dec 06 07:35:30 crc kubenswrapper[4809]: I1206 07:35:30.934734 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_collector-hmjlt_1fe931cb-8318-4749-8be6-286c69da0860/collector/0.log" Dec 06 07:35:30 crc kubenswrapper[4809]: I1206 07:35:30.944957 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-compactor-0_2365b183-7913-4ce1-94bc-b30705d653b2/loki-compactor/0.log" Dec 06 07:35:31 crc kubenswrapper[4809]: I1206 07:35:31.117091 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-distributor-76cc67bf56-r5p25_2ae7975b-9830-47bf-af4f-6f7fe5e47894/loki-distributor/0.log" Dec 06 07:35:31 crc kubenswrapper[4809]: I1206 07:35:31.165071 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-65498c4f8f-48rjq_72a3ed95-3e3d-4faf-88e4-ad1731902910/gateway/0.log" Dec 06 07:35:31 crc kubenswrapper[4809]: I1206 07:35:31.189165 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-65498c4f8f-48rjq_72a3ed95-3e3d-4faf-88e4-ad1731902910/opa/0.log" Dec 06 07:35:31 crc kubenswrapper[4809]: I1206 07:35:31.433609 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-65498c4f8f-9s8qq_ef8f3e7e-57ea-42b8-a777-1778e5ed975b/opa/0.log" Dec 06 07:35:31 crc kubenswrapper[4809]: I1206 07:35:31.444431 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-65498c4f8f-9s8qq_ef8f3e7e-57ea-42b8-a777-1778e5ed975b/gateway/0.log" Dec 06 07:35:31 crc kubenswrapper[4809]: I1206 07:35:31.584645 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-index-gateway-0_535f6f4f-93e7-4483-9c80-598c21bd1e6e/loki-index-gateway/0.log" Dec 06 07:35:31 crc kubenswrapper[4809]: I1206 07:35:31.664303 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-ingester-0_0f3f468b-668a-4b15-af5b-c519c4ef03d7/loki-ingester/0.log" Dec 06 07:35:31 crc kubenswrapper[4809]: I1206 07:35:31.833958 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-querier-5895d59bb8-q5rsl_c4ce3880-869a-4fef-b76c-e669273700e1/loki-querier/0.log" Dec 06 07:35:31 crc kubenswrapper[4809]: I1206 07:35:31.835089 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-query-frontend-84558f7c9f-9mw6j_142a1d98-f469-4d1c-8541-d4ceb151f74e/loki-query-frontend/0.log" Dec 06 07:35:47 crc kubenswrapper[4809]: I1206 07:35:47.465028 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-d4fgg_073938c5-bc80-47d3-91e8-1d6d7da0f000/kube-rbac-proxy/0.log" Dec 06 07:35:47 crc kubenswrapper[4809]: I1206 07:35:47.460688 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-ttn9g_b483382f-c2e3-40bf-ab4b-b58c2f3f9132/frr-k8s-webhook-server/0.log" Dec 06 07:35:47 crc kubenswrapper[4809]: I1206 07:35:47.673693 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-d4fgg_073938c5-bc80-47d3-91e8-1d6d7da0f000/controller/0.log" Dec 06 07:35:47 crc kubenswrapper[4809]: I1206 07:35:47.761365 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z95cs_a3b6197b-05d4-49b2-9a67-56b675046a1d/cp-frr-files/0.log" Dec 06 07:35:47 crc kubenswrapper[4809]: I1206 07:35:47.980911 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z95cs_a3b6197b-05d4-49b2-9a67-56b675046a1d/cp-frr-files/0.log" Dec 06 07:35:47 crc kubenswrapper[4809]: I1206 07:35:47.981225 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z95cs_a3b6197b-05d4-49b2-9a67-56b675046a1d/cp-reloader/0.log" Dec 06 07:35:47 crc kubenswrapper[4809]: I1206 07:35:47.986178 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z95cs_a3b6197b-05d4-49b2-9a67-56b675046a1d/cp-metrics/0.log" Dec 06 07:35:48 crc kubenswrapper[4809]: I1206 07:35:48.047954 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z95cs_a3b6197b-05d4-49b2-9a67-56b675046a1d/cp-reloader/0.log" Dec 06 07:35:48 crc kubenswrapper[4809]: I1206 07:35:48.265700 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z95cs_a3b6197b-05d4-49b2-9a67-56b675046a1d/cp-frr-files/0.log" Dec 06 07:35:48 crc kubenswrapper[4809]: I1206 07:35:48.346462 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z95cs_a3b6197b-05d4-49b2-9a67-56b675046a1d/cp-reloader/0.log" Dec 06 07:35:48 crc kubenswrapper[4809]: I1206 07:35:48.359549 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z95cs_a3b6197b-05d4-49b2-9a67-56b675046a1d/cp-metrics/0.log" Dec 06 07:35:48 crc kubenswrapper[4809]: I1206 07:35:48.384854 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z95cs_a3b6197b-05d4-49b2-9a67-56b675046a1d/cp-metrics/0.log" Dec 06 07:35:48 crc kubenswrapper[4809]: I1206 07:35:48.531781 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z95cs_a3b6197b-05d4-49b2-9a67-56b675046a1d/cp-reloader/0.log" Dec 06 07:35:48 crc kubenswrapper[4809]: I1206 07:35:48.531789 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z95cs_a3b6197b-05d4-49b2-9a67-56b675046a1d/cp-frr-files/0.log" Dec 06 07:35:48 crc kubenswrapper[4809]: I1206 07:35:48.543432 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z95cs_a3b6197b-05d4-49b2-9a67-56b675046a1d/cp-metrics/0.log" Dec 06 07:35:48 crc kubenswrapper[4809]: I1206 07:35:48.614539 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z95cs_a3b6197b-05d4-49b2-9a67-56b675046a1d/controller/0.log" Dec 06 07:35:48 crc kubenswrapper[4809]: I1206 07:35:48.795271 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z95cs_a3b6197b-05d4-49b2-9a67-56b675046a1d/kube-rbac-proxy/0.log" Dec 06 07:35:48 crc kubenswrapper[4809]: I1206 07:35:48.837487 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z95cs_a3b6197b-05d4-49b2-9a67-56b675046a1d/frr-metrics/0.log" Dec 06 07:35:48 crc kubenswrapper[4809]: I1206 07:35:48.946353 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z95cs_a3b6197b-05d4-49b2-9a67-56b675046a1d/kube-rbac-proxy-frr/0.log" Dec 06 07:35:48 crc kubenswrapper[4809]: I1206 07:35:48.976285 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z95cs_a3b6197b-05d4-49b2-9a67-56b675046a1d/reloader/0.log" Dec 06 07:35:49 crc kubenswrapper[4809]: I1206 07:35:49.501572 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-6f667b6558-c67wt_69484422-8317-4050-945c-9c9b127552b1/manager/0.log" Dec 06 07:35:49 crc kubenswrapper[4809]: I1206 07:35:49.789294 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-765b7d7f99-8dngt_60798ea0-75be-4a0c-a8c6-9fb431ba0e67/webhook-server/0.log" Dec 06 07:35:49 crc kubenswrapper[4809]: I1206 07:35:49.851436 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-5t9z6_21411d89-4a48-4556-8ad2-193f1c8be0db/kube-rbac-proxy/0.log" Dec 06 07:35:50 crc kubenswrapper[4809]: I1206 07:35:50.732783 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-5t9z6_21411d89-4a48-4556-8ad2-193f1c8be0db/speaker/0.log" Dec 06 07:35:50 crc kubenswrapper[4809]: I1206 07:35:50.956782 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-z95cs_a3b6197b-05d4-49b2-9a67-56b675046a1d/frr/0.log" Dec 06 07:36:05 crc kubenswrapper[4809]: I1206 07:36:05.361082 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8pfj4q_de7c1cd3-87d5-431f-a971-5474ac3d56ec/util/0.log" Dec 06 07:36:05 crc kubenswrapper[4809]: I1206 07:36:05.541588 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8pfj4q_de7c1cd3-87d5-431f-a971-5474ac3d56ec/util/0.log" Dec 06 07:36:05 crc kubenswrapper[4809]: I1206 07:36:05.616662 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8pfj4q_de7c1cd3-87d5-431f-a971-5474ac3d56ec/pull/0.log" Dec 06 07:36:05 crc kubenswrapper[4809]: I1206 07:36:05.616998 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8pfj4q_de7c1cd3-87d5-431f-a971-5474ac3d56ec/pull/0.log" Dec 06 07:36:05 crc kubenswrapper[4809]: I1206 07:36:05.776005 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8pfj4q_de7c1cd3-87d5-431f-a971-5474ac3d56ec/util/0.log" Dec 06 07:36:05 crc kubenswrapper[4809]: I1206 07:36:05.796397 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8pfj4q_de7c1cd3-87d5-431f-a971-5474ac3d56ec/pull/0.log" Dec 06 07:36:05 crc kubenswrapper[4809]: I1206 07:36:05.875982 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8pfj4q_de7c1cd3-87d5-431f-a971-5474ac3d56ec/extract/0.log" Dec 06 07:36:05 crc kubenswrapper[4809]: I1206 07:36:05.969211 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwt88_d8953b5f-9e44-407f-aab5-6a434e11a7ef/util/0.log" Dec 06 07:36:06 crc kubenswrapper[4809]: I1206 07:36:06.244486 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwt88_d8953b5f-9e44-407f-aab5-6a434e11a7ef/util/0.log" Dec 06 07:36:06 crc kubenswrapper[4809]: I1206 07:36:06.255702 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwt88_d8953b5f-9e44-407f-aab5-6a434e11a7ef/pull/0.log" Dec 06 07:36:06 crc kubenswrapper[4809]: I1206 07:36:06.255785 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwt88_d8953b5f-9e44-407f-aab5-6a434e11a7ef/pull/0.log" Dec 06 07:36:06 crc kubenswrapper[4809]: I1206 07:36:06.662342 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwt88_d8953b5f-9e44-407f-aab5-6a434e11a7ef/extract/0.log" Dec 06 07:36:06 crc kubenswrapper[4809]: I1206 07:36:06.676979 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwt88_d8953b5f-9e44-407f-aab5-6a434e11a7ef/pull/0.log" Dec 06 07:36:06 crc kubenswrapper[4809]: I1206 07:36:06.682684 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwt88_d8953b5f-9e44-407f-aab5-6a434e11a7ef/util/0.log" Dec 06 07:36:06 crc kubenswrapper[4809]: I1206 07:36:06.869680 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210t4wcj_64828a07-1605-4981-a9ca-9442ca370be5/util/0.log" Dec 06 07:36:07 crc kubenswrapper[4809]: I1206 07:36:07.133146 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210t4wcj_64828a07-1605-4981-a9ca-9442ca370be5/pull/0.log" Dec 06 07:36:07 crc kubenswrapper[4809]: I1206 07:36:07.185872 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210t4wcj_64828a07-1605-4981-a9ca-9442ca370be5/pull/0.log" Dec 06 07:36:07 crc kubenswrapper[4809]: I1206 07:36:07.195109 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210t4wcj_64828a07-1605-4981-a9ca-9442ca370be5/util/0.log" Dec 06 07:36:07 crc kubenswrapper[4809]: I1206 07:36:07.339713 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210t4wcj_64828a07-1605-4981-a9ca-9442ca370be5/util/0.log" Dec 06 07:36:07 crc kubenswrapper[4809]: I1206 07:36:07.365066 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210t4wcj_64828a07-1605-4981-a9ca-9442ca370be5/pull/0.log" Dec 06 07:36:07 crc kubenswrapper[4809]: I1206 07:36:07.453192 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210t4wcj_64828a07-1605-4981-a9ca-9442ca370be5/extract/0.log" Dec 06 07:36:07 crc kubenswrapper[4809]: I1206 07:36:07.564950 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpd5xh_e12ceeab-a1f8-4a5b-aefc-77d7cf057f40/util/0.log" Dec 06 07:36:08 crc kubenswrapper[4809]: I1206 07:36:08.027817 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpd5xh_e12ceeab-a1f8-4a5b-aefc-77d7cf057f40/pull/0.log" Dec 06 07:36:08 crc kubenswrapper[4809]: I1206 07:36:08.027915 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpd5xh_e12ceeab-a1f8-4a5b-aefc-77d7cf057f40/pull/0.log" Dec 06 07:36:08 crc kubenswrapper[4809]: I1206 07:36:08.232816 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpd5xh_e12ceeab-a1f8-4a5b-aefc-77d7cf057f40/util/0.log" Dec 06 07:36:08 crc kubenswrapper[4809]: I1206 07:36:08.260445 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpd5xh_e12ceeab-a1f8-4a5b-aefc-77d7cf057f40/util/0.log" Dec 06 07:36:08 crc kubenswrapper[4809]: I1206 07:36:08.261042 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpd5xh_e12ceeab-a1f8-4a5b-aefc-77d7cf057f40/pull/0.log" Dec 06 07:36:08 crc kubenswrapper[4809]: I1206 07:36:08.365991 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpd5xh_e12ceeab-a1f8-4a5b-aefc-77d7cf057f40/extract/0.log" Dec 06 07:36:08 crc kubenswrapper[4809]: I1206 07:36:08.511228 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rhmbp_ade7d3b7-533c-40a1-9964-4e5ac7b88214/util/0.log" Dec 06 07:36:08 crc kubenswrapper[4809]: I1206 07:36:08.760127 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rhmbp_ade7d3b7-533c-40a1-9964-4e5ac7b88214/util/0.log" Dec 06 07:36:08 crc kubenswrapper[4809]: I1206 07:36:08.761771 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rhmbp_ade7d3b7-533c-40a1-9964-4e5ac7b88214/pull/0.log" Dec 06 07:36:08 crc kubenswrapper[4809]: I1206 07:36:08.769173 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rhmbp_ade7d3b7-533c-40a1-9964-4e5ac7b88214/pull/0.log" Dec 06 07:36:08 crc kubenswrapper[4809]: I1206 07:36:08.929258 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rhmbp_ade7d3b7-533c-40a1-9964-4e5ac7b88214/util/0.log" Dec 06 07:36:08 crc kubenswrapper[4809]: I1206 07:36:08.973928 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rhmbp_ade7d3b7-533c-40a1-9964-4e5ac7b88214/pull/0.log" Dec 06 07:36:08 crc kubenswrapper[4809]: I1206 07:36:08.999501 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rhmbp_ade7d3b7-533c-40a1-9964-4e5ac7b88214/extract/0.log" Dec 06 07:36:09 crc kubenswrapper[4809]: I1206 07:36:09.179461 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nqpwx_408b531c-9f2a-4ee5-ad65-db67b2313be3/extract-utilities/0.log" Dec 06 07:36:09 crc kubenswrapper[4809]: I1206 07:36:09.474713 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nqpwx_408b531c-9f2a-4ee5-ad65-db67b2313be3/extract-content/0.log" Dec 06 07:36:09 crc kubenswrapper[4809]: I1206 07:36:09.479282 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nqpwx_408b531c-9f2a-4ee5-ad65-db67b2313be3/extract-content/0.log" Dec 06 07:36:09 crc kubenswrapper[4809]: I1206 07:36:09.485855 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nqpwx_408b531c-9f2a-4ee5-ad65-db67b2313be3/extract-utilities/0.log" Dec 06 07:36:09 crc kubenswrapper[4809]: I1206 07:36:09.697377 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nqpwx_408b531c-9f2a-4ee5-ad65-db67b2313be3/extract-utilities/0.log" Dec 06 07:36:09 crc kubenswrapper[4809]: I1206 07:36:09.723894 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nqpwx_408b531c-9f2a-4ee5-ad65-db67b2313be3/extract-content/0.log" Dec 06 07:36:09 crc kubenswrapper[4809]: I1206 07:36:09.810036 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-72dcl_f8e767ec-8d4b-4a29-8562-115db074c7a7/extract-utilities/0.log" Dec 06 07:36:10 crc kubenswrapper[4809]: I1206 07:36:10.039244 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-72dcl_f8e767ec-8d4b-4a29-8562-115db074c7a7/extract-content/0.log" Dec 06 07:36:10 crc kubenswrapper[4809]: I1206 07:36:10.040587 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-72dcl_f8e767ec-8d4b-4a29-8562-115db074c7a7/extract-utilities/0.log" Dec 06 07:36:10 crc kubenswrapper[4809]: I1206 07:36:10.055708 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-72dcl_f8e767ec-8d4b-4a29-8562-115db074c7a7/extract-content/0.log" Dec 06 07:36:10 crc kubenswrapper[4809]: I1206 07:36:10.255021 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-72dcl_f8e767ec-8d4b-4a29-8562-115db074c7a7/extract-utilities/0.log" Dec 06 07:36:10 crc kubenswrapper[4809]: I1206 07:36:10.298337 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-72dcl_f8e767ec-8d4b-4a29-8562-115db074c7a7/extract-content/0.log" Dec 06 07:36:10 crc kubenswrapper[4809]: I1206 07:36:10.531839 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-r2lg6_286c8de4-0ea9-43ad-bb4f-970319e0f4d3/marketplace-operator/0.log" Dec 06 07:36:10 crc kubenswrapper[4809]: I1206 07:36:10.559444 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-72dcl_f8e767ec-8d4b-4a29-8562-115db074c7a7/registry-server/0.log" Dec 06 07:36:10 crc kubenswrapper[4809]: I1206 07:36:10.837449 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nqpwx_408b531c-9f2a-4ee5-ad65-db67b2313be3/registry-server/0.log" Dec 06 07:36:10 crc kubenswrapper[4809]: I1206 07:36:10.943280 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-zjvtm_0714d252-8cd5-4a99-9050-8e383ad64885/extract-utilities/0.log" Dec 06 07:36:11 crc kubenswrapper[4809]: I1206 07:36:11.215575 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-zjvtm_0714d252-8cd5-4a99-9050-8e383ad64885/extract-content/0.log" Dec 06 07:36:11 crc kubenswrapper[4809]: I1206 07:36:11.227129 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-zjvtm_0714d252-8cd5-4a99-9050-8e383ad64885/extract-utilities/0.log" Dec 06 07:36:11 crc kubenswrapper[4809]: I1206 07:36:11.254577 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-zjvtm_0714d252-8cd5-4a99-9050-8e383ad64885/extract-content/0.log" Dec 06 07:36:11 crc kubenswrapper[4809]: I1206 07:36:11.465282 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-zjvtm_0714d252-8cd5-4a99-9050-8e383ad64885/extract-utilities/0.log" Dec 06 07:36:11 crc kubenswrapper[4809]: I1206 07:36:11.549497 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-zjvtm_0714d252-8cd5-4a99-9050-8e383ad64885/extract-content/0.log" Dec 06 07:36:11 crc kubenswrapper[4809]: I1206 07:36:11.592213 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-czfkp_1c5ff94a-d9b0-41f2-9d47-47f67e0a9895/extract-utilities/0.log" Dec 06 07:36:11 crc kubenswrapper[4809]: I1206 07:36:11.762745 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-zjvtm_0714d252-8cd5-4a99-9050-8e383ad64885/registry-server/0.log" Dec 06 07:36:11 crc kubenswrapper[4809]: I1206 07:36:11.833689 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-czfkp_1c5ff94a-d9b0-41f2-9d47-47f67e0a9895/extract-utilities/0.log" Dec 06 07:36:11 crc kubenswrapper[4809]: I1206 07:36:11.850453 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-czfkp_1c5ff94a-d9b0-41f2-9d47-47f67e0a9895/extract-content/0.log" Dec 06 07:36:11 crc kubenswrapper[4809]: I1206 07:36:11.901351 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-czfkp_1c5ff94a-d9b0-41f2-9d47-47f67e0a9895/extract-content/0.log" Dec 06 07:36:12 crc kubenswrapper[4809]: I1206 07:36:12.084319 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-czfkp_1c5ff94a-d9b0-41f2-9d47-47f67e0a9895/extract-content/0.log" Dec 06 07:36:12 crc kubenswrapper[4809]: I1206 07:36:12.094715 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-czfkp_1c5ff94a-d9b0-41f2-9d47-47f67e0a9895/extract-utilities/0.log" Dec 06 07:36:13 crc kubenswrapper[4809]: I1206 07:36:13.080666 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-czfkp_1c5ff94a-d9b0-41f2-9d47-47f67e0a9895/registry-server/0.log" Dec 06 07:36:26 crc kubenswrapper[4809]: I1206 07:36:26.971638 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-h7k8w_c034c52b-dd97-4722-84fe-b3389a008fbd/prometheus-operator/0.log" Dec 06 07:36:27 crc kubenswrapper[4809]: I1206 07:36:27.449958 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-566bd9db68-t8tqd_51bf2206-9c20-4c9a-8dbd-1f74c0033236/prometheus-operator-admission-webhook/0.log" Dec 06 07:36:27 crc kubenswrapper[4809]: I1206 07:36:27.530301 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-566bd9db68-99254_35c30afa-953e-4808-b4ba-473ffb44c4a5/prometheus-operator-admission-webhook/0.log" Dec 06 07:36:27 crc kubenswrapper[4809]: I1206 07:36:27.733187 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-bfc8s_6d4f55df-5dc2-4e72-b7f8-cee9229da476/operator/0.log" Dec 06 07:36:27 crc kubenswrapper[4809]: I1206 07:36:27.916265 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-ui-dashboards-7d5fb4cbfb-sj6gs_6be1b144-2e51-42a1-9430-201ee08ce602/observability-ui-dashboards/0.log" Dec 06 07:36:28 crc kubenswrapper[4809]: I1206 07:36:28.047537 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-fh2sn_bb9f4b83-1aa5-48ec-ac64-2d4c7835c5af/perses-operator/0.log" Dec 06 07:36:47 crc kubenswrapper[4809]: I1206 07:36:47.013176 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-697657f76c-5stmg_19fc8500-fddb-4650-839a-82166e98a567/manager/1.log" Dec 06 07:36:47 crc kubenswrapper[4809]: I1206 07:36:47.020535 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-697657f76c-5stmg_19fc8500-fddb-4650-839a-82166e98a567/kube-rbac-proxy/0.log" Dec 06 07:36:47 crc kubenswrapper[4809]: I1206 07:36:47.064830 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-697657f76c-5stmg_19fc8500-fddb-4650-839a-82166e98a567/manager/0.log" Dec 06 07:37:04 crc kubenswrapper[4809]: I1206 07:37:04.499400 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:37:04 crc kubenswrapper[4809]: I1206 07:37:04.499879 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:37:06 crc kubenswrapper[4809]: I1206 07:37:06.778435 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-galera-0" podUID="7c596e14-b96f-4e5f-a8ac-3dc66fa0276d" containerName="galera" probeResult="failure" output="command timed out" Dec 06 07:37:06 crc kubenswrapper[4809]: I1206 07:37:06.780142 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="7c596e14-b96f-4e5f-a8ac-3dc66fa0276d" containerName="galera" probeResult="failure" output="command timed out" Dec 06 07:37:15 crc kubenswrapper[4809]: I1206 07:37:15.240310 4809 trace.go:236] Trace[44981590]: "Calculate volume metrics of images for pod openshift-machine-config-operator/machine-config-operator-74547568cd-nsqq5" (06-Dec-2025 07:37:14.175) (total time: 1064ms): Dec 06 07:37:15 crc kubenswrapper[4809]: Trace[44981590]: [1.064621822s] [1.064621822s] END Dec 06 07:37:34 crc kubenswrapper[4809]: I1206 07:37:34.496317 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:37:34 crc kubenswrapper[4809]: I1206 07:37:34.496831 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:38:04 crc kubenswrapper[4809]: I1206 07:38:04.496635 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:38:04 crc kubenswrapper[4809]: I1206 07:38:04.497371 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:38:04 crc kubenswrapper[4809]: I1206 07:38:04.497430 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-npms2" Dec 06 07:38:04 crc kubenswrapper[4809]: I1206 07:38:04.620418 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2b99cfee1a9995b9796e60fb6b1a787ce8771b97d7af6a38edfaa8224097d30d"} pod="openshift-machine-config-operator/machine-config-daemon-npms2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 07:38:04 crc kubenswrapper[4809]: I1206 07:38:04.620571 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" containerID="cri-o://2b99cfee1a9995b9796e60fb6b1a787ce8771b97d7af6a38edfaa8224097d30d" gracePeriod=600 Dec 06 07:38:06 crc kubenswrapper[4809]: I1206 07:38:06.121177 4809 patch_prober.go:28] interesting pod/loki-operator-controller-manager-697657f76c-5stmg container/manager namespace/openshift-operators-redhat: Liveness probe status=failure output="Get \"http://10.217.0.47:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:38:06 crc kubenswrapper[4809]: I1206 07:38:06.121189 4809 patch_prober.go:28] interesting pod/loki-operator-controller-manager-697657f76c-5stmg container/manager namespace/openshift-operators-redhat: Readiness probe status=failure output="Get \"http://10.217.0.47:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:38:06 crc kubenswrapper[4809]: I1206 07:38:06.121557 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators-redhat/loki-operator-controller-manager-697657f76c-5stmg" podUID="19fc8500-fddb-4650-839a-82166e98a567" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.47:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:38:06 crc kubenswrapper[4809]: I1206 07:38:06.121478 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operators-redhat/loki-operator-controller-manager-697657f76c-5stmg" podUID="19fc8500-fddb-4650-839a-82166e98a567" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.47:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:38:07 crc kubenswrapper[4809]: I1206 07:38:07.044449 4809 generic.go:334] "Generic (PLEG): container finished" podID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerID="2b99cfee1a9995b9796e60fb6b1a787ce8771b97d7af6a38edfaa8224097d30d" exitCode=0 Dec 06 07:38:07 crc kubenswrapper[4809]: I1206 07:38:07.044517 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" event={"ID":"cbf4cf62-024e-4703-a8b8-9aecda9cd26a","Type":"ContainerDied","Data":"2b99cfee1a9995b9796e60fb6b1a787ce8771b97d7af6a38edfaa8224097d30d"} Dec 06 07:38:07 crc kubenswrapper[4809]: I1206 07:38:07.044773 4809 scope.go:117] "RemoveContainer" containerID="26a5b8063411f48708112d11ebddb0587d0fe6fc8f6a5c0aa5728fedd5b56f1d" Dec 06 07:38:08 crc kubenswrapper[4809]: I1206 07:38:08.058321 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" event={"ID":"cbf4cf62-024e-4703-a8b8-9aecda9cd26a","Type":"ContainerStarted","Data":"4b28141798e8be5bf2c8cdfd3390a3ddfd37326a66c4ad4407752e73538ea965"} Dec 06 07:38:27 crc kubenswrapper[4809]: I1206 07:38:27.959212 4809 patch_prober.go:28] interesting pod/oauth-openshift-55889b984c-nxw8t container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.56:6443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:38:27 crc kubenswrapper[4809]: I1206 07:38:27.960119 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-55889b984c-nxw8t" podUID="404214e0-30aa-44ff-b2d0-fb3c127068f9" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.56:6443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:38:31 crc kubenswrapper[4809]: I1206 07:38:31.174859 4809 patch_prober.go:28] interesting pod/prometheus-operator-admission-webhook-f54c54754-ljfgz container/prometheus-operator-admission-webhook namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.69:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:38:31 crc kubenswrapper[4809]: I1206 07:38:31.175700 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-ljfgz" podUID="5e0d0108-b36f-47bc-bb5c-9fbc5ed0bf92" containerName="prometheus-operator-admission-webhook" probeResult="failure" output="Get \"https://10.217.0.69:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:38:43 crc kubenswrapper[4809]: I1206 07:38:43.144354 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="b4a36ec8-2831-4f0f-b20c-137fa3d081c8" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Dec 06 07:38:43 crc kubenswrapper[4809]: I1206 07:38:43.222188 4809 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Liveness probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:38:43 crc kubenswrapper[4809]: I1206 07:38:43.299256 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 07:38:43 crc kubenswrapper[4809]: E1206 07:38:43.224463 4809 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="11.714s" Dec 06 07:38:43 crc kubenswrapper[4809]: I1206 07:38:43.252784 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd462lbf" podUID="6fbac224-e82d-482e-aeea-c83a9f978775" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.113:8081/healthz\": dial tcp 10.217.0.113:8081: connect: connection refused" Dec 06 07:38:43 crc kubenswrapper[4809]: I1206 07:38:43.277451 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd462lbf" podUID="6fbac224-e82d-482e-aeea-c83a9f978775" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.113:8081/readyz\": dial tcp 10.217.0.113:8081: connect: connection refused" Dec 06 07:38:43 crc kubenswrapper[4809]: I1206 07:38:43.243802 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7trbt" podUID="45f87079-e852-4b47-a240-6f2bcf247509" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.105:8081/readyz\": EOF" Dec 06 07:38:43 crc kubenswrapper[4809]: I1206 07:38:43.369465 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/test-operator-controller-manager-5854674fcc-lqkvl" podUID="b376f569-0116-41f4-8a57-b5e5769e6a10" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.118:8081/readyz\": dial tcp 10.217.0.118:8081: connect: connection refused" Dec 06 07:38:43 crc kubenswrapper[4809]: E1206 07:38:43.968333 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod45f87079_e852_4b47_a240_6f2bcf247509.slice/crio-be1ecfef1755d34a69c3ac07b04e729bc32095ffdca7c1d32de77e47839fe82a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb376f569_0116_41f4_8a57_b5e5769e6a10.slice/crio-b317e5312c43701b451e86d74c12b7f999f7e606ea9ed3c6f3a5c991a820a635.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6fbac224_e82d_482e_aeea_c83a9f978775.slice/crio-e0d1203b2b4ea6eef7f2343a750bed29c71c8cb6e7a2fb1fa571c99d2d50700f.scope\": RecentStats: unable to find data in memory cache]" Dec 06 07:38:43 crc kubenswrapper[4809]: E1206 07:38:43.969287 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6fbac224_e82d_482e_aeea_c83a9f978775.slice/crio-e0d1203b2b4ea6eef7f2343a750bed29c71c8cb6e7a2fb1fa571c99d2d50700f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod45f87079_e852_4b47_a240_6f2bcf247509.slice/crio-be1ecfef1755d34a69c3ac07b04e729bc32095ffdca7c1d32de77e47839fe82a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb376f569_0116_41f4_8a57_b5e5769e6a10.slice/crio-b317e5312c43701b451e86d74c12b7f999f7e606ea9ed3c6f3a5c991a820a635.scope\": RecentStats: unable to find data in memory cache]" Dec 06 07:38:44 crc kubenswrapper[4809]: I1206 07:38:44.629874 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7trbt" podUID="45f87079-e852-4b47-a240-6f2bcf247509" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.105:8081/readyz\": dial tcp 10.217.0.105:8081: connect: connection refused" Dec 06 07:38:44 crc kubenswrapper[4809]: I1206 07:38:44.629957 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7trbt" podUID="45f87079-e852-4b47-a240-6f2bcf247509" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.105:8081/healthz\": dial tcp 10.217.0.105:8081: connect: connection refused" Dec 06 07:38:45 crc kubenswrapper[4809]: I1206 07:38:45.167322 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/test-operator-controller-manager-5854674fcc-lqkvl" podUID="b376f569-0116-41f4-8a57-b5e5769e6a10" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.118:8081/readyz\": dial tcp 10.217.0.118:8081: connect: connection refused" Dec 06 07:38:45 crc kubenswrapper[4809]: I1206 07:38:45.168684 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/test-operator-controller-manager-5854674fcc-lqkvl" podUID="b376f569-0116-41f4-8a57-b5e5769e6a10" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.118:8081/healthz\": dial tcp 10.217.0.118:8081: connect: connection refused" Dec 06 07:38:48 crc kubenswrapper[4809]: I1206 07:38:47.694682 4809 scope.go:117] "RemoveContainer" containerID="02aabf36b6a93c6a064da2230e09faaf3ccd6ce4facde411766574f66588e4ea" Dec 06 07:38:50 crc kubenswrapper[4809]: I1206 07:38:50.539624 4809 generic.go:334] "Generic (PLEG): container finished" podID="6fbac224-e82d-482e-aeea-c83a9f978775" containerID="e0d1203b2b4ea6eef7f2343a750bed29c71c8cb6e7a2fb1fa571c99d2d50700f" exitCode=1 Dec 06 07:38:50 crc kubenswrapper[4809]: I1206 07:38:50.540298 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd462lbf" event={"ID":"6fbac224-e82d-482e-aeea-c83a9f978775","Type":"ContainerDied","Data":"e0d1203b2b4ea6eef7f2343a750bed29c71c8cb6e7a2fb1fa571c99d2d50700f"} Dec 06 07:38:50 crc kubenswrapper[4809]: I1206 07:38:50.541205 4809 scope.go:117] "RemoveContainer" containerID="e0d1203b2b4ea6eef7f2343a750bed29c71c8cb6e7a2fb1fa571c99d2d50700f" Dec 06 07:38:50 crc kubenswrapper[4809]: I1206 07:38:50.550794 4809 generic.go:334] "Generic (PLEG): container finished" podID="b376f569-0116-41f4-8a57-b5e5769e6a10" containerID="b317e5312c43701b451e86d74c12b7f999f7e606ea9ed3c6f3a5c991a820a635" exitCode=1 Dec 06 07:38:50 crc kubenswrapper[4809]: I1206 07:38:50.550895 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-lqkvl" event={"ID":"b376f569-0116-41f4-8a57-b5e5769e6a10","Type":"ContainerDied","Data":"b317e5312c43701b451e86d74c12b7f999f7e606ea9ed3c6f3a5c991a820a635"} Dec 06 07:38:50 crc kubenswrapper[4809]: I1206 07:38:50.551007 4809 scope.go:117] "RemoveContainer" containerID="5c7db629ed3edf74098384a0a0132f41fe046c98a11cfe620d66c8f600323dc5" Dec 06 07:38:50 crc kubenswrapper[4809]: I1206 07:38:50.552564 4809 scope.go:117] "RemoveContainer" containerID="b317e5312c43701b451e86d74c12b7f999f7e606ea9ed3c6f3a5c991a820a635" Dec 06 07:38:50 crc kubenswrapper[4809]: I1206 07:38:50.556637 4809 generic.go:334] "Generic (PLEG): container finished" podID="45f87079-e852-4b47-a240-6f2bcf247509" containerID="be1ecfef1755d34a69c3ac07b04e729bc32095ffdca7c1d32de77e47839fe82a" exitCode=1 Dec 06 07:38:50 crc kubenswrapper[4809]: I1206 07:38:50.556762 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7trbt" event={"ID":"45f87079-e852-4b47-a240-6f2bcf247509","Type":"ContainerDied","Data":"be1ecfef1755d34a69c3ac07b04e729bc32095ffdca7c1d32de77e47839fe82a"} Dec 06 07:38:50 crc kubenswrapper[4809]: I1206 07:38:50.557593 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd462lbf" Dec 06 07:38:50 crc kubenswrapper[4809]: I1206 07:38:50.558446 4809 scope.go:117] "RemoveContainer" containerID="be1ecfef1755d34a69c3ac07b04e729bc32095ffdca7c1d32de77e47839fe82a" Dec 06 07:38:53 crc kubenswrapper[4809]: I1206 07:38:53.593552 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-lqkvl" event={"ID":"b376f569-0116-41f4-8a57-b5e5769e6a10","Type":"ContainerStarted","Data":"556c2a18c236d0a3073e37c8c98fcbacb2d72ecc327e037679e3e8f13dfae6d4"} Dec 06 07:38:54 crc kubenswrapper[4809]: I1206 07:38:54.626511 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7trbt" event={"ID":"45f87079-e852-4b47-a240-6f2bcf247509","Type":"ContainerStarted","Data":"22d531bb2d69d12e4ebb34f91e6d8df4e509c0bd12e51578bcd174f430affb64"} Dec 06 07:38:54 crc kubenswrapper[4809]: I1206 07:38:54.627424 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7trbt" Dec 06 07:38:54 crc kubenswrapper[4809]: I1206 07:38:54.631218 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd462lbf" event={"ID":"6fbac224-e82d-482e-aeea-c83a9f978775","Type":"ContainerStarted","Data":"536cd82320aa0427cc7a849fc2a26f3dc1521ae85345a974f9d8e14b4bb23cfd"} Dec 06 07:38:54 crc kubenswrapper[4809]: I1206 07:38:54.631419 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-lqkvl" Dec 06 07:38:54 crc kubenswrapper[4809]: I1206 07:38:54.811172 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7xn64"] Dec 06 07:38:54 crc kubenswrapper[4809]: E1206 07:38:54.811735 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f3c416e-5913-4f48-8192-c64e009042b6" containerName="extract-utilities" Dec 06 07:38:54 crc kubenswrapper[4809]: I1206 07:38:54.811761 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f3c416e-5913-4f48-8192-c64e009042b6" containerName="extract-utilities" Dec 06 07:38:54 crc kubenswrapper[4809]: E1206 07:38:54.811831 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f3c416e-5913-4f48-8192-c64e009042b6" containerName="registry-server" Dec 06 07:38:54 crc kubenswrapper[4809]: I1206 07:38:54.811841 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f3c416e-5913-4f48-8192-c64e009042b6" containerName="registry-server" Dec 06 07:38:54 crc kubenswrapper[4809]: E1206 07:38:54.811865 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f3c416e-5913-4f48-8192-c64e009042b6" containerName="extract-content" Dec 06 07:38:54 crc kubenswrapper[4809]: I1206 07:38:54.811876 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f3c416e-5913-4f48-8192-c64e009042b6" containerName="extract-content" Dec 06 07:38:54 crc kubenswrapper[4809]: I1206 07:38:54.812195 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f3c416e-5913-4f48-8192-c64e009042b6" containerName="registry-server" Dec 06 07:38:54 crc kubenswrapper[4809]: I1206 07:38:54.814550 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7xn64" Dec 06 07:38:54 crc kubenswrapper[4809]: I1206 07:38:54.833473 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7xn64"] Dec 06 07:38:54 crc kubenswrapper[4809]: I1206 07:38:54.973187 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09aa0663-ccc7-4b7a-a2b1-1579cac22552-utilities\") pod \"certified-operators-7xn64\" (UID: \"09aa0663-ccc7-4b7a-a2b1-1579cac22552\") " pod="openshift-marketplace/certified-operators-7xn64" Dec 06 07:38:54 crc kubenswrapper[4809]: I1206 07:38:54.974387 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgbpv\" (UniqueName: \"kubernetes.io/projected/09aa0663-ccc7-4b7a-a2b1-1579cac22552-kube-api-access-qgbpv\") pod \"certified-operators-7xn64\" (UID: \"09aa0663-ccc7-4b7a-a2b1-1579cac22552\") " pod="openshift-marketplace/certified-operators-7xn64" Dec 06 07:38:54 crc kubenswrapper[4809]: I1206 07:38:54.974665 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09aa0663-ccc7-4b7a-a2b1-1579cac22552-catalog-content\") pod \"certified-operators-7xn64\" (UID: \"09aa0663-ccc7-4b7a-a2b1-1579cac22552\") " pod="openshift-marketplace/certified-operators-7xn64" Dec 06 07:38:55 crc kubenswrapper[4809]: I1206 07:38:55.076923 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgbpv\" (UniqueName: \"kubernetes.io/projected/09aa0663-ccc7-4b7a-a2b1-1579cac22552-kube-api-access-qgbpv\") pod \"certified-operators-7xn64\" (UID: \"09aa0663-ccc7-4b7a-a2b1-1579cac22552\") " pod="openshift-marketplace/certified-operators-7xn64" Dec 06 07:38:55 crc kubenswrapper[4809]: I1206 07:38:55.077209 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09aa0663-ccc7-4b7a-a2b1-1579cac22552-catalog-content\") pod \"certified-operators-7xn64\" (UID: \"09aa0663-ccc7-4b7a-a2b1-1579cac22552\") " pod="openshift-marketplace/certified-operators-7xn64" Dec 06 07:38:55 crc kubenswrapper[4809]: I1206 07:38:55.077416 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09aa0663-ccc7-4b7a-a2b1-1579cac22552-utilities\") pod \"certified-operators-7xn64\" (UID: \"09aa0663-ccc7-4b7a-a2b1-1579cac22552\") " pod="openshift-marketplace/certified-operators-7xn64" Dec 06 07:38:55 crc kubenswrapper[4809]: I1206 07:38:55.079295 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09aa0663-ccc7-4b7a-a2b1-1579cac22552-catalog-content\") pod \"certified-operators-7xn64\" (UID: \"09aa0663-ccc7-4b7a-a2b1-1579cac22552\") " pod="openshift-marketplace/certified-operators-7xn64" Dec 06 07:38:55 crc kubenswrapper[4809]: I1206 07:38:55.079614 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09aa0663-ccc7-4b7a-a2b1-1579cac22552-utilities\") pod \"certified-operators-7xn64\" (UID: \"09aa0663-ccc7-4b7a-a2b1-1579cac22552\") " pod="openshift-marketplace/certified-operators-7xn64" Dec 06 07:38:55 crc kubenswrapper[4809]: I1206 07:38:55.098955 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgbpv\" (UniqueName: \"kubernetes.io/projected/09aa0663-ccc7-4b7a-a2b1-1579cac22552-kube-api-access-qgbpv\") pod \"certified-operators-7xn64\" (UID: \"09aa0663-ccc7-4b7a-a2b1-1579cac22552\") " pod="openshift-marketplace/certified-operators-7xn64" Dec 06 07:38:55 crc kubenswrapper[4809]: I1206 07:38:55.141893 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7xn64" Dec 06 07:38:55 crc kubenswrapper[4809]: I1206 07:38:55.641238 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd462lbf" Dec 06 07:38:55 crc kubenswrapper[4809]: I1206 07:38:55.840036 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7xn64"] Dec 06 07:38:56 crc kubenswrapper[4809]: I1206 07:38:56.657791 4809 generic.go:334] "Generic (PLEG): container finished" podID="09aa0663-ccc7-4b7a-a2b1-1579cac22552" containerID="25bd87e25e5b13f663de714f70bc3747cc81e9b118039d657abb521a45be18d7" exitCode=0 Dec 06 07:38:56 crc kubenswrapper[4809]: I1206 07:38:56.658151 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7xn64" event={"ID":"09aa0663-ccc7-4b7a-a2b1-1579cac22552","Type":"ContainerDied","Data":"25bd87e25e5b13f663de714f70bc3747cc81e9b118039d657abb521a45be18d7"} Dec 06 07:38:56 crc kubenswrapper[4809]: I1206 07:38:56.658220 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7xn64" event={"ID":"09aa0663-ccc7-4b7a-a2b1-1579cac22552","Type":"ContainerStarted","Data":"98a3130c6edf83b3735df7da6ca99148dce88322d8be7ef734f0a88c757f9d3d"} Dec 06 07:38:56 crc kubenswrapper[4809]: I1206 07:38:56.661333 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 07:38:57 crc kubenswrapper[4809]: I1206 07:38:57.668519 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7xn64" event={"ID":"09aa0663-ccc7-4b7a-a2b1-1579cac22552","Type":"ContainerStarted","Data":"57503d86d77cd14da2d00cff65acbee8c8455eabe6968539cb615e960b2439c7"} Dec 06 07:39:00 crc kubenswrapper[4809]: I1206 07:39:00.703849 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd462lbf" Dec 06 07:39:02 crc kubenswrapper[4809]: I1206 07:39:02.770613 4809 generic.go:334] "Generic (PLEG): container finished" podID="09aa0663-ccc7-4b7a-a2b1-1579cac22552" containerID="57503d86d77cd14da2d00cff65acbee8c8455eabe6968539cb615e960b2439c7" exitCode=0 Dec 06 07:39:02 crc kubenswrapper[4809]: I1206 07:39:02.771190 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7xn64" event={"ID":"09aa0663-ccc7-4b7a-a2b1-1579cac22552","Type":"ContainerDied","Data":"57503d86d77cd14da2d00cff65acbee8c8455eabe6968539cb615e960b2439c7"} Dec 06 07:39:03 crc kubenswrapper[4809]: I1206 07:39:03.786330 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7xn64" event={"ID":"09aa0663-ccc7-4b7a-a2b1-1579cac22552","Type":"ContainerStarted","Data":"2c816d95a9ed60acac9e9464645f73371e22da30576db44c5037302d97559ebb"} Dec 06 07:39:03 crc kubenswrapper[4809]: I1206 07:39:03.807045 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7xn64" podStartSLOduration=3.221947366 podStartE2EDuration="9.807020107s" podCreationTimestamp="2025-12-06 07:38:54 +0000 UTC" firstStartedPulling="2025-12-06 07:38:56.661029557 +0000 UTC m=+6461.550012499" lastFinishedPulling="2025-12-06 07:39:03.246102298 +0000 UTC m=+6468.135085240" observedRunningTime="2025-12-06 07:39:03.804659133 +0000 UTC m=+6468.693642075" watchObservedRunningTime="2025-12-06 07:39:03.807020107 +0000 UTC m=+6468.696003049" Dec 06 07:39:04 crc kubenswrapper[4809]: I1206 07:39:04.649874 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7trbt" Dec 06 07:39:05 crc kubenswrapper[4809]: I1206 07:39:05.142627 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7xn64" Dec 06 07:39:05 crc kubenswrapper[4809]: I1206 07:39:05.143562 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7xn64" Dec 06 07:39:05 crc kubenswrapper[4809]: I1206 07:39:05.168953 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-lqkvl" Dec 06 07:39:06 crc kubenswrapper[4809]: I1206 07:39:06.202468 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-7xn64" podUID="09aa0663-ccc7-4b7a-a2b1-1579cac22552" containerName="registry-server" probeResult="failure" output=< Dec 06 07:39:06 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Dec 06 07:39:06 crc kubenswrapper[4809]: > Dec 06 07:39:15 crc kubenswrapper[4809]: I1206 07:39:15.728009 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7xn64" Dec 06 07:39:15 crc kubenswrapper[4809]: I1206 07:39:15.788808 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7xn64" Dec 06 07:39:15 crc kubenswrapper[4809]: I1206 07:39:15.972883 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7xn64"] Dec 06 07:39:17 crc kubenswrapper[4809]: I1206 07:39:17.430983 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7xn64" podUID="09aa0663-ccc7-4b7a-a2b1-1579cac22552" containerName="registry-server" containerID="cri-o://2c816d95a9ed60acac9e9464645f73371e22da30576db44c5037302d97559ebb" gracePeriod=2 Dec 06 07:39:18 crc kubenswrapper[4809]: I1206 07:39:18.446088 4809 generic.go:334] "Generic (PLEG): container finished" podID="09aa0663-ccc7-4b7a-a2b1-1579cac22552" containerID="2c816d95a9ed60acac9e9464645f73371e22da30576db44c5037302d97559ebb" exitCode=0 Dec 06 07:39:18 crc kubenswrapper[4809]: I1206 07:39:18.446217 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7xn64" event={"ID":"09aa0663-ccc7-4b7a-a2b1-1579cac22552","Type":"ContainerDied","Data":"2c816d95a9ed60acac9e9464645f73371e22da30576db44c5037302d97559ebb"} Dec 06 07:39:18 crc kubenswrapper[4809]: I1206 07:39:18.588456 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7xn64" Dec 06 07:39:18 crc kubenswrapper[4809]: I1206 07:39:18.742458 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qgbpv\" (UniqueName: \"kubernetes.io/projected/09aa0663-ccc7-4b7a-a2b1-1579cac22552-kube-api-access-qgbpv\") pod \"09aa0663-ccc7-4b7a-a2b1-1579cac22552\" (UID: \"09aa0663-ccc7-4b7a-a2b1-1579cac22552\") " Dec 06 07:39:18 crc kubenswrapper[4809]: I1206 07:39:18.742748 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09aa0663-ccc7-4b7a-a2b1-1579cac22552-utilities\") pod \"09aa0663-ccc7-4b7a-a2b1-1579cac22552\" (UID: \"09aa0663-ccc7-4b7a-a2b1-1579cac22552\") " Dec 06 07:39:18 crc kubenswrapper[4809]: I1206 07:39:18.742795 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09aa0663-ccc7-4b7a-a2b1-1579cac22552-catalog-content\") pod \"09aa0663-ccc7-4b7a-a2b1-1579cac22552\" (UID: \"09aa0663-ccc7-4b7a-a2b1-1579cac22552\") " Dec 06 07:39:18 crc kubenswrapper[4809]: I1206 07:39:18.744269 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09aa0663-ccc7-4b7a-a2b1-1579cac22552-utilities" (OuterVolumeSpecName: "utilities") pod "09aa0663-ccc7-4b7a-a2b1-1579cac22552" (UID: "09aa0663-ccc7-4b7a-a2b1-1579cac22552"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:39:18 crc kubenswrapper[4809]: I1206 07:39:18.754136 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09aa0663-ccc7-4b7a-a2b1-1579cac22552-kube-api-access-qgbpv" (OuterVolumeSpecName: "kube-api-access-qgbpv") pod "09aa0663-ccc7-4b7a-a2b1-1579cac22552" (UID: "09aa0663-ccc7-4b7a-a2b1-1579cac22552"). InnerVolumeSpecName "kube-api-access-qgbpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:39:18 crc kubenswrapper[4809]: I1206 07:39:18.795485 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09aa0663-ccc7-4b7a-a2b1-1579cac22552-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "09aa0663-ccc7-4b7a-a2b1-1579cac22552" (UID: "09aa0663-ccc7-4b7a-a2b1-1579cac22552"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:39:18 crc kubenswrapper[4809]: I1206 07:39:18.845708 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09aa0663-ccc7-4b7a-a2b1-1579cac22552-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 07:39:18 crc kubenswrapper[4809]: I1206 07:39:18.845754 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09aa0663-ccc7-4b7a-a2b1-1579cac22552-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 07:39:18 crc kubenswrapper[4809]: I1206 07:39:18.845766 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qgbpv\" (UniqueName: \"kubernetes.io/projected/09aa0663-ccc7-4b7a-a2b1-1579cac22552-kube-api-access-qgbpv\") on node \"crc\" DevicePath \"\"" Dec 06 07:39:19 crc kubenswrapper[4809]: I1206 07:39:19.489079 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7xn64" event={"ID":"09aa0663-ccc7-4b7a-a2b1-1579cac22552","Type":"ContainerDied","Data":"98a3130c6edf83b3735df7da6ca99148dce88322d8be7ef734f0a88c757f9d3d"} Dec 06 07:39:19 crc kubenswrapper[4809]: I1206 07:39:19.489424 4809 scope.go:117] "RemoveContainer" containerID="2c816d95a9ed60acac9e9464645f73371e22da30576db44c5037302d97559ebb" Dec 06 07:39:19 crc kubenswrapper[4809]: I1206 07:39:19.489164 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7xn64" Dec 06 07:39:19 crc kubenswrapper[4809]: I1206 07:39:19.526446 4809 scope.go:117] "RemoveContainer" containerID="57503d86d77cd14da2d00cff65acbee8c8455eabe6968539cb615e960b2439c7" Dec 06 07:39:19 crc kubenswrapper[4809]: I1206 07:39:19.529864 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7xn64"] Dec 06 07:39:19 crc kubenswrapper[4809]: I1206 07:39:19.553633 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7xn64"] Dec 06 07:39:19 crc kubenswrapper[4809]: I1206 07:39:19.577269 4809 scope.go:117] "RemoveContainer" containerID="25bd87e25e5b13f663de714f70bc3747cc81e9b118039d657abb521a45be18d7" Dec 06 07:39:21 crc kubenswrapper[4809]: I1206 07:39:21.402430 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09aa0663-ccc7-4b7a-a2b1-1579cac22552" path="/var/lib/kubelet/pods/09aa0663-ccc7-4b7a-a2b1-1579cac22552/volumes" Dec 06 07:39:32 crc kubenswrapper[4809]: I1206 07:39:32.651050 4809 generic.go:334] "Generic (PLEG): container finished" podID="3603c2a8-988f-4ecf-ad1b-51b74176a0a1" containerID="e9785bbc5d7c66af1d9400512f3e1cf9163eba7b5c7fffd5163e48af26640ad4" exitCode=0 Dec 06 07:39:32 crc kubenswrapper[4809]: I1206 07:39:32.651176 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bn94p/must-gather-t7bfm" event={"ID":"3603c2a8-988f-4ecf-ad1b-51b74176a0a1","Type":"ContainerDied","Data":"e9785bbc5d7c66af1d9400512f3e1cf9163eba7b5c7fffd5163e48af26640ad4"} Dec 06 07:39:32 crc kubenswrapper[4809]: I1206 07:39:32.652428 4809 scope.go:117] "RemoveContainer" containerID="e9785bbc5d7c66af1d9400512f3e1cf9163eba7b5c7fffd5163e48af26640ad4" Dec 06 07:39:33 crc kubenswrapper[4809]: I1206 07:39:33.392772 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-bn94p_must-gather-t7bfm_3603c2a8-988f-4ecf-ad1b-51b74176a0a1/gather/0.log" Dec 06 07:39:45 crc kubenswrapper[4809]: I1206 07:39:45.738657 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-bn94p/must-gather-t7bfm"] Dec 06 07:39:45 crc kubenswrapper[4809]: I1206 07:39:45.739398 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-bn94p/must-gather-t7bfm" podUID="3603c2a8-988f-4ecf-ad1b-51b74176a0a1" containerName="copy" containerID="cri-o://58bf556966fa07e4d904e4d1811fec480bf1a6508c911e97b07a41f00dc004e0" gracePeriod=2 Dec 06 07:39:45 crc kubenswrapper[4809]: I1206 07:39:45.752579 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-bn94p/must-gather-t7bfm"] Dec 06 07:39:46 crc kubenswrapper[4809]: I1206 07:39:46.510669 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-bn94p_must-gather-t7bfm_3603c2a8-988f-4ecf-ad1b-51b74176a0a1/copy/0.log" Dec 06 07:39:46 crc kubenswrapper[4809]: I1206 07:39:46.511706 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bn94p/must-gather-t7bfm" Dec 06 07:39:46 crc kubenswrapper[4809]: I1206 07:39:46.680425 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3603c2a8-988f-4ecf-ad1b-51b74176a0a1-must-gather-output\") pod \"3603c2a8-988f-4ecf-ad1b-51b74176a0a1\" (UID: \"3603c2a8-988f-4ecf-ad1b-51b74176a0a1\") " Dec 06 07:39:46 crc kubenswrapper[4809]: I1206 07:39:46.680510 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rghj2\" (UniqueName: \"kubernetes.io/projected/3603c2a8-988f-4ecf-ad1b-51b74176a0a1-kube-api-access-rghj2\") pod \"3603c2a8-988f-4ecf-ad1b-51b74176a0a1\" (UID: \"3603c2a8-988f-4ecf-ad1b-51b74176a0a1\") " Dec 06 07:39:46 crc kubenswrapper[4809]: I1206 07:39:46.694910 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3603c2a8-988f-4ecf-ad1b-51b74176a0a1-kube-api-access-rghj2" (OuterVolumeSpecName: "kube-api-access-rghj2") pod "3603c2a8-988f-4ecf-ad1b-51b74176a0a1" (UID: "3603c2a8-988f-4ecf-ad1b-51b74176a0a1"). InnerVolumeSpecName "kube-api-access-rghj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:39:46 crc kubenswrapper[4809]: I1206 07:39:46.783516 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rghj2\" (UniqueName: \"kubernetes.io/projected/3603c2a8-988f-4ecf-ad1b-51b74176a0a1-kube-api-access-rghj2\") on node \"crc\" DevicePath \"\"" Dec 06 07:39:46 crc kubenswrapper[4809]: I1206 07:39:46.813745 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-bn94p_must-gather-t7bfm_3603c2a8-988f-4ecf-ad1b-51b74176a0a1/copy/0.log" Dec 06 07:39:46 crc kubenswrapper[4809]: I1206 07:39:46.814381 4809 generic.go:334] "Generic (PLEG): container finished" podID="3603c2a8-988f-4ecf-ad1b-51b74176a0a1" containerID="58bf556966fa07e4d904e4d1811fec480bf1a6508c911e97b07a41f00dc004e0" exitCode=143 Dec 06 07:39:46 crc kubenswrapper[4809]: I1206 07:39:46.814427 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bn94p/must-gather-t7bfm" Dec 06 07:39:46 crc kubenswrapper[4809]: I1206 07:39:46.814444 4809 scope.go:117] "RemoveContainer" containerID="58bf556966fa07e4d904e4d1811fec480bf1a6508c911e97b07a41f00dc004e0" Dec 06 07:39:46 crc kubenswrapper[4809]: I1206 07:39:46.841379 4809 scope.go:117] "RemoveContainer" containerID="e9785bbc5d7c66af1d9400512f3e1cf9163eba7b5c7fffd5163e48af26640ad4" Dec 06 07:39:46 crc kubenswrapper[4809]: I1206 07:39:46.874790 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3603c2a8-988f-4ecf-ad1b-51b74176a0a1-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "3603c2a8-988f-4ecf-ad1b-51b74176a0a1" (UID: "3603c2a8-988f-4ecf-ad1b-51b74176a0a1"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:39:46 crc kubenswrapper[4809]: I1206 07:39:46.885536 4809 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3603c2a8-988f-4ecf-ad1b-51b74176a0a1-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 06 07:39:46 crc kubenswrapper[4809]: I1206 07:39:46.899021 4809 scope.go:117] "RemoveContainer" containerID="58bf556966fa07e4d904e4d1811fec480bf1a6508c911e97b07a41f00dc004e0" Dec 06 07:39:46 crc kubenswrapper[4809]: E1206 07:39:46.899762 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58bf556966fa07e4d904e4d1811fec480bf1a6508c911e97b07a41f00dc004e0\": container with ID starting with 58bf556966fa07e4d904e4d1811fec480bf1a6508c911e97b07a41f00dc004e0 not found: ID does not exist" containerID="58bf556966fa07e4d904e4d1811fec480bf1a6508c911e97b07a41f00dc004e0" Dec 06 07:39:46 crc kubenswrapper[4809]: I1206 07:39:46.899926 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58bf556966fa07e4d904e4d1811fec480bf1a6508c911e97b07a41f00dc004e0"} err="failed to get container status \"58bf556966fa07e4d904e4d1811fec480bf1a6508c911e97b07a41f00dc004e0\": rpc error: code = NotFound desc = could not find container \"58bf556966fa07e4d904e4d1811fec480bf1a6508c911e97b07a41f00dc004e0\": container with ID starting with 58bf556966fa07e4d904e4d1811fec480bf1a6508c911e97b07a41f00dc004e0 not found: ID does not exist" Dec 06 07:39:46 crc kubenswrapper[4809]: I1206 07:39:46.900101 4809 scope.go:117] "RemoveContainer" containerID="e9785bbc5d7c66af1d9400512f3e1cf9163eba7b5c7fffd5163e48af26640ad4" Dec 06 07:39:46 crc kubenswrapper[4809]: E1206 07:39:46.900490 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9785bbc5d7c66af1d9400512f3e1cf9163eba7b5c7fffd5163e48af26640ad4\": container with ID starting with e9785bbc5d7c66af1d9400512f3e1cf9163eba7b5c7fffd5163e48af26640ad4 not found: ID does not exist" containerID="e9785bbc5d7c66af1d9400512f3e1cf9163eba7b5c7fffd5163e48af26640ad4" Dec 06 07:39:46 crc kubenswrapper[4809]: I1206 07:39:46.900534 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9785bbc5d7c66af1d9400512f3e1cf9163eba7b5c7fffd5163e48af26640ad4"} err="failed to get container status \"e9785bbc5d7c66af1d9400512f3e1cf9163eba7b5c7fffd5163e48af26640ad4\": rpc error: code = NotFound desc = could not find container \"e9785bbc5d7c66af1d9400512f3e1cf9163eba7b5c7fffd5163e48af26640ad4\": container with ID starting with e9785bbc5d7c66af1d9400512f3e1cf9163eba7b5c7fffd5163e48af26640ad4 not found: ID does not exist" Dec 06 07:39:47 crc kubenswrapper[4809]: I1206 07:39:47.414895 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3603c2a8-988f-4ecf-ad1b-51b74176a0a1" path="/var/lib/kubelet/pods/3603c2a8-988f-4ecf-ad1b-51b74176a0a1/volumes" Dec 06 07:40:20 crc kubenswrapper[4809]: I1206 07:40:20.931969 4809 trace.go:236] Trace[275183976]: "Calculate volume metrics of config-data-generated for pod openstack/openstack-cell1-galera-0" (06-Dec-2025 07:40:19.647) (total time: 1284ms): Dec 06 07:40:20 crc kubenswrapper[4809]: Trace[275183976]: [1.284578052s] [1.284578052s] END Dec 06 07:40:34 crc kubenswrapper[4809]: I1206 07:40:34.496965 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:40:34 crc kubenswrapper[4809]: I1206 07:40:34.497423 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:41:04 crc kubenswrapper[4809]: I1206 07:41:04.496293 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:41:04 crc kubenswrapper[4809]: I1206 07:41:04.497095 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:41:34 crc kubenswrapper[4809]: I1206 07:41:34.496803 4809 patch_prober.go:28] interesting pod/machine-config-daemon-npms2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:41:34 crc kubenswrapper[4809]: I1206 07:41:34.497321 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:41:34 crc kubenswrapper[4809]: I1206 07:41:34.497368 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-npms2" Dec 06 07:41:34 crc kubenswrapper[4809]: I1206 07:41:34.498328 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4b28141798e8be5bf2c8cdfd3390a3ddfd37326a66c4ad4407752e73538ea965"} pod="openshift-machine-config-operator/machine-config-daemon-npms2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 07:41:34 crc kubenswrapper[4809]: I1206 07:41:34.498389 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerName="machine-config-daemon" containerID="cri-o://4b28141798e8be5bf2c8cdfd3390a3ddfd37326a66c4ad4407752e73538ea965" gracePeriod=600 Dec 06 07:41:38 crc kubenswrapper[4809]: I1206 07:41:38.182884 4809 generic.go:334] "Generic (PLEG): container finished" podID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" containerID="4b28141798e8be5bf2c8cdfd3390a3ddfd37326a66c4ad4407752e73538ea965" exitCode=0 Dec 06 07:41:38 crc kubenswrapper[4809]: E1206 07:41:38.182990 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:41:38 crc kubenswrapper[4809]: I1206 07:41:38.183250 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-npms2" event={"ID":"cbf4cf62-024e-4703-a8b8-9aecda9cd26a","Type":"ContainerDied","Data":"4b28141798e8be5bf2c8cdfd3390a3ddfd37326a66c4ad4407752e73538ea965"} Dec 06 07:41:38 crc kubenswrapper[4809]: I1206 07:41:38.183771 4809 scope.go:117] "RemoveContainer" containerID="2b99cfee1a9995b9796e60fb6b1a787ce8771b97d7af6a38edfaa8224097d30d" Dec 06 07:41:39 crc kubenswrapper[4809]: I1206 07:41:39.196924 4809 scope.go:117] "RemoveContainer" containerID="4b28141798e8be5bf2c8cdfd3390a3ddfd37326a66c4ad4407752e73538ea965" Dec 06 07:41:39 crc kubenswrapper[4809]: E1206 07:41:39.197779 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:41:53 crc kubenswrapper[4809]: I1206 07:41:53.388669 4809 scope.go:117] "RemoveContainer" containerID="4b28141798e8be5bf2c8cdfd3390a3ddfd37326a66c4ad4407752e73538ea965" Dec 06 07:41:53 crc kubenswrapper[4809]: E1206 07:41:53.390056 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:42:08 crc kubenswrapper[4809]: I1206 07:42:08.389222 4809 scope.go:117] "RemoveContainer" containerID="4b28141798e8be5bf2c8cdfd3390a3ddfd37326a66c4ad4407752e73538ea965" Dec 06 07:42:08 crc kubenswrapper[4809]: E1206 07:42:08.390121 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:42:22 crc kubenswrapper[4809]: I1206 07:42:22.388659 4809 scope.go:117] "RemoveContainer" containerID="4b28141798e8be5bf2c8cdfd3390a3ddfd37326a66c4ad4407752e73538ea965" Dec 06 07:42:22 crc kubenswrapper[4809]: E1206 07:42:22.391793 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:42:29 crc kubenswrapper[4809]: I1206 07:42:25.302148 4809 patch_prober.go:28] interesting pod/authentication-operator-69f744f599-7qpjs container/authentication-operator namespace/openshift-authentication-operator: Liveness probe status=failure output="Get \"https://10.217.0.18:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:42:29 crc kubenswrapper[4809]: I1206 07:42:25.302567 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication-operator/authentication-operator-69f744f599-7qpjs" podUID="2a1ed741-489c-455f-a344-404bbab66c22" containerName="authentication-operator" probeResult="failure" output="Get \"https://10.217.0.18:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:42:29 crc kubenswrapper[4809]: I1206 07:42:25.431780 4809 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-pkq7h container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.23:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:42:29 crc kubenswrapper[4809]: I1206 07:42:25.431851 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pkq7h" podUID="00bf9f05-7bc3-40e1-a2e9-1af1bf93f014" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.23:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 07:42:29 crc kubenswrapper[4809]: I1206 07:42:28.395121 4809 patch_prober.go:28] interesting pod/logging-loki-gateway-65498c4f8f-9s8qq container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.54:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:42:29 crc kubenswrapper[4809]: I1206 07:42:28.395562 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq" podUID="ef8f3e7e-57ea-42b8-a777-1778e5ed975b" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.54:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:42:29 crc kubenswrapper[4809]: I1206 07:42:28.395191 4809 patch_prober.go:28] interesting pod/logging-loki-gateway-65498c4f8f-9s8qq container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.54:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 07:42:29 crc kubenswrapper[4809]: I1206 07:42:28.395638 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-65498c4f8f-9s8qq" podUID="ef8f3e7e-57ea-42b8-a777-1778e5ed975b" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.54:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 07:42:29 crc kubenswrapper[4809]: E1206 07:42:29.665958 4809 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="4.278s" Dec 06 07:42:29 crc kubenswrapper[4809]: I1206 07:42:29.759171 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qn4w9"] Dec 06 07:42:29 crc kubenswrapper[4809]: E1206 07:42:29.760668 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09aa0663-ccc7-4b7a-a2b1-1579cac22552" containerName="registry-server" Dec 06 07:42:29 crc kubenswrapper[4809]: I1206 07:42:29.760687 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="09aa0663-ccc7-4b7a-a2b1-1579cac22552" containerName="registry-server" Dec 06 07:42:29 crc kubenswrapper[4809]: E1206 07:42:29.760695 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09aa0663-ccc7-4b7a-a2b1-1579cac22552" containerName="extract-content" Dec 06 07:42:29 crc kubenswrapper[4809]: I1206 07:42:29.760702 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="09aa0663-ccc7-4b7a-a2b1-1579cac22552" containerName="extract-content" Dec 06 07:42:29 crc kubenswrapper[4809]: E1206 07:42:29.760723 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3603c2a8-988f-4ecf-ad1b-51b74176a0a1" containerName="gather" Dec 06 07:42:29 crc kubenswrapper[4809]: I1206 07:42:29.760729 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3603c2a8-988f-4ecf-ad1b-51b74176a0a1" containerName="gather" Dec 06 07:42:29 crc kubenswrapper[4809]: E1206 07:42:29.760744 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09aa0663-ccc7-4b7a-a2b1-1579cac22552" containerName="extract-utilities" Dec 06 07:42:29 crc kubenswrapper[4809]: I1206 07:42:29.760750 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="09aa0663-ccc7-4b7a-a2b1-1579cac22552" containerName="extract-utilities" Dec 06 07:42:29 crc kubenswrapper[4809]: E1206 07:42:29.760763 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3603c2a8-988f-4ecf-ad1b-51b74176a0a1" containerName="copy" Dec 06 07:42:29 crc kubenswrapper[4809]: I1206 07:42:29.760768 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3603c2a8-988f-4ecf-ad1b-51b74176a0a1" containerName="copy" Dec 06 07:42:29 crc kubenswrapper[4809]: I1206 07:42:29.761165 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="09aa0663-ccc7-4b7a-a2b1-1579cac22552" containerName="registry-server" Dec 06 07:42:29 crc kubenswrapper[4809]: I1206 07:42:29.761202 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="3603c2a8-988f-4ecf-ad1b-51b74176a0a1" containerName="gather" Dec 06 07:42:29 crc kubenswrapper[4809]: I1206 07:42:29.761213 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="3603c2a8-988f-4ecf-ad1b-51b74176a0a1" containerName="copy" Dec 06 07:42:29 crc kubenswrapper[4809]: I1206 07:42:29.762801 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qn4w9"] Dec 06 07:42:29 crc kubenswrapper[4809]: I1206 07:42:29.762882 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qn4w9" Dec 06 07:42:29 crc kubenswrapper[4809]: I1206 07:42:29.878346 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c633de75-e128-4c94-929e-37ed0421f065-catalog-content\") pod \"redhat-operators-qn4w9\" (UID: \"c633de75-e128-4c94-929e-37ed0421f065\") " pod="openshift-marketplace/redhat-operators-qn4w9" Dec 06 07:42:29 crc kubenswrapper[4809]: I1206 07:42:29.878890 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trqnw\" (UniqueName: \"kubernetes.io/projected/c633de75-e128-4c94-929e-37ed0421f065-kube-api-access-trqnw\") pod \"redhat-operators-qn4w9\" (UID: \"c633de75-e128-4c94-929e-37ed0421f065\") " pod="openshift-marketplace/redhat-operators-qn4w9" Dec 06 07:42:29 crc kubenswrapper[4809]: I1206 07:42:29.878974 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c633de75-e128-4c94-929e-37ed0421f065-utilities\") pod \"redhat-operators-qn4w9\" (UID: \"c633de75-e128-4c94-929e-37ed0421f065\") " pod="openshift-marketplace/redhat-operators-qn4w9" Dec 06 07:42:29 crc kubenswrapper[4809]: I1206 07:42:29.980925 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trqnw\" (UniqueName: \"kubernetes.io/projected/c633de75-e128-4c94-929e-37ed0421f065-kube-api-access-trqnw\") pod \"redhat-operators-qn4w9\" (UID: \"c633de75-e128-4c94-929e-37ed0421f065\") " pod="openshift-marketplace/redhat-operators-qn4w9" Dec 06 07:42:29 crc kubenswrapper[4809]: I1206 07:42:29.981237 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c633de75-e128-4c94-929e-37ed0421f065-utilities\") pod \"redhat-operators-qn4w9\" (UID: \"c633de75-e128-4c94-929e-37ed0421f065\") " pod="openshift-marketplace/redhat-operators-qn4w9" Dec 06 07:42:29 crc kubenswrapper[4809]: I1206 07:42:29.981445 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c633de75-e128-4c94-929e-37ed0421f065-catalog-content\") pod \"redhat-operators-qn4w9\" (UID: \"c633de75-e128-4c94-929e-37ed0421f065\") " pod="openshift-marketplace/redhat-operators-qn4w9" Dec 06 07:42:29 crc kubenswrapper[4809]: I1206 07:42:29.982415 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c633de75-e128-4c94-929e-37ed0421f065-utilities\") pod \"redhat-operators-qn4w9\" (UID: \"c633de75-e128-4c94-929e-37ed0421f065\") " pod="openshift-marketplace/redhat-operators-qn4w9" Dec 06 07:42:29 crc kubenswrapper[4809]: I1206 07:42:29.982607 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c633de75-e128-4c94-929e-37ed0421f065-catalog-content\") pod \"redhat-operators-qn4w9\" (UID: \"c633de75-e128-4c94-929e-37ed0421f065\") " pod="openshift-marketplace/redhat-operators-qn4w9" Dec 06 07:42:30 crc kubenswrapper[4809]: I1206 07:42:30.014464 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trqnw\" (UniqueName: \"kubernetes.io/projected/c633de75-e128-4c94-929e-37ed0421f065-kube-api-access-trqnw\") pod \"redhat-operators-qn4w9\" (UID: \"c633de75-e128-4c94-929e-37ed0421f065\") " pod="openshift-marketplace/redhat-operators-qn4w9" Dec 06 07:42:30 crc kubenswrapper[4809]: I1206 07:42:30.087615 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qn4w9" Dec 06 07:42:30 crc kubenswrapper[4809]: I1206 07:42:30.621700 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qn4w9"] Dec 06 07:42:30 crc kubenswrapper[4809]: I1206 07:42:30.643898 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qn4w9" event={"ID":"c633de75-e128-4c94-929e-37ed0421f065","Type":"ContainerStarted","Data":"018d2e9cab51c3efbdf83e0f4eb5281bfd318238fbe39dd2e8c7778fd70d74ed"} Dec 06 07:42:31 crc kubenswrapper[4809]: I1206 07:42:31.663308 4809 generic.go:334] "Generic (PLEG): container finished" podID="c633de75-e128-4c94-929e-37ed0421f065" containerID="c27b2a4b6415615498a0e138f5144ca828097a4a83ee1a603660ddb02751af2b" exitCode=0 Dec 06 07:42:31 crc kubenswrapper[4809]: I1206 07:42:31.663672 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qn4w9" event={"ID":"c633de75-e128-4c94-929e-37ed0421f065","Type":"ContainerDied","Data":"c27b2a4b6415615498a0e138f5144ca828097a4a83ee1a603660ddb02751af2b"} Dec 06 07:42:35 crc kubenswrapper[4809]: I1206 07:42:35.728335 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qn4w9" event={"ID":"c633de75-e128-4c94-929e-37ed0421f065","Type":"ContainerStarted","Data":"68da5872bc1157b9945f9becc4fecaf706c0a36674eab968e98dbd4213e7b39e"} Dec 06 07:42:36 crc kubenswrapper[4809]: I1206 07:42:36.994361 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="hostpath-provisioner/csi-hostpathplugin-p2lkt" podUID="c598f580-a438-4015-9a8a-f98b3b2b0b3d" containerName="hostpath-provisioner" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 07:42:37 crc kubenswrapper[4809]: I1206 07:42:37.389294 4809 scope.go:117] "RemoveContainer" containerID="4b28141798e8be5bf2c8cdfd3390a3ddfd37326a66c4ad4407752e73538ea965" Dec 06 07:42:37 crc kubenswrapper[4809]: E1206 07:42:37.389616 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:42:40 crc kubenswrapper[4809]: I1206 07:42:40.793748 4809 generic.go:334] "Generic (PLEG): container finished" podID="c633de75-e128-4c94-929e-37ed0421f065" containerID="68da5872bc1157b9945f9becc4fecaf706c0a36674eab968e98dbd4213e7b39e" exitCode=0 Dec 06 07:42:40 crc kubenswrapper[4809]: I1206 07:42:40.793814 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qn4w9" event={"ID":"c633de75-e128-4c94-929e-37ed0421f065","Type":"ContainerDied","Data":"68da5872bc1157b9945f9becc4fecaf706c0a36674eab968e98dbd4213e7b39e"} Dec 06 07:42:45 crc kubenswrapper[4809]: I1206 07:42:45.886630 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qn4w9" event={"ID":"c633de75-e128-4c94-929e-37ed0421f065","Type":"ContainerStarted","Data":"0498a6b058c06b918133f22ae9e6dd33f016cde1f09ade5fb27362ef54c41f53"} Dec 06 07:42:45 crc kubenswrapper[4809]: I1206 07:42:45.923626 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qn4w9" podStartSLOduration=5.503548321 podStartE2EDuration="18.923602735s" podCreationTimestamp="2025-12-06 07:42:27 +0000 UTC" firstStartedPulling="2025-12-06 07:42:31.665870528 +0000 UTC m=+6676.554853470" lastFinishedPulling="2025-12-06 07:42:45.085924922 +0000 UTC m=+6689.974907884" observedRunningTime="2025-12-06 07:42:45.908408654 +0000 UTC m=+6690.797391606" watchObservedRunningTime="2025-12-06 07:42:45.923602735 +0000 UTC m=+6690.812585687" Dec 06 07:42:49 crc kubenswrapper[4809]: I1206 07:42:49.389464 4809 scope.go:117] "RemoveContainer" containerID="4b28141798e8be5bf2c8cdfd3390a3ddfd37326a66c4ad4407752e73538ea965" Dec 06 07:42:49 crc kubenswrapper[4809]: E1206 07:42:49.390338 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:42:50 crc kubenswrapper[4809]: I1206 07:42:50.088462 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qn4w9" Dec 06 07:42:50 crc kubenswrapper[4809]: I1206 07:42:50.088831 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qn4w9" Dec 06 07:42:51 crc kubenswrapper[4809]: I1206 07:42:51.144423 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qn4w9" podUID="c633de75-e128-4c94-929e-37ed0421f065" containerName="registry-server" probeResult="failure" output=< Dec 06 07:42:51 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Dec 06 07:42:51 crc kubenswrapper[4809]: > Dec 06 07:43:00 crc kubenswrapper[4809]: I1206 07:43:00.149663 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qn4w9" Dec 06 07:43:00 crc kubenswrapper[4809]: I1206 07:43:00.240226 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qn4w9" Dec 06 07:43:00 crc kubenswrapper[4809]: I1206 07:43:00.401763 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qn4w9"] Dec 06 07:43:01 crc kubenswrapper[4809]: I1206 07:43:01.389543 4809 scope.go:117] "RemoveContainer" containerID="4b28141798e8be5bf2c8cdfd3390a3ddfd37326a66c4ad4407752e73538ea965" Dec 06 07:43:01 crc kubenswrapper[4809]: E1206 07:43:01.391882 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:43:02 crc kubenswrapper[4809]: I1206 07:43:02.103545 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qn4w9" podUID="c633de75-e128-4c94-929e-37ed0421f065" containerName="registry-server" containerID="cri-o://0498a6b058c06b918133f22ae9e6dd33f016cde1f09ade5fb27362ef54c41f53" gracePeriod=2 Dec 06 07:43:02 crc kubenswrapper[4809]: I1206 07:43:02.865885 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qn4w9" Dec 06 07:43:03 crc kubenswrapper[4809]: I1206 07:43:03.032449 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c633de75-e128-4c94-929e-37ed0421f065-utilities\") pod \"c633de75-e128-4c94-929e-37ed0421f065\" (UID: \"c633de75-e128-4c94-929e-37ed0421f065\") " Dec 06 07:43:03 crc kubenswrapper[4809]: I1206 07:43:03.032896 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c633de75-e128-4c94-929e-37ed0421f065-catalog-content\") pod \"c633de75-e128-4c94-929e-37ed0421f065\" (UID: \"c633de75-e128-4c94-929e-37ed0421f065\") " Dec 06 07:43:03 crc kubenswrapper[4809]: I1206 07:43:03.033154 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-trqnw\" (UniqueName: \"kubernetes.io/projected/c633de75-e128-4c94-929e-37ed0421f065-kube-api-access-trqnw\") pod \"c633de75-e128-4c94-929e-37ed0421f065\" (UID: \"c633de75-e128-4c94-929e-37ed0421f065\") " Dec 06 07:43:03 crc kubenswrapper[4809]: I1206 07:43:03.035047 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c633de75-e128-4c94-929e-37ed0421f065-utilities" (OuterVolumeSpecName: "utilities") pod "c633de75-e128-4c94-929e-37ed0421f065" (UID: "c633de75-e128-4c94-929e-37ed0421f065"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:43:03 crc kubenswrapper[4809]: I1206 07:43:03.042067 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c633de75-e128-4c94-929e-37ed0421f065-kube-api-access-trqnw" (OuterVolumeSpecName: "kube-api-access-trqnw") pod "c633de75-e128-4c94-929e-37ed0421f065" (UID: "c633de75-e128-4c94-929e-37ed0421f065"). InnerVolumeSpecName "kube-api-access-trqnw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:43:03 crc kubenswrapper[4809]: I1206 07:43:03.119807 4809 generic.go:334] "Generic (PLEG): container finished" podID="c633de75-e128-4c94-929e-37ed0421f065" containerID="0498a6b058c06b918133f22ae9e6dd33f016cde1f09ade5fb27362ef54c41f53" exitCode=0 Dec 06 07:43:03 crc kubenswrapper[4809]: I1206 07:43:03.120092 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qn4w9" event={"ID":"c633de75-e128-4c94-929e-37ed0421f065","Type":"ContainerDied","Data":"0498a6b058c06b918133f22ae9e6dd33f016cde1f09ade5fb27362ef54c41f53"} Dec 06 07:43:03 crc kubenswrapper[4809]: I1206 07:43:03.120225 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qn4w9" event={"ID":"c633de75-e128-4c94-929e-37ed0421f065","Type":"ContainerDied","Data":"018d2e9cab51c3efbdf83e0f4eb5281bfd318238fbe39dd2e8c7778fd70d74ed"} Dec 06 07:43:03 crc kubenswrapper[4809]: I1206 07:43:03.120309 4809 scope.go:117] "RemoveContainer" containerID="0498a6b058c06b918133f22ae9e6dd33f016cde1f09ade5fb27362ef54c41f53" Dec 06 07:43:03 crc kubenswrapper[4809]: I1206 07:43:03.120551 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qn4w9" Dec 06 07:43:03 crc kubenswrapper[4809]: I1206 07:43:03.136820 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c633de75-e128-4c94-929e-37ed0421f065-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 07:43:03 crc kubenswrapper[4809]: I1206 07:43:03.137005 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-trqnw\" (UniqueName: \"kubernetes.io/projected/c633de75-e128-4c94-929e-37ed0421f065-kube-api-access-trqnw\") on node \"crc\" DevicePath \"\"" Dec 06 07:43:03 crc kubenswrapper[4809]: I1206 07:43:03.157032 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c633de75-e128-4c94-929e-37ed0421f065-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c633de75-e128-4c94-929e-37ed0421f065" (UID: "c633de75-e128-4c94-929e-37ed0421f065"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:43:03 crc kubenswrapper[4809]: I1206 07:43:03.158125 4809 scope.go:117] "RemoveContainer" containerID="68da5872bc1157b9945f9becc4fecaf706c0a36674eab968e98dbd4213e7b39e" Dec 06 07:43:03 crc kubenswrapper[4809]: I1206 07:43:03.182062 4809 scope.go:117] "RemoveContainer" containerID="c27b2a4b6415615498a0e138f5144ca828097a4a83ee1a603660ddb02751af2b" Dec 06 07:43:03 crc kubenswrapper[4809]: I1206 07:43:03.239229 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c633de75-e128-4c94-929e-37ed0421f065-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 07:43:03 crc kubenswrapper[4809]: I1206 07:43:03.267385 4809 scope.go:117] "RemoveContainer" containerID="0498a6b058c06b918133f22ae9e6dd33f016cde1f09ade5fb27362ef54c41f53" Dec 06 07:43:03 crc kubenswrapper[4809]: E1206 07:43:03.268093 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0498a6b058c06b918133f22ae9e6dd33f016cde1f09ade5fb27362ef54c41f53\": container with ID starting with 0498a6b058c06b918133f22ae9e6dd33f016cde1f09ade5fb27362ef54c41f53 not found: ID does not exist" containerID="0498a6b058c06b918133f22ae9e6dd33f016cde1f09ade5fb27362ef54c41f53" Dec 06 07:43:03 crc kubenswrapper[4809]: I1206 07:43:03.268461 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0498a6b058c06b918133f22ae9e6dd33f016cde1f09ade5fb27362ef54c41f53"} err="failed to get container status \"0498a6b058c06b918133f22ae9e6dd33f016cde1f09ade5fb27362ef54c41f53\": rpc error: code = NotFound desc = could not find container \"0498a6b058c06b918133f22ae9e6dd33f016cde1f09ade5fb27362ef54c41f53\": container with ID starting with 0498a6b058c06b918133f22ae9e6dd33f016cde1f09ade5fb27362ef54c41f53 not found: ID does not exist" Dec 06 07:43:03 crc kubenswrapper[4809]: I1206 07:43:03.268497 4809 scope.go:117] "RemoveContainer" containerID="68da5872bc1157b9945f9becc4fecaf706c0a36674eab968e98dbd4213e7b39e" Dec 06 07:43:03 crc kubenswrapper[4809]: E1206 07:43:03.269574 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68da5872bc1157b9945f9becc4fecaf706c0a36674eab968e98dbd4213e7b39e\": container with ID starting with 68da5872bc1157b9945f9becc4fecaf706c0a36674eab968e98dbd4213e7b39e not found: ID does not exist" containerID="68da5872bc1157b9945f9becc4fecaf706c0a36674eab968e98dbd4213e7b39e" Dec 06 07:43:03 crc kubenswrapper[4809]: I1206 07:43:03.269741 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68da5872bc1157b9945f9becc4fecaf706c0a36674eab968e98dbd4213e7b39e"} err="failed to get container status \"68da5872bc1157b9945f9becc4fecaf706c0a36674eab968e98dbd4213e7b39e\": rpc error: code = NotFound desc = could not find container \"68da5872bc1157b9945f9becc4fecaf706c0a36674eab968e98dbd4213e7b39e\": container with ID starting with 68da5872bc1157b9945f9becc4fecaf706c0a36674eab968e98dbd4213e7b39e not found: ID does not exist" Dec 06 07:43:03 crc kubenswrapper[4809]: I1206 07:43:03.269865 4809 scope.go:117] "RemoveContainer" containerID="c27b2a4b6415615498a0e138f5144ca828097a4a83ee1a603660ddb02751af2b" Dec 06 07:43:03 crc kubenswrapper[4809]: E1206 07:43:03.270298 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c27b2a4b6415615498a0e138f5144ca828097a4a83ee1a603660ddb02751af2b\": container with ID starting with c27b2a4b6415615498a0e138f5144ca828097a4a83ee1a603660ddb02751af2b not found: ID does not exist" containerID="c27b2a4b6415615498a0e138f5144ca828097a4a83ee1a603660ddb02751af2b" Dec 06 07:43:03 crc kubenswrapper[4809]: I1206 07:43:03.270341 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c27b2a4b6415615498a0e138f5144ca828097a4a83ee1a603660ddb02751af2b"} err="failed to get container status \"c27b2a4b6415615498a0e138f5144ca828097a4a83ee1a603660ddb02751af2b\": rpc error: code = NotFound desc = could not find container \"c27b2a4b6415615498a0e138f5144ca828097a4a83ee1a603660ddb02751af2b\": container with ID starting with c27b2a4b6415615498a0e138f5144ca828097a4a83ee1a603660ddb02751af2b not found: ID does not exist" Dec 06 07:43:03 crc kubenswrapper[4809]: I1206 07:43:03.474955 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qn4w9"] Dec 06 07:43:03 crc kubenswrapper[4809]: I1206 07:43:03.490080 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qn4w9"] Dec 06 07:43:05 crc kubenswrapper[4809]: I1206 07:43:05.407546 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c633de75-e128-4c94-929e-37ed0421f065" path="/var/lib/kubelet/pods/c633de75-e128-4c94-929e-37ed0421f065/volumes" Dec 06 07:43:12 crc kubenswrapper[4809]: I1206 07:43:12.389244 4809 scope.go:117] "RemoveContainer" containerID="4b28141798e8be5bf2c8cdfd3390a3ddfd37326a66c4ad4407752e73538ea965" Dec 06 07:43:12 crc kubenswrapper[4809]: E1206 07:43:12.390182 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:43:26 crc kubenswrapper[4809]: I1206 07:43:26.388904 4809 scope.go:117] "RemoveContainer" containerID="4b28141798e8be5bf2c8cdfd3390a3ddfd37326a66c4ad4407752e73538ea965" Dec 06 07:43:26 crc kubenswrapper[4809]: E1206 07:43:26.390269 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:43:37 crc kubenswrapper[4809]: I1206 07:43:37.391677 4809 scope.go:117] "RemoveContainer" containerID="4b28141798e8be5bf2c8cdfd3390a3ddfd37326a66c4ad4407752e73538ea965" Dec 06 07:43:37 crc kubenswrapper[4809]: E1206 07:43:37.392749 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:43:51 crc kubenswrapper[4809]: I1206 07:43:51.388785 4809 scope.go:117] "RemoveContainer" containerID="4b28141798e8be5bf2c8cdfd3390a3ddfd37326a66c4ad4407752e73538ea965" Dec 06 07:43:51 crc kubenswrapper[4809]: E1206 07:43:51.389858 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:43:55 crc kubenswrapper[4809]: I1206 07:43:55.705820 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-w946g"] Dec 06 07:43:55 crc kubenswrapper[4809]: E1206 07:43:55.707005 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c633de75-e128-4c94-929e-37ed0421f065" containerName="extract-content" Dec 06 07:43:55 crc kubenswrapper[4809]: I1206 07:43:55.707023 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c633de75-e128-4c94-929e-37ed0421f065" containerName="extract-content" Dec 06 07:43:55 crc kubenswrapper[4809]: E1206 07:43:55.707056 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c633de75-e128-4c94-929e-37ed0421f065" containerName="registry-server" Dec 06 07:43:55 crc kubenswrapper[4809]: I1206 07:43:55.707063 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c633de75-e128-4c94-929e-37ed0421f065" containerName="registry-server" Dec 06 07:43:55 crc kubenswrapper[4809]: E1206 07:43:55.707093 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c633de75-e128-4c94-929e-37ed0421f065" containerName="extract-utilities" Dec 06 07:43:55 crc kubenswrapper[4809]: I1206 07:43:55.707103 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c633de75-e128-4c94-929e-37ed0421f065" containerName="extract-utilities" Dec 06 07:43:55 crc kubenswrapper[4809]: I1206 07:43:55.707411 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="c633de75-e128-4c94-929e-37ed0421f065" containerName="registry-server" Dec 06 07:43:55 crc kubenswrapper[4809]: I1206 07:43:55.709875 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w946g" Dec 06 07:43:55 crc kubenswrapper[4809]: I1206 07:43:55.725037 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-w946g"] Dec 06 07:43:55 crc kubenswrapper[4809]: I1206 07:43:55.823294 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqws5\" (UniqueName: \"kubernetes.io/projected/1f9ae1c5-dde6-4899-a336-6cdabd235ce0-kube-api-access-vqws5\") pod \"redhat-marketplace-w946g\" (UID: \"1f9ae1c5-dde6-4899-a336-6cdabd235ce0\") " pod="openshift-marketplace/redhat-marketplace-w946g" Dec 06 07:43:55 crc kubenswrapper[4809]: I1206 07:43:55.823399 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f9ae1c5-dde6-4899-a336-6cdabd235ce0-catalog-content\") pod \"redhat-marketplace-w946g\" (UID: \"1f9ae1c5-dde6-4899-a336-6cdabd235ce0\") " pod="openshift-marketplace/redhat-marketplace-w946g" Dec 06 07:43:55 crc kubenswrapper[4809]: I1206 07:43:55.823549 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f9ae1c5-dde6-4899-a336-6cdabd235ce0-utilities\") pod \"redhat-marketplace-w946g\" (UID: \"1f9ae1c5-dde6-4899-a336-6cdabd235ce0\") " pod="openshift-marketplace/redhat-marketplace-w946g" Dec 06 07:43:55 crc kubenswrapper[4809]: I1206 07:43:55.927386 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqws5\" (UniqueName: \"kubernetes.io/projected/1f9ae1c5-dde6-4899-a336-6cdabd235ce0-kube-api-access-vqws5\") pod \"redhat-marketplace-w946g\" (UID: \"1f9ae1c5-dde6-4899-a336-6cdabd235ce0\") " pod="openshift-marketplace/redhat-marketplace-w946g" Dec 06 07:43:55 crc kubenswrapper[4809]: I1206 07:43:55.927790 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f9ae1c5-dde6-4899-a336-6cdabd235ce0-catalog-content\") pod \"redhat-marketplace-w946g\" (UID: \"1f9ae1c5-dde6-4899-a336-6cdabd235ce0\") " pod="openshift-marketplace/redhat-marketplace-w946g" Dec 06 07:43:55 crc kubenswrapper[4809]: I1206 07:43:55.928054 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f9ae1c5-dde6-4899-a336-6cdabd235ce0-utilities\") pod \"redhat-marketplace-w946g\" (UID: \"1f9ae1c5-dde6-4899-a336-6cdabd235ce0\") " pod="openshift-marketplace/redhat-marketplace-w946g" Dec 06 07:43:55 crc kubenswrapper[4809]: I1206 07:43:55.928781 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f9ae1c5-dde6-4899-a336-6cdabd235ce0-catalog-content\") pod \"redhat-marketplace-w946g\" (UID: \"1f9ae1c5-dde6-4899-a336-6cdabd235ce0\") " pod="openshift-marketplace/redhat-marketplace-w946g" Dec 06 07:43:55 crc kubenswrapper[4809]: I1206 07:43:55.928795 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f9ae1c5-dde6-4899-a336-6cdabd235ce0-utilities\") pod \"redhat-marketplace-w946g\" (UID: \"1f9ae1c5-dde6-4899-a336-6cdabd235ce0\") " pod="openshift-marketplace/redhat-marketplace-w946g" Dec 06 07:43:55 crc kubenswrapper[4809]: I1206 07:43:55.952477 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqws5\" (UniqueName: \"kubernetes.io/projected/1f9ae1c5-dde6-4899-a336-6cdabd235ce0-kube-api-access-vqws5\") pod \"redhat-marketplace-w946g\" (UID: \"1f9ae1c5-dde6-4899-a336-6cdabd235ce0\") " pod="openshift-marketplace/redhat-marketplace-w946g" Dec 06 07:43:56 crc kubenswrapper[4809]: I1206 07:43:56.040049 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w946g" Dec 06 07:43:56 crc kubenswrapper[4809]: I1206 07:43:56.534286 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-w946g"] Dec 06 07:43:56 crc kubenswrapper[4809]: I1206 07:43:56.779260 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w946g" event={"ID":"1f9ae1c5-dde6-4899-a336-6cdabd235ce0","Type":"ContainerStarted","Data":"4b845f2fa9db23b463ec8b81ae78fb5ef5220e00a4a231bff3092dce9bf00574"} Dec 06 07:43:57 crc kubenswrapper[4809]: I1206 07:43:57.793761 4809 generic.go:334] "Generic (PLEG): container finished" podID="1f9ae1c5-dde6-4899-a336-6cdabd235ce0" containerID="e0c7145690a9e93a6c06eecf2b67f34505af936294862b57212de6a5a447e007" exitCode=0 Dec 06 07:43:57 crc kubenswrapper[4809]: I1206 07:43:57.793825 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w946g" event={"ID":"1f9ae1c5-dde6-4899-a336-6cdabd235ce0","Type":"ContainerDied","Data":"e0c7145690a9e93a6c06eecf2b67f34505af936294862b57212de6a5a447e007"} Dec 06 07:43:57 crc kubenswrapper[4809]: I1206 07:43:57.796642 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 07:43:59 crc kubenswrapper[4809]: I1206 07:43:59.817319 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w946g" event={"ID":"1f9ae1c5-dde6-4899-a336-6cdabd235ce0","Type":"ContainerStarted","Data":"c31de653e268da9172b3778935eb07128747d631293594045dafbeb2fdb919f5"} Dec 06 07:44:00 crc kubenswrapper[4809]: I1206 07:44:00.837345 4809 generic.go:334] "Generic (PLEG): container finished" podID="1f9ae1c5-dde6-4899-a336-6cdabd235ce0" containerID="c31de653e268da9172b3778935eb07128747d631293594045dafbeb2fdb919f5" exitCode=0 Dec 06 07:44:00 crc kubenswrapper[4809]: I1206 07:44:00.837456 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w946g" event={"ID":"1f9ae1c5-dde6-4899-a336-6cdabd235ce0","Type":"ContainerDied","Data":"c31de653e268da9172b3778935eb07128747d631293594045dafbeb2fdb919f5"} Dec 06 07:44:02 crc kubenswrapper[4809]: I1206 07:44:02.390822 4809 scope.go:117] "RemoveContainer" containerID="4b28141798e8be5bf2c8cdfd3390a3ddfd37326a66c4ad4407752e73538ea965" Dec 06 07:44:02 crc kubenswrapper[4809]: E1206 07:44:02.391626 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:44:02 crc kubenswrapper[4809]: I1206 07:44:02.866729 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w946g" event={"ID":"1f9ae1c5-dde6-4899-a336-6cdabd235ce0","Type":"ContainerStarted","Data":"da680fa79fb1d5b53593a3486672c749e6766639eb460df8c7957471e670fada"} Dec 06 07:44:02 crc kubenswrapper[4809]: I1206 07:44:02.924132 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-w946g" podStartSLOduration=3.405593604 podStartE2EDuration="7.924102875s" podCreationTimestamp="2025-12-06 07:43:55 +0000 UTC" firstStartedPulling="2025-12-06 07:43:57.796412882 +0000 UTC m=+6762.685395824" lastFinishedPulling="2025-12-06 07:44:02.314922123 +0000 UTC m=+6767.203905095" observedRunningTime="2025-12-06 07:44:02.893516169 +0000 UTC m=+6767.782499121" watchObservedRunningTime="2025-12-06 07:44:02.924102875 +0000 UTC m=+6767.813085817" Dec 06 07:44:06 crc kubenswrapper[4809]: I1206 07:44:06.041192 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-w946g" Dec 06 07:44:06 crc kubenswrapper[4809]: I1206 07:44:06.041784 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-w946g" Dec 06 07:44:06 crc kubenswrapper[4809]: I1206 07:44:06.106623 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-w946g" Dec 06 07:44:16 crc kubenswrapper[4809]: I1206 07:44:16.111114 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-w946g" Dec 06 07:44:16 crc kubenswrapper[4809]: I1206 07:44:16.167953 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-w946g"] Dec 06 07:44:16 crc kubenswrapper[4809]: I1206 07:44:16.389331 4809 scope.go:117] "RemoveContainer" containerID="4b28141798e8be5bf2c8cdfd3390a3ddfd37326a66c4ad4407752e73538ea965" Dec 06 07:44:16 crc kubenswrapper[4809]: E1206 07:44:16.389690 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:44:17 crc kubenswrapper[4809]: I1206 07:44:17.043223 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-w946g" podUID="1f9ae1c5-dde6-4899-a336-6cdabd235ce0" containerName="registry-server" containerID="cri-o://da680fa79fb1d5b53593a3486672c749e6766639eb460df8c7957471e670fada" gracePeriod=2 Dec 06 07:44:18 crc kubenswrapper[4809]: I1206 07:44:18.057480 4809 generic.go:334] "Generic (PLEG): container finished" podID="1f9ae1c5-dde6-4899-a336-6cdabd235ce0" containerID="da680fa79fb1d5b53593a3486672c749e6766639eb460df8c7957471e670fada" exitCode=0 Dec 06 07:44:18 crc kubenswrapper[4809]: I1206 07:44:18.057576 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w946g" event={"ID":"1f9ae1c5-dde6-4899-a336-6cdabd235ce0","Type":"ContainerDied","Data":"da680fa79fb1d5b53593a3486672c749e6766639eb460df8c7957471e670fada"} Dec 06 07:44:18 crc kubenswrapper[4809]: I1206 07:44:18.522359 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w946g" Dec 06 07:44:18 crc kubenswrapper[4809]: I1206 07:44:18.693980 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f9ae1c5-dde6-4899-a336-6cdabd235ce0-catalog-content\") pod \"1f9ae1c5-dde6-4899-a336-6cdabd235ce0\" (UID: \"1f9ae1c5-dde6-4899-a336-6cdabd235ce0\") " Dec 06 07:44:18 crc kubenswrapper[4809]: I1206 07:44:18.694813 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vqws5\" (UniqueName: \"kubernetes.io/projected/1f9ae1c5-dde6-4899-a336-6cdabd235ce0-kube-api-access-vqws5\") pod \"1f9ae1c5-dde6-4899-a336-6cdabd235ce0\" (UID: \"1f9ae1c5-dde6-4899-a336-6cdabd235ce0\") " Dec 06 07:44:18 crc kubenswrapper[4809]: I1206 07:44:18.694884 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f9ae1c5-dde6-4899-a336-6cdabd235ce0-utilities\") pod \"1f9ae1c5-dde6-4899-a336-6cdabd235ce0\" (UID: \"1f9ae1c5-dde6-4899-a336-6cdabd235ce0\") " Dec 06 07:44:18 crc kubenswrapper[4809]: I1206 07:44:18.697995 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f9ae1c5-dde6-4899-a336-6cdabd235ce0-utilities" (OuterVolumeSpecName: "utilities") pod "1f9ae1c5-dde6-4899-a336-6cdabd235ce0" (UID: "1f9ae1c5-dde6-4899-a336-6cdabd235ce0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:44:18 crc kubenswrapper[4809]: I1206 07:44:18.705349 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f9ae1c5-dde6-4899-a336-6cdabd235ce0-kube-api-access-vqws5" (OuterVolumeSpecName: "kube-api-access-vqws5") pod "1f9ae1c5-dde6-4899-a336-6cdabd235ce0" (UID: "1f9ae1c5-dde6-4899-a336-6cdabd235ce0"). InnerVolumeSpecName "kube-api-access-vqws5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:44:18 crc kubenswrapper[4809]: I1206 07:44:18.735283 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f9ae1c5-dde6-4899-a336-6cdabd235ce0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1f9ae1c5-dde6-4899-a336-6cdabd235ce0" (UID: "1f9ae1c5-dde6-4899-a336-6cdabd235ce0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 07:44:18 crc kubenswrapper[4809]: I1206 07:44:18.799274 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f9ae1c5-dde6-4899-a336-6cdabd235ce0-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 07:44:18 crc kubenswrapper[4809]: I1206 07:44:18.799329 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f9ae1c5-dde6-4899-a336-6cdabd235ce0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 07:44:18 crc kubenswrapper[4809]: I1206 07:44:18.799355 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vqws5\" (UniqueName: \"kubernetes.io/projected/1f9ae1c5-dde6-4899-a336-6cdabd235ce0-kube-api-access-vqws5\") on node \"crc\" DevicePath \"\"" Dec 06 07:44:19 crc kubenswrapper[4809]: I1206 07:44:19.083956 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w946g" event={"ID":"1f9ae1c5-dde6-4899-a336-6cdabd235ce0","Type":"ContainerDied","Data":"4b845f2fa9db23b463ec8b81ae78fb5ef5220e00a4a231bff3092dce9bf00574"} Dec 06 07:44:19 crc kubenswrapper[4809]: I1206 07:44:19.083998 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w946g" Dec 06 07:44:19 crc kubenswrapper[4809]: I1206 07:44:19.084009 4809 scope.go:117] "RemoveContainer" containerID="da680fa79fb1d5b53593a3486672c749e6766639eb460df8c7957471e670fada" Dec 06 07:44:19 crc kubenswrapper[4809]: I1206 07:44:19.105102 4809 scope.go:117] "RemoveContainer" containerID="c31de653e268da9172b3778935eb07128747d631293594045dafbeb2fdb919f5" Dec 06 07:44:19 crc kubenswrapper[4809]: I1206 07:44:19.137347 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-w946g"] Dec 06 07:44:19 crc kubenswrapper[4809]: I1206 07:44:19.153807 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-w946g"] Dec 06 07:44:19 crc kubenswrapper[4809]: I1206 07:44:19.158175 4809 scope.go:117] "RemoveContainer" containerID="e0c7145690a9e93a6c06eecf2b67f34505af936294862b57212de6a5a447e007" Dec 06 07:44:19 crc kubenswrapper[4809]: I1206 07:44:19.401131 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f9ae1c5-dde6-4899-a336-6cdabd235ce0" path="/var/lib/kubelet/pods/1f9ae1c5-dde6-4899-a336-6cdabd235ce0/volumes" Dec 06 07:44:30 crc kubenswrapper[4809]: I1206 07:44:30.390274 4809 scope.go:117] "RemoveContainer" containerID="4b28141798e8be5bf2c8cdfd3390a3ddfd37326a66c4ad4407752e73538ea965" Dec 06 07:44:30 crc kubenswrapper[4809]: E1206 07:44:30.391887 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" Dec 06 07:44:44 crc kubenswrapper[4809]: I1206 07:44:44.389952 4809 scope.go:117] "RemoveContainer" containerID="4b28141798e8be5bf2c8cdfd3390a3ddfd37326a66c4ad4407752e73538ea965" Dec 06 07:44:44 crc kubenswrapper[4809]: E1206 07:44:44.390779 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-npms2_openshift-machine-config-operator(cbf4cf62-024e-4703-a8b8-9aecda9cd26a)\"" pod="openshift-machine-config-operator/machine-config-daemon-npms2" podUID="cbf4cf62-024e-4703-a8b8-9aecda9cd26a" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515114757374024463 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015114757374017400 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015114741567016520 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015114741567015470 5ustar corecore